Listing Thumbnail

    HAMi - GPU Virtualization & Unified Scheduling for K8s

     Info
    Deployed on AWS
    Quick Launch
    HAMi is an open-source Kubernetes middleware providing GPU compute and memory isolation, flexible GPU slicing, and topology-aware scheduling maximizing utilization for AI inference and training workloads across NVIDIA GPUs and AWS Neuron devices; part of the CNCF ecosystem. Works with NVIDIA GPU Operator, vLLM Production Stack, and Xinference.

    Overview

    HAMi (Heterogeneous AI Computing Virtualization Middleware) is an open-source solution for GPU virtualization and unified scheduling in Kubernetes clusters. It enables compute power limiting and memory isolation for GPUs at the Kubernetes layer, allowing multiple workloads to share the same physical GPU without interference. For AWS Neuron devices , HAMi integrates with the AWS Neuron Device Plugin to support shared access, while providing a unified scheduling layer for both GPU and Neuron resources. Key capabilities:

    1. GPU virtualization: Split NVIDIA GPUs by compute percentage and memory size, enabling fine-grained sharing for inference and training.
    2. AWS Neuron integration: Works with AWS Neuron Device Plugin to enable shared usage of Inferentia and Trainium devices, with HAMi handling unified scheduling.
    3. Topology-aware scheduling: Optimize workload placement based on NUMA and NVLink topology for better performance.
    4. Works with popular AI stacks: Fully compatible with NVIDIA GPU Operator, Xinference, and vLLM Production Stack. HAMi is ideal for AI inference, large model serving, and training workloads on Kubernetes, helping maximize hardware utilization across mixed GPU and AWS Neuron environments.

    Highlights

    • Unified Kubernetes management for NVIDIA GPUs and AWS Neuron devices.
    • Fine-grained GPU slicing and topology-aware scheduling for optimal utilization.
    • Works with NVIDIA GPU Operator, Xinference, and vLLM Production Stack.

    Details

    Delivery method

    Supported services

    Delivery option
    helm delivery method

    Latest version

    Operating system
    Linux

    Deployed on AWS

    Unlock automation with AI agent solutions

    Fast-track AI initiatives with agents, tools, and solutions from AWS Partners.
    AI Agents

    Features and programs

    Financing for AWS Marketplace purchases

    AWS Marketplace now accepts line of credit payments through the PNC Vendor Finance program. This program is available to select AWS customers in the US, excluding NV, NC, ND, TN, & VT.
    Financing for AWS Marketplace purchases

    Quick Launch

    Leverage AWS CloudFormation templates to reduce the time and resources required to configure, deploy, and launch your software.

    Pricing

    HAMi - GPU Virtualization & Unified Scheduling for K8s

     Info
    This product is available free of charge. Free subscriptions have no end date and may be canceled any time.
    Additional AWS infrastructure costs may apply. Use the AWS Pricing Calculator  to estimate your infrastructure costs.

    Vendor refund policy

    This product is free of charge. Refunds are not applicable.

    How can we make this page better?

    We'd like to hear your feedback and ideas on how to improve this page.
    We'd like to hear your feedback and ideas on how to improve this page.

    Legal

    Vendor terms and conditions

    Upon subscribing to this product, you must acknowledge and agree to the terms and conditions outlined in the vendor's End User License Agreement (EULA) .

    Content disclaimer

    Vendors are responsible for their product descriptions and other product content. AWS does not warrant that vendors' product descriptions or other product content are accurate, complete, reliable, current, or error-free.

    Usage information

     Info

    Delivery details

    helm delivery method

    Supported services: Learn more 
    • Amazon EKS
    • Amazon EKS Anywhere
    Helm chart

    Helm charts are Kubernetes YAML manifests combined into a single package that can be installed on Kubernetes clusters. The containerized application is deployed on a cluster by running a single Helm install command to install the seller-provided Helm chart.

    Version release notes

    Basing on latest open-source release, changed chart to using it at AWS.

    Additional details

    Usage instructions

    Installation: https://project-hami.io/docs/next/installation/aws-installation/  Usage: First, Label your GPU nodes for scheduling with HAMi by adding the label "gpu=on". Without this label, the nodes cannot be managed by our scheduler. kubectl label nodes {nodeid} gpu=on Then you can try the example as example.yaml: apiVersion: v1 kind: Pod metadata: name: gpu-pod spec: containers: - name: ubuntu-container image: ubuntu:22.04 command: ["bash", "-c", "sleep 86400"] resources: limits: nvidia.com/gpu: 1 # declare how many physical GPUs the pod needs nvidia.com/gpumem: 3000 # identifies 3000M GPU memory each physical GPU allocates to the pod (Optional Integer) nvidia.com/gpucores: 30 # identifies 30% GPU GPU core each physical GPU allocates to the pod (Optional Integer) kubectl apply -f example.yaml You can follow the example to declear your own manifest to create workloads using the virtualized.

    Support

    Vendor support

    Email: info@dynamia.ai  GitHub Issues: https://github.com/Project-HAMi/HAMi/issues  Community support is available via GitHub issues and email.

    AWS infrastructure support

    AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.

    Customer reviews

    Ratings and reviews

     Info
    0 ratings
    5 star
    4 star
    3 star
    2 star
    1 star
    0%
    0%
    0%
    0%
    0%
    0 AWS reviews
    No customer reviews yet
    Be the first to review this product . We've partnered with PeerSpot to gather customer feedback. You can share your experience by writing or recording a review, or scheduling a call with a PeerSpot analyst.