Overview
HAMi (Heterogeneous AI Computing Virtualization Middleware) is an open-source solution for GPU virtualization and unified scheduling in Kubernetes clusters. It enables compute power limiting and memory isolation for GPUs at the Kubernetes layer, allowing multiple workloads to share the same physical GPU without interference. For AWS Neuron devices , HAMi integrates with the AWS Neuron Device Plugin to support shared access, while providing a unified scheduling layer for both GPU and Neuron resources. Key capabilities:
- GPU virtualization: Split NVIDIA GPUs by compute percentage and memory size, enabling fine-grained sharing for inference and training.
- AWS Neuron integration: Works with AWS Neuron Device Plugin to enable shared usage of Inferentia and Trainium devices, with HAMi handling unified scheduling.
- Topology-aware scheduling: Optimize workload placement based on NUMA and NVLink topology for better performance.
- Works with popular AI stacks: Fully compatible with NVIDIA GPU Operator, Xinference, and vLLM Production Stack. HAMi is ideal for AI inference, large model serving, and training workloads on Kubernetes, helping maximize hardware utilization across mixed GPU and AWS Neuron environments.
Highlights
- Unified Kubernetes management for NVIDIA GPUs and AWS Neuron devices.
- Fine-grained GPU slicing and topology-aware scheduling for optimal utilization.
- Works with NVIDIA GPU Operator, Xinference, and vLLM Production Stack.
Details
Unlock automation with AI agent solutions

Features and programs
Financing for AWS Marketplace purchases
Quick Launch
Pricing
Vendor refund policy
This product is free of charge. Refunds are not applicable.
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
helm delivery method
- Amazon EKS
- Amazon EKS Anywhere
Helm chart
Helm charts are Kubernetes YAML manifests combined into a single package that can be installed on Kubernetes clusters. The containerized application is deployed on a cluster by running a single Helm install command to install the seller-provided Helm chart.
Version release notes
Basing on latest open-source release, changed chart to using it at AWS.
Additional details
Usage instructions
Installation: https://project-hami.io/docs/next/installation/aws-installation/Â Usage: First, Label your GPU nodes for scheduling with HAMi by adding the label "gpu=on". Without this label, the nodes cannot be managed by our scheduler. kubectl label nodes {nodeid} gpu=on Then you can try the example as example.yaml: apiVersion: v1 kind: Pod metadata: name: gpu-pod spec: containers: - name: ubuntu-container image: ubuntu:22.04 command: ["bash", "-c", "sleep 86400"] resources: limits: nvidia.com/gpu: 1 # declare how many physical GPUs the pod needs nvidia.com/gpumem: 3000 # identifies 3000M GPU memory each physical GPU allocates to the pod (Optional Integer) nvidia.com/gpucores: 30 # identifies 30% GPU GPU core each physical GPU allocates to the pod (Optional Integer) kubectl apply -f example.yaml You can follow the example to declear your own manifest to create workloads using the virtualized.
Resources
Vendor resources
Support
Vendor support
Email: info@dynamia.ai GitHub Issues: https://github.com/Project-HAMi/HAMi/issues Community support is available via GitHub issues and email.
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.