AWS HPC Blog
Category: Technical How-to
Accelerating HPC Deployment with AWS Parallel Computing Service and Kiro CLI
Research teams moving from on-premises HPC environments often struggle with the complexity of cloud deployment. Traditional approaches require deep expertise in AWS networking, storage architectures, and Slurm configuration management. A typical manual deployment involves weeks of infrastructure provisioning, network topology design, scheduler configuration, and performance tuning. Research teams with limited platform engineering resources find themselves […]
End-to-end scalable vision intelligence pipeline using LIDAR 3D Point Clouds on AWS
Elevate your mining and construction business to new heights with the power of 3D world modeling and AI-driven scene interpretation. Our latest blog post delves into the underlying technologies, from SLAM and photogrammetry to point-cloud analysis, and demonstrates how to scale the execution of these compute-intensive algorithms on AWS. Unlock the potential to optimize your workflows, improve decision-making, and drive sustainable growth.
Enhanced Performance for Whisper Audio Transcription on AWS Batch and AWS Inferentia
In this post, we’ll review the key optimizations and performance gains for our Whisper audio transcription solution powered by AWS Batch and AWS Inferentia.
Introducing managed accounting for AWS Parallel Computing Service
AWS Parallel Computing Service (AWS PCS) now supports accounting, a Slurm feature that enables you to monitor resource utilization, enforce resource limits, and manage access-control to specific capacity across users and projects in a cluster. AWS PCS manages the accounting database for the cluster, so that you don’t have to setup and manage a separate accounting database. In this post, we’ll show you how this works, and point you to some actual use cases you can try yourself.
Adding functionality to your applications using multiple containers in AWS Batch
Discover how to coordinate multiple applications in separate containers within a single AWS Batch job definition. Learn the benefits of this approach and how to share resources between containers for more efficient, scalable deployments.
How to use rate-limited resources in AWS Batch jobs with resource aware scheduling
Struggling with bottlenecks in your batch processing? AWS Batch’s new resource aware scheduling capability could be the solution your business needs. This feature allows you to define and manage consumable resources, helping maximize the use of your compute power. Check out our blog to learn more.
Accelerating file reads with a storage caching server
Looking to boost storage performance for your #HPC workloads on AWS? Find out how to build a high-speed EC2 cache for under $4/hour that’s fit for purpose in this new blog post.
Scaling your LLM inference workloads: multi-node deployment with TensorRT-LLM and Triton on Amazon EKS
LLMs are scaling exponentially. Learn how advanced technologies like Triton, TRT-LLM and EKS enable seamless deployment of models like the 405B parameter Llama 3.1. Let’s go large.
On-demand visual login nodes – using RES with AWS Parallel Computing Service
Running large-scale modeling and simulations just got easier. Check out our new post on integrating AWS Parallel Computing Service with Research and Engineering Studio for individualized access to HPC resources.
Building a secure and compliant HPC environment on AWS following NIST SP 800-223
Check out our latest blog post to learn how AWS enables building secure, compliant high performance computing (HPC) environments aligned with NIST SP 800-223 guidelines. We walk through the key components, security considerations, and steps for deploying a zone-based HPC architecture on AWS.








