Overview
This product can help to stress the inference server with concurrent queries with custom large data and analyse the server resource utilization (e.g. GPU utilization, GPU memory, CPU utilization and CPU memory) against one of multiple GPUs. Monthly charge is for support and customization on the go.
Highlights
- This product can help to determine and analyse the large data
- You can input any JSON-based data url. The server is able to ingest data and using those data, you can chat anything with those data
- Prior support provide on-mail and customization on the go
Details
Unlock automation with AI agent solutions

Features and programs
Financing for AWS Marketplace purchases
Pricing
Vendor refund policy
No refund policy
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
Run the container
- Amazon ECS
- Amazon ECS Anywhere
Container image
Containers are lightweight, portable execution environments that wrap server application software in a filesystem that includes everything it needs to run. Container applications run on supported container runtimes and orchestration services, such as Amazon Elastic Container Service (Amazon ECS) or Amazon Elastic Kubernetes Service (Amazon EKS). Both eliminate the need for you to install and operate your own container orchestration software by managing and scheduling containers on a scalable cluster of virtual machines.
Version release notes
Release notes
- docker image can be downloaded from the continer registry withoiut requireing AMI without any AWS dependency
- AIS docker can be directly deployed to local servers
Additional details
Usage instructions
Here is guide to run the service in your docker based machine.
Prerequisites
- AWS cli
- Docker
Step 1: Authenticate with AWS ECR
- Before pulling the container image, authenticate your local machine or AWS service with ECR. You can use below command
- aws ecr get-login-password --region us-east-1 | docker login --username AWS --password-stdin 709825985650.dkr.ecr.us-east-1.amazonaws.com
Step 2: Define the Container Images
- CONTAINER_IMAGES="709825985650.dkr.ecr.us-east-1.amazonaws.com/bhojr/bhojr/ai-inference-str ess-container:2.2"
Step 3: Pull the Docker Images for i in $(echo $CONTAINER_IMAGES | sed "s/,/ /g"); do docker pull $i; done
Step 4: Run the Docker Images docker run -p 8080:8080 -d 709825985650.dkr.ecr.us-east-1.amazonaws.com/bhojr/bhojr/ai-inference-stress-container:2.2
Once all the above steps are done,. You can go to the site as http://<your__ip>:<service_port>.
You need to enter the license key, which can be obtained from the authorized section of the Baideac web page: [https://www.baideac.com/licensing.html ].
Licensing Instructions:
We provide product services under a free recurring license. You can visit the official licensing page at [https://www.baideac.com/licensing.html ]. Once you obtain the free license, you will be able to access the platform. This product is under BYOL [bring your own license].
Steps to Obtain a License:
- Create an account at [https://www.baideac.com/licensing.html ] and select "Trial Version" as the account type.
- In the "Product Dashboard," submit a request for a trial key by selecting the product "AI Inference Server."
- Copy the license key from the table.
You can use this license key to access the platform.
Here is product link, can find all required instructions: https://www.baideac.com/ai-inference-stresser.htmlÂ
Resources
Vendor resources
Support
Vendor support
Support can be available at the mail address support@baideac.comÂ
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.
Similar products
Customer reviews
Great tool to validate the underlying AI Inference Infrastructure
Easy to deploy container image. Once installed, we successfully ran some basic AI models (as included in the image) to confirm the performance of our potential AI server. This helped avoid the under-provisioning and over-provisioning of GPUs and memory.