Overview
Reduce inference costs and increase inference speed by using Multiverse Computing's CompactifAI Mistral Small 3.1 Slim. A 50% compression of the widely known Mistral Small 3.1 model.
Highlights
- 70% Infrastructure OpEx Reduction
- 2x Faster Inference Throughput
Details
Introducing multi-product solutions
You can now purchase comprehensive solutions tailored to use cases and industries.
Features and programs
Financing for AWS Marketplace purchases
Pricing
Dimension | Description | Cost/host/hour |
|---|---|---|
ml.g5.12xlarge Inference (Batch) Recommended | Model inference on the ml.g5.12xlarge instance type, batch mode | $0.98 |
ml.g5.12xlarge Inference (Real-Time) Recommended | Model inference on the ml.g5.12xlarge instance type, real-time mode | $0.98 |
ml.g5.16xlarge Inference (Batch) | Model inference on the ml.g5.16xlarge instance type, batch mode | $0.74 |
ml.g5.24xlarge Inference (Batch) | Model inference on the ml.g5.24xlarge instance type, batch mode | $1.35 |
ml.g5.48xlarge Inference (Batch) | Model inference on the ml.g5.48xlarge instance type, batch mode | $2.57 |
ml.g5.16xlarge Inference (Real-Time) | Model inference on the ml.g5.16xlarge instance type, real-time mode | $0.74 |
ml.g5.24xlarge Inference (Real-Time) | Model inference on the ml.g5.24xlarge instance type, real-time mode | $1.35 |
ml.g5.48xlarge Inference (Real-Time) | Model inference on the ml.g5.48xlarge instance type, real-time mode | $2.57 |
ml.p4d.24xlarge Inference (Real-Time) | Model inference on the ml.p4d.24xlarge instance type, real-time mode | $3.15 |
ml.p5.48xlarge Inference (Real-Time) | Model inference on the ml.p5.48xlarge instance type, real-time mode | $7.72 |
Vendor refund policy
Placeholder Refund Policy
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
Amazon SageMaker model
An Amazon SageMaker model package is a pre-trained machine learning model ready to use without additional training. Use the model package to create a model on Amazon SageMaker for real-time inference or batch processing. Amazon SageMaker is a fully managed platform for building, training, and deploying machine learning models at scale.
Version release notes
CompactifAI Mistral Small 3.1 Slim inference on a vLLM engine with text-only support
Additional details
Inputs
- Summary
Our model accepts an input compatible with openAI's chat completions endpoint.
- Input MIME type
- application/json
Support
Vendor support
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.