Listing Thumbnail

    CompactifAI

     Info
    Deployed on AWS
    CompactifAI API empowers organizations with ultra-efficient and scalable AI models that slash compute and energy costs, accelerate deployment, and fuel innovation, all without compromising performance or reliability!

    Overview

    Play video

    CompactifAI API delivers a new class of compressed, ultra-efficient AI models engineered for performance, sustainability, and adaptability. These dramatically reduce compute and energy costs while maintaining enterprise-grade accuracy and speed. CompactifAI adapts seamlessly to diverse environments, ensuring consistent, high-performance AI. With turnkey software and an intuitive API, it offers cost effective inference enabling enterprises to innovate faster, scale cost-effectively, and unlock competitive edge.

    Highlights

    • 50% Reduction on Operational Costs: Significantly lowers compute and energy expenses across AI workloads.
    • 2x Faster Inference: Delivers faster inference speeds with lower latency and minimal resource consumption.
    • High Throughput: Enables higher volume processing to serve more users simultaneously without sacrificing quality.

    Details

    Delivery method

    Deployed on AWS

    Unlock automation with AI agent solutions

    Fast-track AI initiatives with agents, tools, and solutions from AWS Partners.
    AI Agents

    Features and programs

    Financing for AWS Marketplace purchases

    AWS Marketplace now accepts line of credit payments through the PNC Vendor Finance program. This program is available to select AWS customers in the US, excluding NV, NC, ND, TN, & VT.
    Financing for AWS Marketplace purchases

    Pricing

    Pricing is based on actual usage, with charges varying according to how much you consume. Subscriptions have no end date and may be canceled any time.
    Additional AWS infrastructure costs may apply. Use the AWS Pricing Calculator  to estimate your infrastructure costs.

    Usage costs (24)

     Info
    Dimension
    Description
    Cost/unit
    Llama 3.3 70B Slim by CompactifAI
    Price per 1 million input tokens
    $0.16
    Mistral Small 3.1 Slim by CompactifAI
    Price per 1 million input tokens
    $0.05
    DeepSeek R1 Slim by CompactifAI (coming soon)
    Price per 1 million input tokens
    $0.28
    Llama 4 Scout Slim by CompactifAI
    Price per 1 million input tokens
    $0.07
    Llama 3.1 8B Slim by CompactifAI
    Price per 1 million input tokens
    $0.05
    Llama 3.3 70B Slim by CompactifAI
    Price per 1 million output tokens
    $0.31
    Mistral Small 3.1 Slim by CompactifAI
    Price per 1 million output tokens
    $0.08
    DeepSeek R1 Slim by CompactifAI (coming soon)
    Price per 1 million output tokens
    $0.44
    Llama 4 Scout Slim by CompactifAI
    Price per 1 million output tokens
    $0.10
    Llama 3.1 8B Slim by CompactifAI
    Price per 1 million output tokens
    $0.07

    AI Insights

     Info

    Dimensions summary

    The pricing dimensions for CompactifAI are organized around different opensource language models (like Llama, Mistral, and DeepSeek) with their respective Slim versions — compressed variants featuring up to 95% size reduction, enabling up to 2x faster inference speed and up to 50% lower inference costs, with only an average 3% precision drop. Each of them is priced separately for input and output tokens: Input tokens represent the text sent to the model, while output tokens represent the text generated by the model, with pricing calculated per million tokens.

    Top-of-mind questions for buyers like you

    How am I charged for using these language models?
    You are charged based on the number of tokens processed, with separate pricing for input tokens (text you send) and output tokens (text generated). Pricing is calculated per million tokens, and varies depending on which model version you choose (standard or Slim versions).
    What is the difference between standard and Slim versions of the models?
    Slim versions are optimized variants of the standard models, developed using CompactifAI technology. These versions feature up to 95% size reduction, enabling up to 2x faster inference speed and up to 50% lower inference costs, with only an average 3% precision drop. All in addition to a highly competitive pricing, significantly below the industry average.
    How do I choose between different models like Llama 3.3 70B Slim or Mistral Small 3.1 Slim?
    Slim models available via the CompactifAI API offer different trade-offs in performance and efficiency, much alike their standard versions. Larger models like Llama 3.3 70B Slim provide higher accuracy and reasoning power, while smaller ones like Mistral Small 3.1 Slim are ideal for cost-effective, high-speed tasks. The advantage of the Slim versions compared to their original counterparts is that, on average, they can deliver up to 2Ă— faster inference and up to 50% cost reduction with minimal accuracy drop.

    Vendor refund policy

    Get in touch for Refund Policy.

    Custom pricing options

    Request a private offer to receive a custom quote.

    How can we make this page better?

    We'd like to hear your feedback and ideas on how to improve this page.
    We'd like to hear your feedback and ideas on how to improve this page.

    Legal

    Vendor terms and conditions

    Upon subscribing to this product, you must acknowledge and agree to the terms and conditions outlined in the vendor's End User License Agreement (EULA) .

    Content disclaimer

    Vendors are responsible for their product descriptions and other product content. AWS does not warrant that vendors' product descriptions or other product content are accurate, complete, reliable, current, or error-free.

    Usage information

     Info

    Delivery details

    Software as a Service (SaaS)

    SaaS delivers cloud-based software applications directly to customers over the internet. You can access these applications through a subscription model. You will pay recurring monthly usage fees through your AWS bill, while AWS handles deployment and infrastructure management, ensuring scalability, reliability, and seamless integration with other AWS services.

    Resources

    Vendor resources

    Support

    Vendor support

    Use our technical support form  to request both technical support or other type of enquiries like billing, etc. Alternatively you can also email us at support@compactif.ai  or resources page 

    AWS infrastructure support

    AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.

    Customer reviews

    Ratings and reviews

     Info
    0 ratings
    5 star
    4 star
    3 star
    2 star
    1 star
    0%
    0%
    0%
    0%
    0%
    0 AWS reviews
    No customer reviews yet
    Be the first to review this product . We've partnered with PeerSpot to gather customer feedback. You can share your experience by writing or recording a review, or scheduling a call with a PeerSpot analyst.