Listing Thumbnail

    Fireworks

     Info
    Deployed on AWS
    Fireworks.ai offers a generative AI platform as a service. We optimize for rapid product iteration building on top of gen AI as well as minimizing cost to serve.

    Overview

    Experience the fastest inference and fine-tuning platform with Fireworks AI. Utilize state-of-the-art open-source models, fine-tune them, or deploy your own at no additional cost. Access a diverse library of models across various modalities - including text, vision, embedding, audio, image, and multimodal - to build and scale your AI applications efficiently.

    • Blazing fast inference for 100+ models
    • Fine-tune and deploy in minutes
    • Building blocks for compound AI systems

    Start in seconds and pay-per-token with our serverless deployment. Or Use our dedicated deployments, fully optimized to your use case.

    Highlights

    • Instantly run popular and specialized models, including DeepSeek R1, Llama3, Mixtral, and Stable Diffusion, optimized for peak latency, throughput, and context length. Fireattention custom CUDA kernel, serves models four times faster than vLLM without compromising quality.
    • Fine-tune with our LoRA-based service, twice as cost-efficient as other providers. Instantly deploy and switch between up to 100 fine-tuned models to experiment without extra costs. Serve models at blazing-fast speeds of up to 300 tokens per second on our serverless inference platform.
    • Leverage the building blocks for compound AI systems. Handle tasks with multiple models, modalities, and external APIs and data instead of relying on a single model. Use FireFunction, a SOTA function calling model, to compose compound AI systems for RAG, search, and domain-expert copilots for automation, code, math, medicine, and more.

    Details

    Delivery method

    Deployed on AWS

    Unlock automation with AI agent solutions

    Fast-track AI initiatives with agents, tools, and solutions from AWS Partners.
    AI Agents

    Features and programs

    Financing for AWS Marketplace purchases

    AWS Marketplace now accepts line of credit payments through the PNC Vendor Finance program. This program is available to select AWS customers in the US, excluding NV, NC, ND, TN, & VT.
    Financing for AWS Marketplace purchases

    Pricing

    Pricing is based on the duration and terms of your contract with the vendor, and additional usage. You pay upfront or in installments according to your contract terms with the vendor. This entitles you to a specified quantity of use for the contract duration. Usage-based pricing is in effect for overages or additional usage not covered in the contract. These charges are applied on top of the contract price. If you choose not to renew or replace your contract before the contract end date, access to your entitlements will expire.
    Additional AWS infrastructure costs may apply. Use the AWS Pricing Calculator  to estimate your infrastructure costs.

    12-month contract (1)

     Info
    Dimension
    Description
    Cost/12 months
    Enterprise
    Unlimited deployment models
    $500,000.00

    Additional usage costs (1)

     Info

    The following dimensions are not included in the contract terms, which will be charged based on your usage.

    Dimension
    Description
    Cost/unit
    additionalusage
    Additional Usage
    $1.00

    Vendor refund policy

    All fees are non-refundable and non-cancellable except as required by law.

    How can we make this page better?

    We'd like to hear your feedback and ideas on how to improve this page.
    We'd like to hear your feedback and ideas on how to improve this page.

    Legal

    Vendor terms and conditions

    Upon subscribing to this product, you must acknowledge and agree to the terms and conditions outlined in the vendor's End User License Agreement (EULA) .

    Content disclaimer

    Vendors are responsible for their product descriptions and other product content. AWS does not warrant that vendors' product descriptions or other product content are accurate, complete, reliable, current, or error-free.

    Usage information

     Info

    Delivery details

    Software as a Service (SaaS)

    SaaS delivers cloud-based software applications directly to customers over the internet. You can access these applications through a subscription model. You will pay recurring monthly usage fees through your AWS bill, while AWS handles deployment and infrastructure management, ensuring scalability, reliability, and seamless integration with other AWS services.

    Support

    Vendor support

    Email support services are available from Monday to Friday.
    support@fireworks.ai 

    AWS infrastructure support

    AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.

    Product comparison

     Info
    Updated weekly

    Accolades

     Info
    Top
    10
    In Finance & Accounting, Research
    Top
    10
    In Summarization-Text, Generation-Text
    Top
    10
    In Procurement & Supply Chain

    Overview

     Info
    AI generated from product descriptions
    Model Inference Performance
    Supports blazing fast inference for 100+ models with custom CUDA kernel Fireattention, serving models four times faster than vLLM
    Model Fine-tuning Capabilities
    Provides LoRA-based fine-tuning service with ability to instantly deploy and switch between up to 100 fine-tuned models
    Multi-modal Model Support
    Offers diverse library of models across text, vision, embedding, audio, image, and multimodal modalities
    Compound AI System Architecture
    Enables task handling with multiple models, modalities, external APIs and data using advanced function calling capabilities
    Advanced Model Optimization
    Supports specialized open-source models with optimizations for latency, throughput, and extended context length
    Model Quantization Support
    Supports multi-bit integer quantization from 2-bit to 8-bit, enabling efficient model inference on limited GPU memory
    GPU and CPU Inference
    Capable of running inference simultaneously on GPU and CPU, allowing processing of larger models across different hardware resources
    Model Compatibility
    Supports diverse language models including LLaMA, LLaMA 2, Falcon, Alpaca, GPT4All, Vicuna, Mistral AI, and multiple multilingual models
    Inference Framework
    Utilizes llama.cpp with plain C/C++ implementation, offering efficient and lightweight model inference without complex dependencies
    Architecture Optimization
    Provides support for x86 architecture extensions including AVX, AVX2, and AVX512 for enhanced computational performance
    Large Language Model Support
    Provides access to latest large language models (LLMs) with a versatile experimentation environment
    No-Code Application Development
    Enables building and deploying generative AI applications through visual interface without complex coding
    Enterprise Security Framework
    Implements enterprise-level security with credentials management, PII masking, data encryption, and role-based access controls
    AI Workflow Observability
    Offers detailed operational dashboards to track generative AI usage, performance, trends, and model interactions
    Trust and Safety Mechanisms
    Includes content filtering controls to reduce noise, block harmful content, and provide relevant citations for generated outputs

    Contract

     Info
    Standard contract

    Customer reviews

    Ratings and reviews

     Info
    0 ratings
    5 star
    4 star
    3 star
    2 star
    1 star
    0%
    0%
    0%
    0%
    0%
    0 AWS reviews
    |
    3 external reviews
    Star ratings include only reviews from verified AWS customers. External reviews can also include a star rating, but star ratings from external reviews are not averaged in with the AWS customer star ratings.
    Liraz A.

    One Stop AI Model Shop

    Reviewed on Nov 14, 2024
    Review provided by G2
    What do you like best about the product?
    So many AI models to choos from... Love the option of the playground
    What do you dislike about the product?
    pretty hard to get started. they really need a quickstart guide.
    and beacuse the site is so full of featurs - a tour would be nice.
    What problems is the product solving and how is that benefiting you?
    helping me choose the right model for my day to day use.
    reviewer2588646

    Enhanced text-to-image creation with solid API and fine-tuning support

    Reviewed on Nov 06, 2024
    Review provided by PeerSpot

    What is our primary use case?

    We primarily use Fireworks AI  for text-to-image generation. We are developing a platform for artists to sell their art styles, where the system helps them tune a model and then sell images generated from their signature.

    How has it helped my organization?

    Fireworks AI  has helped our organization by enabling us to create a platform for artists to sell their art styles. I am not the user of the solution. I'm the developer. It helps me do my job effectively.

    What is most valuable?

    Fireworks AI has a solid API and is quite easy to interact with. It has better documentation and logs, which are important for me as a developer. Additionally, it has a bigger infrastructure and provides nice support for fine-tuning the Flux  AI model.

    What needs improvement?

    Returning the values charged for each event generation would improve Fireworks AI. When using the API, it does not return information about the charges for image generation, which would be useful for our solution.

    For how long have I used the solution?

    I have been using Fireworks AI for about four months.

    What do I think about the stability of the solution?

    Fireworks AI is pretty stable, and I have not encountered any problems.

    What do I think about the scalability of the solution?

    Fireworks AI offers a very complete API, and its scalability is impressive.

    Which solution did I use previously and why did I switch?

    I previously used Okta. It was discontinued, so we opted for Fireworks AI.

    How was the initial setup?

    The initial setup was fairly easy. It took about eight to ten days, including integrating it into our solution, testing, and moving from scratch to production.

    What's my experience with pricing, setup cost, and licensing?

    I cannot comment on pricing or setup cost since others handle that aspect. As a developer, I primarily use the API.

    Which other solutions did I evaluate?

    I have evaluated SAL as an alternative solution.

    What other advice do I have?

    I'd rate the solution ten out of ten.

    Which deployment model are you using for this solution?

    Public Cloud

    If public cloud, private cloud, or hybrid cloud, which cloud provider do you use?

    Other
    Pratiksh S.

    Review for Fireworks AI

    Reviewed on Sep 05, 2024
    Review provided by G2
    What do you like best about the product?
    They have categorised the models according to users requirements and user have to pay for the products they use. No extra costing.
    What do you dislike about the product?
    They need to use more dependable parameters. And should increase their serverless model limits.
    What problems is the product solving and how is that benefiting you?
    AI is the booming condition in the industry and with the Fireworks it feels easy to deploy models to organisational servers. Additionally they use Meta Llama.
    View all reviews