Listing Thumbnail

    Instant RAGFlow: Ready-to-Use AI Knowledge Retrieval Engine

     Info
    Deployed on AWS
    This product has charges associated with it for seller support. Deploy advanced retrieval-augmented workflows with local or remote LLMs like OpenAI, Deepseek, Qwen, Mistral, complete control, high flexibility, and enterprise-grade security

    Overview

    This is a repackaged open source software product wherein additional charges apply for support by TechLatest.net.

    Important: For step by step guide on how to setup this vm , please refer to our Getting Started guide 

    Deploy a ready-to-use virtual machine powered by Ragflow and Ollama, fully loaded with lead-ing open-source language models and optimized for high-performance GPU inference.

    What is Inside:

    1 - Ragflow: End-to-End RAG Workflow Orchestration

    Ragflow is an open-source framework purpose-built for Retrieval-Augmented Generation (RAG) pipelines for deep document understanding. It lets you easily build, manage, and deploy AI systems that combine LLM reasoning with your proprietary or domain-specific data.

    Offering features such as:

    • Deep Document Understanding: Intelligent layout analysis, template-based chunking (for PDFs, tables, resumes, legal docs, etc.), visual chunking, and explainable citations that reduce hallucinations and support traceability Quality-In, Quality-Out: High-fidelity input leads to accurate, grounded outputs, even with large contexts or complex formats

    • Broad Multimodal Support: Works across diverse sources, including Word, PPT, Excel, images, scanned docs, web pages, structured data

    • Seamless Pipeline Orchestration: Provides both Workflow and Agentic Workflow, a unified canvas for low-code and prompt-driven logic, simplifying complex orchestration

    • Deep Research Multi Agent Engine: Built-in template enabling dynamic, iterative ex-ploration of user queries across internal and external sources, using a robust agent hi-erarchy and prompt-engineered decision flows:

    2 - Ollama: Local LLM Inference

    Ollama allows you to run large language models locally with ease. It is designed for perfor-mance, portability, and low latency, making it perfect for developers and enterprises alike.

    Preinstalled and ready to go with GPU acceleration, Ollama on this VM includes the following models:

    • Deepseek-R1: family of open reasoning models

    • Qwen 2.5: High-performing general-purpose model

    • Mistral: Compact and efficient model for reasoning tasks

    • Gemma: Open, lightweight LLM by Google

    • LLaVA: Vision-language model for image + text use cases

    • LLaMA 3.3 - optimized for dialogue/chat use cases

    3 - NVIDIA GPU Support

    • Fully configured GPU-ready environment

    • Harness the power of GPU-accelerated inference to drastically reduce latency and in-crease throughput for LLM tasks

    • Works seamlessly with Ollama and Ragflow for high-speed GenAI workflows

    Use Cases

    • Deep Research Agents: Autonomously break down research tasks into sub-tasks, re-trieve across multiple sources, and synthesize executive-level reports.

    • Document Q&A & Knowledge Assistants: Tap into structured data across formats with accurate citation and transparency.

    • AI Copilots & Knowledge Workers: Leverage visual and text inputs to power multi-modal assistants.

    • Secure, Scalable RAG Applications: Everything runs within your own cloud environ-ment with full workflow control and observability.

    • Low-Latency LLM APIs: Direct deployment of Ollama LLMs for high-performance AI endpoints.

    Why Choose This VM?

    • Full Data Control & Security: Everything runs in your isolated cloud environment giv-ing you Full control over your environment and data, Ideal for sensitive workloads, in-ternal documents, and enterprise-grade compliance.

    • Flexible Model Support: Use your own embeddings, documents, and vector DBs with Ragflow. Comes with preinstalled LLMs (Deepseek-R1, Qwen 2.5, Mistral, Gemma, Llama, LLaVA) and allows you to easily add your own models via Ollama or any Other LLM provider, giving you complete control over what models you use and how you run them.

    • All-in-One: Everything you need for GenAI development in a single VM

    • Instant Setup: No need to install anything , spin up and start working

    • Multimodal Ready: Includes LLaVA for image+text inference

    Disclaimer: Other trademarks and trade names may be used in this document to refer to either the entities claiming the marks and/or names or their products and are the property of their respective owners. We disclaim proprietary interest in the marks and names of others.

    Highlights

    • Agentic RAG with Ragflow, Local Ollama LLMs, GPU-Acceleration & Secure AI Workflow Platform

    Details

    Delivery method

    Delivery option
    64-bit (x86) Amazon Machine Image (AMI)

    Latest version

    Operating system
    Ubuntu 24.04 LTS

    Deployed on AWS

    Unlock automation with AI agent solutions

    Fast-track AI initiatives with agents, tools, and solutions from AWS Partners.
    AI Agents

    Features and programs

    Financing for AWS Marketplace purchases

    AWS Marketplace now accepts line of credit payments through the PNC Vendor Finance program. This program is available to select AWS customers in the US, excluding NV, NC, ND, TN, & VT.
    Financing for AWS Marketplace purchases

    Pricing

    Instant RAGFlow: Ready-to-Use AI Knowledge Retrieval Engine

     Info
    Pricing is based on actual usage, with charges varying according to how much you consume. Subscriptions have no end date and may be canceled any time.
    Additional AWS infrastructure costs may apply. Use the AWS Pricing Calculator  to estimate your infrastructure costs.

    Usage costs (572)

     Info
    • ...
    Dimension
    Cost/hour
    t2.2xlarge
    Recommended
    $0.10
    m5ad.2xlarge
    $0.10
    m6i.12xlarge
    $0.10
    r6i.8xlarge
    $0.10
    r6in.4xlarge
    $0.10
    r5n.2xlarge
    $0.10
    i7ie.3xlarge
    $0.10
    t3.xlarge
    $0.10
    r6id.xlarge
    $0.10
    r5d.24xlarge
    $0.10

    Vendor refund policy

    Will be charged for usage, can be canceled anytime and usage fee is non refundable

    How can we make this page better?

    We'd like to hear your feedback and ideas on how to improve this page.
    We'd like to hear your feedback and ideas on how to improve this page.

    Legal

    Vendor terms and conditions

    Upon subscribing to this product, you must acknowledge and agree to the terms and conditions outlined in the vendor's End User License Agreement (EULA) .

    Content disclaimer

    Vendors are responsible for their product descriptions and other product content. AWS does not warrant that vendors' product descriptions or other product content are accurate, complete, reliable, current, or error-free.

    Usage information

     Info

    Delivery details

    64-bit (x86) Amazon Machine Image (AMI)

    Amazon Machine Image (AMI)

    An AMI is a virtual image that provides the information required to launch an instance. Amazon EC2 (Elastic Compute Cloud) instances are virtual servers on which you can run your applications and workloads, offering varying combinations of CPU, memory, storage, and networking resources. You can launch as many instances from as many different AMIs as you need.

    Version release notes

    first release

    Additional details

    Usage instructions

    1. On the EC2 Console page, instance is up and running. To connect to this instance through putty, copy the IPv4 Public IP Address. (refer Putty Guide available at https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/connect-linux-inst-from-windows.html  for details on how to connect using putty/ssh).

    2. Open putty, paste the IP address and browse your private key you downloaded while deploying the VM, by going to SSH- >Auth->Credentials , click on Open.

    3. Login as ubuntu user.

    4. Update the password of ubuntu user using below command :

      sudo passwd ubuntu

    5. Once ubuntu user password is set, access the GUI environment using RDP on Windows machine or Remmina on Linux machine.

    6. Copy the Public IP of the VM and paste it in the RDP. Login with ubuntu user and its password.

    7. To access the RAGFlow web interface , open your browser and copy paste the public IP of the VM as https://public_ip_of_vm

    8. Create your first admin account on the registration page by clicking Sign up button.

    For more details please visit - https://www.techlatest.net/support/ragflow_support/aws_gettingstartedguide 

    Support

    Vendor support

    AWS infrastructure support

    AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.

    Similar products

    Customer reviews

    Ratings and reviews

     Info
    0 ratings
    5 star
    4 star
    3 star
    2 star
    1 star
    0%
    0%
    0%
    0%
    0%
    0 AWS reviews
    No customer reviews yet
    Be the first to review this product . We've partnered with PeerSpot to gather customer feedback. You can share your experience by writing or recording a review, or scheduling a call with a PeerSpot analyst.