Listing Thumbnail

    Custom LLMs, Ready in Minutes with LLaMa Factory

     Info
    Deployed on AWS
    Preconfigured VM with LLaMA Factory for fine-tuning and deploying open-source LLMs. Includes model training tools, web UI, CLI workflows, and GPU-optimized libraries for fast experimentation.

    Overview

    Important: For step by step guide on how to setup this vm, please refer to our Getting Started guide 

    This virtual machine is configured with LLaMA Factory, a powerful framework designed to simplify large language model fine-tuning and deployment. The VM provides a fully prepared environment for experimenting, training, and serving modern LLMs without the complexity of manual setup.

    LLaMA Factory is an open-source framework built on PyTorch that unifies the workflows needed to customize and deploy large language and multimodal models. It brings together training, evaluation, optimization, and inference into a single, consistent toolkit.

    Core Capabilities:

    Extensive Model Compatibility

    • Works with a broad range of LLMs and multimodal models such as LLaMA, Qwen, Gemma, Mistral, ChatGLM, Phi, and others, enabling flexibility across use cases.

    Multiple Fine-Tuning Methods

    • Supports full parameter training, LoRA and QLoRA, partial freezing, reward modeling, and preference-based learning methods such as DPO and PPO.

    Performance-Focused Design

    • Includes modern efficiency techniques like optimized attention mechanisms and memory-saving strategies to reduce GPU usage and speed up training.

    Intuitive User Experience

    • Offers both CLI workflows and a browser-based UI where users can manage datasets, adjust training settings, and track progress visually.

    Why Choose This LLaMA Factory VM?

    • Everything is preinstalled and configured, removing the friction of dependency management and GPU setup.
    • The VM works equally well for newcomers using the graphical interface and experienced practitioners who prefer scriptable, low-level control.
    • Built for Growth: Highly Scalable for your workload with cloud infrastructure
    • Faster Time to Results: By eliminating setup overhead and providing optimized defaults, this VM helps you focus on building and refining models instead of managing infrastructure.

    Disclaimer: Other trademarks and trade names may be used in this document to refer to either the entities claiming the marks and/or names or their products and are the property of their respective owners. We disclaim proprietary interest in the marks and names of others.

    NVIDIA License disclaimer: Container image Copyright (c) 2016-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. This container image and its contents are governed by the NVIDIA Deep Learning Container License. By pulling and using the container, you accept the terms and conditions of this license: https://developer.nvidia.com/ngc/nvidia-deep-learning-container-license 

    Highlights

    • Fine-tune and deploy open-source LLMs faster with a ready-to-run LLamaFactory VM

    Details

    Categories

    Delivery method

    Delivery option
    64-bit (x86) Amazon Machine Image (AMI)

    Latest version

    Operating system
    Ubuntu 24,04 LTS

    Deployed on AWS
    New

    Introducing multi-product solutions

    You can now purchase comprehensive solutions tailored to use cases and industries.

    Multi-product solutions

    Features and programs

    Financing for AWS Marketplace purchases

    AWS Marketplace now accepts line of credit payments through the PNC Vendor Finance program. This program is available to select AWS customers in the US, excluding NV, NC, ND, TN, & VT.
    Financing for AWS Marketplace purchases

    Pricing

    Custom LLMs, Ready in Minutes with LLaMa Factory

     Info
    Pricing is based on actual usage, with charges varying according to how much you consume. Subscriptions have no end date and may be canceled any time.
    Additional AWS infrastructure costs may apply. Use the AWS Pricing Calculator  to estimate your infrastructure costs.

    Usage costs (715)

     Info
    • ...
    Dimension
    Cost/hour
    t2.2xlarge
    Recommended
    $0.13
    m7i-flex.large
    $0.13
    m7i.24xlarge
    $0.13
    m5ad.4xlarge
    $0.13
    r7a.32xlarge
    $0.13
    m4.large
    $0.13
    z1d.2xlarge
    $0.13
    i4i.12xlarge
    $0.13
    c7a.2xlarge
    $0.13
    m8i.12xlarge
    $0.13

    Vendor refund policy

    will be charged for usage, can be canceled anytime and usage fee is non refundable.

    How can we make this page better?

    We'd like to hear your feedback and ideas on how to improve this page.
    We'd like to hear your feedback and ideas on how to improve this page.

    Legal

    Vendor terms and conditions

    Upon subscribing to this product, you must acknowledge and agree to the terms and conditions outlined in the vendor's End User License Agreement (EULA) .

    Content disclaimer

    Vendors are responsible for their product descriptions and other product content. AWS does not warrant that vendors' product descriptions or other product content are accurate, complete, reliable, current, or error-free.

    Usage information

     Info

    Delivery details

    64-bit (x86) Amazon Machine Image (AMI)

    Amazon Machine Image (AMI)

    An AMI is a virtual image that provides the information required to launch an instance. Amazon EC2 (Elastic Compute Cloud) instances are virtual servers on which you can run your applications and workloads, offering varying combinations of CPU, memory, storage, and networking resources. You can launch as many instances from as many different AMIs as you need.

    Version release notes

    Llama factory v0.9.4 on ubuntu 24.04

    Additional details

    Usage instructions

    1. On the EC2 Console page, instance is up and running. To connect to this instance through putty, copy the IPv4 Public IP Address (refer Putty Guide available at https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/connect-linux-inst-from-windows.html  for details on how to connect using putty/ssh).

    2. Open putty, paste the IP address and browse your private key you downloaded while deploying the VM, by going to SSH->Auth->Credentials, click on Open.

    3. login as ubuntu.

    4. Set the password of Ubuntu user using: sudo passwd ubuntu

    5. Open Remote Desktop client from your Windows machine (or Remmina if you are on Linux system), copy paste the public ip of the VM. Login with same "ubuntu" user and password set in above step.

    6. To access the Milvus WebUI, open the Firefox browser from your RDP session as explained in above step 5. enter the URL as http://localhost:9091/webui/ 

    7. To access the jupyterhub in your local browser (not in RDP), copy paste the public IP of the VM as "https://public_ip_of_vm". Login with same "ubuntu" user and its password.

    For more details please visit - https://techlatest.net/support/llama_factory_support/aws_gettingstartedguide/index.html 

    Support

    Vendor support

    AWS infrastructure support

    AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.

    Similar products

    Customer reviews

    Ratings and reviews

     Info
    0 ratings
    5 star
    4 star
    3 star
    2 star
    1 star
    0%
    0%
    0%
    0%
    0%
    0 reviews
    No customer reviews yet
    Be the first to review this product . We've partnered with PeerSpot to gather customer feedback. You can share your experience by writing or recording a review, or scheduling a call with a PeerSpot analyst.