<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
    <channel>
        <title>ChatGLM on Producthunt daily</title>
        <link>https://producthunt.programnotes.cn/en/tags/chatglm/</link>
        <description>Recent content in ChatGLM on Producthunt daily</description>
        <generator>Hugo -- gohugo.io</generator>
        <language>en</language>
        <lastBuildDate>Tue, 27 May 2025 15:31:11 +0800</lastBuildDate><atom:link href="https://producthunt.programnotes.cn/en/tags/chatglm/index.xml" rel="self" type="application/rss+xml" /><item>
        <title>LLaMA-Factory</title>
        <link>https://producthunt.programnotes.cn/en/p/llama-factory/</link>
        <pubDate>Tue, 27 May 2025 15:31:11 +0800</pubDate>
        
        <guid>https://producthunt.programnotes.cn/en/p/llama-factory/</guid>
        <description>&lt;img src="https://images.unsplash.com/photo-1680153527310-1a70b47af6e9?ixid=M3w0NjAwMjJ8MHwxfHJhbmRvbXx8fHx8fHx8fDE3NDgzMzA5MjJ8&amp;ixlib=rb-4.1.0" alt="Featured image of post LLaMA-Factory" /&gt;&lt;h1 id=&#34;hiyougallama-factory&#34;&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;hiyouga/LLaMA-Factory&lt;/a&gt;
&lt;/h1&gt;&lt;p&gt;&lt;img src=&#34;https://producthunt.programnotes.cn/assets/logo.png&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;# LLaMA Factory&#34;
	
	
&gt;&lt;/p&gt;
&lt;p&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory/stargazers&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://img.shields.io/github/stars/hiyouga/LLaMA-Factory?style=social&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;GitHub Repo stars&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory/commits/main&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://img.shields.io/github/last-commit/hiyouga/LLaMA-Factory&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;GitHub last commit&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory/graphs/contributors&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://img.shields.io/github/contributors/hiyouga/LLaMA-Factory?color=orange&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;GitHub contributors&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory/actions/workflows/tests.yml&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://github.com/hiyouga/LLaMA-Factory/actions/workflows/tests.yml/badge.svg&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;GitHub workflow&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://pypi.org/project/llamafactory/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://img.shields.io/pypi/v/llamafactory&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;PyPI&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://scholar.google.com/scholar?cites=12620864006390196564&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://img.shields.io/badge/citation-476-green&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;Citation&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory/pulls&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://img.shields.io/badge/PRs-welcome-blue&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;GitHub pull request&#34;
	
	
&gt;&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;a class=&#34;link&#34; href=&#34;https://twitter.com/llamafactory_ai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://img.shields.io/twitter/follow/llamafactory_ai&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;Twitter&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://discord.gg/rKfvV9r9FK&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://dcbadge.vercel.app/api/server/rKfvV9r9FK?compact=true&amp;amp;style=flat&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;Discord&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://gitcode.com/zhengyaowei/LLaMA-Factory&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://gitcode.com/zhengyaowei/LLaMA-Factory/star/badge.svg&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;GitCode&#34;
	
	
&gt;&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&lt;a class=&#34;link&#34; href=&#34;https://colab.research.google.com/drive/1eRTPn37ltBbYsISy9Aw2NuI2Aq5CQrD9?usp=sharing&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://colab.research.google.com/assets/colab-badge.svg&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;Open in Colab&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://gallery.pai-ml.com/#/preview/deepLearning/nlp/llama_factory&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://gallery.pai-ml.com/assets/open-in-dsw.svg&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;Open in DSW&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/spaces/hiyouga/LLaMA-Board&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://img.shields.io/badge/%f0%9f%a4%97-Open%20in%20Spaces-blue&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;Spaces&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://modelscope.cn/studios/hiyouga/LLaMA-Board&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://img.shields.io/badge/ModelScope-Open%20in%20Studios-blue&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;Studios&#34;
	
	
&gt;&lt;/a&gt;
&lt;a class=&#34;link&#34; href=&#34;https://aws.amazon.com/cn/blogs/machine-learning/how-apoidea-group-enhances-visual-information-extraction-from-banking-documents-with-multimodal-models-using-llama-factory-on-amazon-sagemaker-hyperpod/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;&lt;img src=&#34;https://img.shields.io/badge/SageMaker-Open%20in%20AWS-blue&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;SageMaker&#34;
	
	
&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3 id=&#34;used-by-amazon-nvidia-aliyun-etc&#34;&gt;Used by &lt;a class=&#34;link&#34; href=&#34;https://aws.amazon.com/cn/blogs/machine-learning/how-apoidea-group-enhances-visual-information-extraction-from-banking-documents-with-multimodal-models-using-llama-factory-on-amazon-sagemaker-hyperpod/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Amazon&lt;/a&gt;, &lt;a class=&#34;link&#34; href=&#34;https://developer.nvidia.com/rtx/ai-toolkit&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;NVIDIA&lt;/a&gt;, &lt;a class=&#34;link&#34; href=&#34;https://help.aliyun.com/zh/pai/use-cases/fine-tune-a-llama-3-model-with-llama-factory&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Aliyun&lt;/a&gt;, etc.
&lt;/h3&gt;&lt;div align=&#34;center&#34; markdown=&#34;1&#34;&gt;
&lt;h3 id=&#34;supporters-&#34;&gt;Supporters ❤️
&lt;/h3&gt;&lt;a href=&#34;https://warp.dev/llama-factory&#34;&gt;
    &lt;img alt=&#34;Warp sponsorship&#34; width=&#34;400&#34; src=&#34;https://github.com/user-attachments/assets/ab8dd143-b0fd-4904-bdc5-dd7ecac94eae&#34;&gt;
&lt;/a&gt;
&lt;h4 id=&#34;warp-the-agentic-terminal-for-developers&#34;&gt;&lt;a class=&#34;link&#34; href=&#34;https://warp.dev/llama-factory&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Warp, the agentic terminal for developers&lt;/a&gt;
&lt;/h4&gt;&lt;p&gt;&lt;a class=&#34;link&#34; href=&#34;https://warp.dev/llama-factory&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Available for MacOS, Linux, &amp;amp; Windows&lt;/a&gt;&lt;/p&gt;
&lt;hr&gt;
&lt;h3 id=&#34;easily-fine-tune-100-large-language-models-with-zero-code-cli-and-web-ui&#34;&gt;Easily fine-tune 100+ large language models with zero-code &lt;a class=&#34;link&#34; href=&#34;#quickstart&#34; &gt;CLI&lt;/a&gt; and &lt;a class=&#34;link&#34; href=&#34;#fine-tuning-with-llama-board-gui-powered-by-gradio&#34; &gt;Web UI&lt;/a&gt;
&lt;/h3&gt;&lt;p&gt;&lt;img src=&#34;https://trendshift.io/api/badge/repositories/4535&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;GitHub Trend&#34;
	
	
&gt;&lt;/p&gt;
&lt;/div&gt;
&lt;p&gt;👋 Join our &lt;a class=&#34;link&#34; href=&#34;assets/wechat.jpg&#34; &gt;WeChat&lt;/a&gt; or &lt;a class=&#34;link&#34; href=&#34;assets/wechat_npu.jpg&#34; &gt;NPU user group&lt;/a&gt;.&lt;/p&gt;
\[ English | [中文](README_zh.md) \]&lt;p&gt;&lt;strong&gt;Fine-tuning a large language model can be easy as&amp;hellip;&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/user-attachments/assets/3991a3a8-4276-4d30-9cab-4cb0c4b9b99e&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;https://github.com/user-attachments/assets/3991a3a8-4276-4d30-9cab-4cb0c4b9b99e&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;Choose your path:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Documentation&lt;/strong&gt;: &lt;a class=&#34;link&#34; href=&#34;https://llamafactory.readthedocs.io/en/latest/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;https://llamafactory.readthedocs.io/en/latest/&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Colab (free)&lt;/strong&gt;: &lt;a class=&#34;link&#34; href=&#34;https://colab.research.google.com/drive/1eRTPn37ltBbYsISy9Aw2NuI2Aq5CQrD9?usp=sharing&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;https://colab.research.google.com/drive/1eRTPn37ltBbYsISy9Aw2NuI2Aq5CQrD9?usp=sharing&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Local machine&lt;/strong&gt;: Please refer to &lt;a class=&#34;link&#34; href=&#34;#getting-started&#34; &gt;usage&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;PAI-DSW (free trial)&lt;/strong&gt;: &lt;a class=&#34;link&#34; href=&#34;https://gallery.pai-ml.com/#/preview/deepLearning/nlp/llama_factory&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;https://gallery.pai-ml.com/#/preview/deepLearning/nlp/llama_factory&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;blockquote&gt;
&lt;p&gt;[!NOTE]
Except for the above links, all other websites are unauthorized third-party websites. Please carefully use them.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;h2 id=&#34;table-of-contents&#34;&gt;Table of Contents
&lt;/h2&gt;&lt;ul&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#features&#34; &gt;Features&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#blogs&#34; &gt;Blogs&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#changelog&#34; &gt;Changelog&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#supported-models&#34; &gt;Supported Models&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#supported-training-approaches&#34; &gt;Supported Training Approaches&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#provided-datasets&#34; &gt;Provided Datasets&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#requirement&#34; &gt;Requirement&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#getting-started&#34; &gt;Getting Started&lt;/a&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#installation&#34; &gt;Installation&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#data-preparation&#34; &gt;Data Preparation&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#quickstart&#34; &gt;Quickstart&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#fine-tuning-with-llama-board-gui-powered-by-gradio&#34; &gt;Fine-Tuning with LLaMA Board GUI&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#build-docker&#34; &gt;Build Docker&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#deploy-with-openai-style-api-and-vllm&#34; &gt;Deploy with OpenAI-style API and vLLM&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#download-from-modelscope-hub&#34; &gt;Download from ModelScope Hub&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#download-from-modelers-hub&#34; &gt;Download from Modelers Hub&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#use-wb-logger&#34; &gt;Use W&amp;amp;B Logger&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#use-swanlab-logger&#34; &gt;Use SwanLab Logger&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#projects-using-llama-factory&#34; &gt;Projects using LLaMA Factory&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#license&#34; &gt;License&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#citation&#34; &gt;Citation&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;#acknowledgement&#34; &gt;Acknowledgement&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;features&#34;&gt;Features
&lt;/h2&gt;&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Various models&lt;/strong&gt;: LLaMA, LLaVA, Mistral, Mixtral-MoE, Qwen, Qwen2-VL, DeepSeek, Yi, Gemma, ChatGLM, Phi, etc.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Integrated methods&lt;/strong&gt;: (Continuous) pre-training, (multimodal) supervised fine-tuning, reward modeling, PPO, DPO, KTO, ORPO, etc.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Scalable resources&lt;/strong&gt;: 16-bit full-tuning, freeze-tuning, LoRA and 2/3/4/5/6/8-bit QLoRA via AQLM/AWQ/GPTQ/LLM.int8/HQQ/EETQ.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Advanced algorithms&lt;/strong&gt;: &lt;a class=&#34;link&#34; href=&#34;https://github.com/jiaweizzhao/GaLore&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;GaLore&lt;/a&gt;, &lt;a class=&#34;link&#34; href=&#34;https://github.com/Ledzy/BAdam&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;BAdam&lt;/a&gt;, &lt;a class=&#34;link&#34; href=&#34;https://github.com/zhuhanqing/APOLLO&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;APOLLO&lt;/a&gt;, &lt;a class=&#34;link&#34; href=&#34;https://github.com/zyushun/Adam-mini&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Adam-mini&lt;/a&gt;, &lt;a class=&#34;link&#34; href=&#34;https://github.com/KellerJordan/Muon&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Muon&lt;/a&gt;, DoRA, LongLoRA, LLaMA Pro, Mixture-of-Depths, LoRA+, LoftQ and PiSSA.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Practical tricks&lt;/strong&gt;: &lt;a class=&#34;link&#34; href=&#34;https://github.com/Dao-AILab/flash-attention&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;FlashAttention-2&lt;/a&gt;, &lt;a class=&#34;link&#34; href=&#34;https://github.com/unslothai/unsloth&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Unsloth&lt;/a&gt;, &lt;a class=&#34;link&#34; href=&#34;https://github.com/linkedin/Liger-Kernel&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Liger Kernel&lt;/a&gt;, RoPE scaling, NEFTune and rsLoRA.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Wide tasks&lt;/strong&gt;: Multi-turn dialogue, tool using, image understanding, visual grounding, video recognition, audio understanding, etc.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Experiment monitors&lt;/strong&gt;: LlamaBoard, TensorBoard, Wandb, MLflow, &lt;a class=&#34;link&#34; href=&#34;https://github.com/SwanHubX/SwanLab&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;SwanLab&lt;/a&gt;, etc.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Faster inference&lt;/strong&gt;: OpenAI-style API, Gradio UI and CLI with &lt;a class=&#34;link&#34; href=&#34;https://github.com/vllm-project/vllm&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;vLLM worker&lt;/a&gt; or &lt;a class=&#34;link&#34; href=&#34;https://github.com/sgl-project/sglang&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;SGLang worker&lt;/a&gt;.&lt;/li&gt;
&lt;/ul&gt;
&lt;h3 id=&#34;day-n-support-for-fine-tuning-cutting-edge-models&#34;&gt;Day-N Support for Fine-Tuning Cutting-Edge Models
&lt;/h3&gt;&lt;table&gt;
  &lt;thead&gt;
      &lt;tr&gt;
          &lt;th&gt;Support Date&lt;/th&gt;
          &lt;th&gt;Model Name&lt;/th&gt;
      &lt;/tr&gt;
  &lt;/thead&gt;
  &lt;tbody&gt;
      &lt;tr&gt;
          &lt;td&gt;Day 0&lt;/td&gt;
          &lt;td&gt;Qwen3 / Qwen2.5-VL / Gemma 3 / InternLM 3 / MiniCPM-o-2.6&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;Day 1&lt;/td&gt;
          &lt;td&gt;Llama 3 / GLM-4 / Mistral Small / PaliGemma2 / Llama 4&lt;/td&gt;
      &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;
&lt;h2 id=&#34;blogs&#34;&gt;Blogs
&lt;/h2&gt;&lt;ul&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://aws.amazon.com/cn/blogs/machine-learning/how-apoidea-group-enhances-visual-information-extraction-from-banking-documents-with-multimodal-models-using-llama-factory-on-amazon-sagemaker-hyperpod/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;How Apoidea Group enhances visual information extraction from banking documents with multimodal models using LLaMA-Factory on Amazon SageMaker HyperPod&lt;/a&gt; (English)&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://buaa-act.feishu.cn/wiki/GVzlwYcRFiR8OLkHbL6cQpYin7g&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Easy Dataset × LLaMA Factory: Enabling LLMs to Efficiently Learn Domain Knowledge&lt;/a&gt; (English)&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://gallery.pai-ml.com/#/preview/deepLearning/nlp/llama_factory_deepseek_r1_distill_7b&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LLaMA Factory: Fine-tuning the DeepSeek-R1-Distill-Qwen-7B Model for News Classifier&lt;/a&gt; (Chinese)&lt;/li&gt;
&lt;/ul&gt;
&lt;details&gt;&lt;summary&gt;All Blogs&lt;/summary&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://aws.amazon.com/cn/blogs/china/a-one-stop-code-free-model-fine-tuning-deployment-platform-based-on-sagemaker-and-llama-factory/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;A One-Stop Code-Free Model Fine-Tuning &amp;amp; Deployment Platform based on SageMaker and LLaMA-Factory&lt;/a&gt; (Chinese)&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://gallery.pai-ml.com/#/preview/deepLearning/nlp/llama_factory_qwen2vl&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LLaMA Factory Multi-Modal Fine-Tuning Practice: Fine-Tuning Qwen2-VL for Personal Tourist Guide&lt;/a&gt; (Chinese)&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://gallery.pai-ml.com/#/preview/deepLearning/nlp/llama_factory&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LLaMA Factory: Fine-tuning the LLaMA3 Model for Role-Playing&lt;/a&gt; (Chinese)&lt;/li&gt;
&lt;/ul&gt;
&lt;/details&gt;
&lt;h2 id=&#34;changelog&#34;&gt;Changelog
&lt;/h2&gt;&lt;p&gt;[25/04/28] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://qwenlm.github.io/blog/qwen3/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen3&lt;/a&gt;&lt;/strong&gt; model family.&lt;/p&gt;
&lt;p&gt;[25/04/21] We supported the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/KellerJordan/Muon&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Muon&lt;/a&gt;&lt;/strong&gt; optimizer. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage. Thank &lt;a class=&#34;link&#34; href=&#34;https://github.com/tianshijing&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;@tianshijing&lt;/a&gt;&amp;rsquo;s PR.&lt;/p&gt;
&lt;p&gt;[25/04/16] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/OpenGVLab/InternVL3-8B&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;InternVL3&lt;/a&gt;&lt;/strong&gt; model. See &lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory/pull/7258&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;PR #7258&lt;/a&gt; to get started.&lt;/p&gt;
&lt;p&gt;[25/04/14] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/THUDM/GLM-Z1-9B-0414&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;GLM-Z1&lt;/a&gt;&lt;/strong&gt; and &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/moonshotai/Kimi-VL-A3B-Instruct&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Kimi-VL&lt;/a&gt;&lt;/strong&gt; models.&lt;/p&gt;
&lt;p&gt;[25/04/06] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://ai.meta.com/blog/llama-4-multimodal-intelligence/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama 4&lt;/a&gt;&lt;/strong&gt; model. See &lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory/pull/7611&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;PR #7611&lt;/a&gt; to get started.&lt;/p&gt;
&lt;details&gt;&lt;summary&gt;Full Changelog&lt;/summary&gt;
&lt;p&gt;[25/03/31] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://qwenlm.github.io/blog/qwen2.5-omni/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen2.5 Omni&lt;/a&gt;&lt;/strong&gt; model. See &lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory/pull/7537&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;PR #7537&lt;/a&gt; to get started.&lt;/p&gt;
&lt;p&gt;[25/03/15] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/sgl-project/sglang&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;SGLang&lt;/a&gt;&lt;/strong&gt; as inference backend. Try &lt;code&gt;infer_backend: sglang&lt;/code&gt; to accelerate inference.&lt;/p&gt;
&lt;p&gt;[25/03/12] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/blog/gemma3&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Gemma 3&lt;/a&gt;&lt;/strong&gt; model.&lt;/p&gt;
&lt;p&gt;[25/02/24] Announcing &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/EasyR1&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;EasyR1&lt;/a&gt;&lt;/strong&gt;, an efficient, scalable and multi-modality RL training framework for efficient GRPO training.&lt;/p&gt;
&lt;p&gt;[25/02/11] We supported saving the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/ollama/ollama&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Ollama&lt;/a&gt;&lt;/strong&gt; modelfile when exporting the model checkpoints. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[25/02/05] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;Qwen/Qwen2-Audio-7B-Instruct&#34; &gt;Qwen2-Audio&lt;/a&gt;&lt;/strong&gt; and &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/openbmb/MiniCPM-o-2_6&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;MiniCPM-o-2.6&lt;/a&gt;&lt;/strong&gt; on audio understanding tasks.&lt;/p&gt;
&lt;p&gt;[25/01/31] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/deepseek-ai/DeepSeek-R1&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;DeepSeek-R1&lt;/a&gt;&lt;/strong&gt; and &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen2.5-VL&lt;/a&gt;&lt;/strong&gt; models.&lt;/p&gt;
&lt;p&gt;[25/01/15] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2412.05270&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;APOLLO&lt;/a&gt;&lt;/strong&gt; optimizer. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[25/01/14] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/openbmb/MiniCPM-o-2_6&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;MiniCPM-o-2.6&lt;/a&gt;&lt;/strong&gt; and &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/openbmb/MiniCPM-V-2_6&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;MiniCPM-V-2.6&lt;/a&gt;&lt;/strong&gt; models. Thank &lt;a class=&#34;link&#34; href=&#34;https://github.com/BUAADreamer&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;@BUAADreamer&lt;/a&gt;&amp;rsquo;s PR.&lt;/p&gt;
&lt;p&gt;[25/01/14] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/collections/internlm/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;InternLM 3&lt;/a&gt;&lt;/strong&gt; models. Thank &lt;a class=&#34;link&#34; href=&#34;https://github.com/hhaAndroid&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;@hhaAndroid&lt;/a&gt;&amp;rsquo;s PR.&lt;/p&gt;
&lt;p&gt;[25/01/10] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/microsoft/phi-4&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Phi-4&lt;/a&gt;&lt;/strong&gt; model.&lt;/p&gt;
&lt;p&gt;[24/12/21] We supported using &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/SwanHubX/SwanLab&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;SwanLab&lt;/a&gt;&lt;/strong&gt; for experiment tracking and visualization. See &lt;a class=&#34;link&#34; href=&#34;#use-swanlab-logger&#34; &gt;this section&lt;/a&gt; for details.&lt;/p&gt;
&lt;p&gt;[24/11/27] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Skywork/Skywork-o1-Open-Llama-3.1-8B&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Skywork-o1&lt;/a&gt;&lt;/strong&gt; model and the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/O1-OPEN/OpenO1-SFT&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;OpenO1&lt;/a&gt;&lt;/strong&gt; dataset.&lt;/p&gt;
&lt;p&gt;[24/10/09] We supported downloading pre-trained models and datasets from the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://modelers.cn/models&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Modelers Hub&lt;/a&gt;&lt;/strong&gt;. See &lt;a class=&#34;link&#34; href=&#34;#download-from-modelers-hub&#34; &gt;this tutorial&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/09/19] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://qwenlm.github.io/blog/qwen2.5/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen2.5&lt;/a&gt;&lt;/strong&gt; models.&lt;/p&gt;
&lt;p&gt;[24/08/30] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://qwenlm.github.io/blog/qwen2-vl/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen2-VL&lt;/a&gt;&lt;/strong&gt; models. Thank &lt;a class=&#34;link&#34; href=&#34;https://github.com/simonJJJ&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;@simonJJJ&lt;/a&gt;&amp;rsquo;s PR.&lt;/p&gt;
&lt;p&gt;[24/08/27] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/linkedin/Liger-Kernel&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Liger Kernel&lt;/a&gt;&lt;/strong&gt;. Try &lt;code&gt;enable_liger_kernel: true&lt;/code&gt; for efficient training.&lt;/p&gt;
&lt;p&gt;[24/08/09] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/zyushun/Adam-mini&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Adam-mini&lt;/a&gt;&lt;/strong&gt; optimizer. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage. Thank &lt;a class=&#34;link&#34; href=&#34;https://github.com/relic-yuexi&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;@relic-yuexi&lt;/a&gt;&amp;rsquo;s PR.&lt;/p&gt;
&lt;p&gt;[24/07/04] We supported &lt;a class=&#34;link&#34; href=&#34;https://github.com/MeetKai/functionary/tree/main/functionary/train/packing&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;contamination-free packed training&lt;/a&gt;. Use &lt;code&gt;neat_packing: true&lt;/code&gt; to activate it. Thank &lt;a class=&#34;link&#34; href=&#34;https://github.com/chuan298&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;@chuan298&lt;/a&gt;&amp;rsquo;s PR.&lt;/p&gt;
&lt;p&gt;[24/06/16] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.02948&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;PiSSA&lt;/a&gt;&lt;/strong&gt; algorithm. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/06/07] We supported fine-tuning the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://qwenlm.github.io/blog/qwen2/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen2&lt;/a&gt;&lt;/strong&gt; and &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/THUDM/GLM-4&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;GLM-4&lt;/a&gt;&lt;/strong&gt; models.&lt;/p&gt;
&lt;p&gt;[24/05/26] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2405.14734&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;SimPO&lt;/a&gt;&lt;/strong&gt; algorithm for preference learning. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/05/20] We supported fine-tuning the &lt;strong&gt;PaliGemma&lt;/strong&gt; series models. Note that the PaliGemma models are pre-trained models, you need to fine-tune them with &lt;code&gt;paligemma&lt;/code&gt; template for chat completion.&lt;/p&gt;
&lt;p&gt;[24/05/18] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.01306&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;KTO&lt;/a&gt;&lt;/strong&gt; algorithm for preference learning. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/05/14] We supported training and inference on the Ascend NPU devices. Check &lt;a class=&#34;link&#34; href=&#34;#installation&#34; &gt;installation&lt;/a&gt; section for details.&lt;/p&gt;
&lt;p&gt;[24/04/26] We supported fine-tuning the &lt;strong&gt;LLaVA-1.5&lt;/strong&gt; multimodal LLMs. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/04/22] We provided a &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://colab.research.google.com/drive/1eRTPn37ltBbYsISy9Aw2NuI2Aq5CQrD9?usp=sharing&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Colab notebook&lt;/a&gt;&lt;/strong&gt; for fine-tuning the Llama-3 model on a free T4 GPU. Two Llama-3-derived models fine-tuned using LLaMA Factory are available at Hugging Face, check &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/shenzhi-wang/Llama3-8B-Chinese-Chat&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama3-8B-Chinese-Chat&lt;/a&gt; and &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/zhichen/Llama3-Chinese&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama3-Chinese&lt;/a&gt; for details.&lt;/p&gt;
&lt;p&gt;[24/04/21] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.02258&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Mixture-of-Depths&lt;/a&gt;&lt;/strong&gt; according to &lt;a class=&#34;link&#34; href=&#34;https://github.com/astramind-ai/Mixture-of-depths&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;AstraMindAI&amp;rsquo;s implementation&lt;/a&gt;. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/04/16] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.02827&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;BAdam&lt;/a&gt;&lt;/strong&gt; optimizer. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/04/16] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/unslothai/unsloth&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;unsloth&lt;/a&gt;&lt;/strong&gt;&amp;rsquo;s long-sequence training (Llama-2-7B-56k within 24GB). It achieves &lt;strong&gt;117%&lt;/strong&gt; speed and &lt;strong&gt;50%&lt;/strong&gt; memory compared with FlashAttention-2, more benchmarks can be found in &lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory/wiki/Performance-comparison&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;this page&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;[24/03/31] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2403.07691&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;ORPO&lt;/a&gt;&lt;/strong&gt;. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/03/21] Our paper &amp;ldquo;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2403.13372&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LlamaFactory: Unified Efficient Fine-Tuning of 100+ Language Models&lt;/a&gt;&amp;rdquo; is available at arXiv!&lt;/p&gt;
&lt;p&gt;[24/03/20] We supported &lt;strong&gt;FSDP+QLoRA&lt;/strong&gt; that fine-tunes a 70B model on 2x24GB GPUs. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/03/13] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.12354&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LoRA+&lt;/a&gt;&lt;/strong&gt;. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/03/07] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2403.03507&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;GaLore&lt;/a&gt;&lt;/strong&gt; optimizer. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/03/07] We integrated &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/vllm-project/vllm&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;vLLM&lt;/a&gt;&lt;/strong&gt; for faster and concurrent inference. Try &lt;code&gt;infer_backend: vllm&lt;/code&gt; to enjoy &lt;strong&gt;270%&lt;/strong&gt; inference speed.&lt;/p&gt;
&lt;p&gt;[24/02/28] We supported weight-decomposed LoRA (&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.09353&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;DoRA&lt;/a&gt;&lt;/strong&gt;). Try &lt;code&gt;use_dora: true&lt;/code&gt; to activate DoRA training.&lt;/p&gt;
&lt;p&gt;[24/02/15] We supported &lt;strong&gt;block expansion&lt;/strong&gt; proposed by &lt;a class=&#34;link&#34; href=&#34;https://github.com/TencentARC/LLaMA-Pro&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LLaMA Pro&lt;/a&gt;. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[24/02/05] Qwen1.5 (Qwen2 beta version) series models are supported in LLaMA-Factory. Check this &lt;a class=&#34;link&#34; href=&#34;https://qwenlm.github.io/blog/qwen1.5/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;blog post&lt;/a&gt; for details.&lt;/p&gt;
&lt;p&gt;[24/01/18] We supported &lt;strong&gt;agent tuning&lt;/strong&gt; for most models, equipping model with tool using abilities by fine-tuning with &lt;code&gt;dataset: glaive_toolcall_en&lt;/code&gt;.&lt;/p&gt;
&lt;p&gt;[23/12/23] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/unslothai/unsloth&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;unsloth&lt;/a&gt;&lt;/strong&gt;&amp;rsquo;s implementation to boost LoRA tuning for the LLaMA, Mistral and Yi models. Try &lt;code&gt;use_unsloth: true&lt;/code&gt; argument to activate unsloth patch. It achieves &lt;strong&gt;170%&lt;/strong&gt; speed in our benchmark, check &lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory/wiki/Performance-comparison&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;this page&lt;/a&gt; for details.&lt;/p&gt;
&lt;p&gt;[23/12/12] We supported fine-tuning the latest MoE model &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/mistralai/Mixtral-8x7B-v0.1&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Mixtral 8x7B&lt;/a&gt;&lt;/strong&gt; in our framework. See hardware requirement &lt;a class=&#34;link&#34; href=&#34;#hardware-requirement&#34; &gt;here&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;[23/12/01] We supported downloading pre-trained models and datasets from the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://modelscope.cn/models&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;ModelScope Hub&lt;/a&gt;&lt;/strong&gt;. See &lt;a class=&#34;link&#34; href=&#34;#download-from-modelscope-hub&#34; &gt;this tutorial&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[23/10/21] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2310.05914&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;NEFTune&lt;/a&gt;&lt;/strong&gt; trick for fine-tuning. Try &lt;code&gt;neftune_noise_alpha: 5&lt;/code&gt; argument to activate NEFTune.&lt;/p&gt;
&lt;p&gt;[23/09/27] We supported &lt;strong&gt;$S^2$-Attn&lt;/strong&gt; proposed by &lt;a class=&#34;link&#34; href=&#34;https://github.com/dvlab-research/LongLoRA&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LongLoRA&lt;/a&gt; for the LLaMA models. Try &lt;code&gt;shift_attn: true&lt;/code&gt; argument to enable shift short attention.&lt;/p&gt;
&lt;p&gt;[23/09/23] We integrated MMLU, C-Eval and CMMLU benchmarks in this repo. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[23/09/10] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/Dao-AILab/flash-attention&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;FlashAttention-2&lt;/a&gt;&lt;/strong&gt;. Try &lt;code&gt;flash_attn: fa2&lt;/code&gt; argument to enable FlashAttention-2 if you are using RTX4090, A100 or H100 GPUs.&lt;/p&gt;
&lt;p&gt;[23/08/12] We supported &lt;strong&gt;RoPE scaling&lt;/strong&gt; to extend the context length of the LLaMA models. Try &lt;code&gt;rope_scaling: linear&lt;/code&gt; argument in training and &lt;code&gt;rope_scaling: dynamic&lt;/code&gt; argument at inference to extrapolate the position embeddings.&lt;/p&gt;
&lt;p&gt;[23/08/11] We supported &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2305.18290&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;DPO training&lt;/a&gt;&lt;/strong&gt; for instruction-tuned models. See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;p&gt;[23/07/31] We supported &lt;strong&gt;dataset streaming&lt;/strong&gt;. Try &lt;code&gt;streaming: true&lt;/code&gt; and &lt;code&gt;max_steps: 10000&lt;/code&gt; arguments to load your dataset in streaming mode.&lt;/p&gt;
&lt;p&gt;[23/07/29] We released two instruction-tuned 13B models at Hugging Face. See these Hugging Face Repos (&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/hiyouga/Llama-2-Chinese-13b-chat&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LLaMA-2&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/hiyouga/Baichuan-13B-sft&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Baichuan&lt;/a&gt;) for details.&lt;/p&gt;
&lt;p&gt;[23/07/18] We developed an &lt;strong&gt;all-in-one Web UI&lt;/strong&gt; for training, evaluation and inference. Try &lt;code&gt;train_web.py&lt;/code&gt; to fine-tune models in your Web browser. Thank &lt;a class=&#34;link&#34; href=&#34;https://github.com/KanadeSiina&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;@KanadeSiina&lt;/a&gt; and &lt;a class=&#34;link&#34; href=&#34;https://github.com/codemayq&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;@codemayq&lt;/a&gt; for their efforts in the development.&lt;/p&gt;
&lt;p&gt;[23/07/09] We released &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/FastEdit&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;FastEdit&lt;/a&gt;&lt;/strong&gt; ⚡🩹, an easy-to-use package for editing the factual knowledge of large language models efficiently. Please follow &lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/FastEdit&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;FastEdit&lt;/a&gt; if you are interested.&lt;/p&gt;
&lt;p&gt;[23/06/29] We provided a &lt;strong&gt;reproducible example&lt;/strong&gt; of training a chat model using instruction-following datasets, see &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/hiyouga/Baichuan-7B-sft&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Baichuan-7B-sft&lt;/a&gt; for details.&lt;/p&gt;
&lt;p&gt;[23/06/22] We aligned the &lt;a class=&#34;link&#34; href=&#34;src/api_demo.py&#34; &gt;demo API&lt;/a&gt; with the &lt;a class=&#34;link&#34; href=&#34;https://platform.openai.com/docs/api-reference/chat&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;OpenAI&amp;rsquo;s&lt;/a&gt; format where you can insert the fine-tuned model in &lt;strong&gt;arbitrary ChatGPT-based applications&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;[23/06/03] We supported quantized training and inference (aka &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/artidoro/qlora&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;QLoRA&lt;/a&gt;&lt;/strong&gt;). See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples&lt;/a&gt; for usage.&lt;/p&gt;
&lt;/details&gt;
&lt;blockquote&gt;
&lt;p&gt;[!TIP]
If you cannot use the latest feature, please pull the latest code and install LLaMA-Factory again.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;h2 id=&#34;supported-models&#34;&gt;Supported Models
&lt;/h2&gt;&lt;table&gt;
  &lt;thead&gt;
      &lt;tr&gt;
          &lt;th&gt;Model&lt;/th&gt;
          &lt;th&gt;Model size&lt;/th&gt;
          &lt;th&gt;Template&lt;/th&gt;
      &lt;/tr&gt;
  &lt;/thead&gt;
  &lt;tbody&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/baichuan-inc&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Baichuan 2&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B/13B&lt;/td&gt;
          &lt;td&gt;baichuan2&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/bigscience&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;BLOOM/BLOOMZ&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;560M/1.1B/1.7B/3B/7.1B/176B&lt;/td&gt;
          &lt;td&gt;-&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/THUDM&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;ChatGLM3&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;6B&lt;/td&gt;
          &lt;td&gt;chatglm3&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/CohereForAI&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Command R&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;35B/104B&lt;/td&gt;
          &lt;td&gt;cohere&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/deepseek-ai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;DeepSeek (Code/MoE)&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B/16B/67B/236B&lt;/td&gt;
          &lt;td&gt;deepseek&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/deepseek-ai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;DeepSeek 2.5/3&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;236B/671B&lt;/td&gt;
          &lt;td&gt;deepseek3&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/deepseek-ai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;DeepSeek R1 (Distill)&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;1.5B/7B/8B/14B/32B/70B/671B&lt;/td&gt;
          &lt;td&gt;deepseekr1&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/tiiuae&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Falcon&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B/11B/40B/180B&lt;/td&gt;
          &lt;td&gt;falcon&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/google&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Gemma/Gemma 2/CodeGemma&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;2B/7B/9B/27B&lt;/td&gt;
          &lt;td&gt;gemma&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/google&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Gemma 3&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;1B/4B/12B/27B&lt;/td&gt;
          &lt;td&gt;gemma3/gemma (1B)&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/THUDM&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;GLM-4/GLM-4-0414/GLM-Z1&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;9B/32B&lt;/td&gt;
          &lt;td&gt;glm4/glmz1&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/openai-community&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;GPT-2&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;0.1B/0.4B/0.8B/1.5B&lt;/td&gt;
          &lt;td&gt;-&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/ibm-granite&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Granite 3.0-3.3&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;1B/2B/3B/8B&lt;/td&gt;
          &lt;td&gt;granite3&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/tencent/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Hunyuan&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B&lt;/td&gt;
          &lt;td&gt;hunyuan&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/IndexTeam&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Index&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;1.9B&lt;/td&gt;
          &lt;td&gt;index&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/internlm&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;InternLM 2-3&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B/8B/20B&lt;/td&gt;
          &lt;td&gt;intern2&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/OpenGVLab&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;InternVL 2.5-3&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;1B/2B/8B/14B/38B/78B&lt;/td&gt;
          &lt;td&gt;intern_vl&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/moonshotai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Kimi-VL&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;16B&lt;/td&gt;
          &lt;td&gt;kimi_vl&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/facebookresearch/llama&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B/13B/33B/65B&lt;/td&gt;
          &lt;td&gt;-&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/meta-llama&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama 2&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B/13B/70B&lt;/td&gt;
          &lt;td&gt;llama2&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/meta-llama&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama 3-3.3&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;1B/3B/8B/70B&lt;/td&gt;
          &lt;td&gt;llama3&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/meta-llama&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama 4&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;109B/402B&lt;/td&gt;
          &lt;td&gt;llama4&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/meta-llama&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama 3.2 Vision&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;11B/90B&lt;/td&gt;
          &lt;td&gt;mllama&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/llava-hf&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LLaVA-1.5&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B/13B&lt;/td&gt;
          &lt;td&gt;llava&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/llava-hf&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LLaVA-NeXT&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B/8B/13B/34B/72B/110B&lt;/td&gt;
          &lt;td&gt;llava_next&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/llava-hf&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LLaVA-NeXT-Video&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B/34B&lt;/td&gt;
          &lt;td&gt;llava_next_video&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/XiaomiMiMo&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;MiMo&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B&lt;/td&gt;
          &lt;td&gt;mimo&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/openbmb&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;MiniCPM&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;1B/2B/4B&lt;/td&gt;
          &lt;td&gt;cpm/cpm3&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/openbmb&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;MiniCPM-o-2.6/MiniCPM-V-2.6&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;8B&lt;/td&gt;
          &lt;td&gt;minicpm_o/minicpm_v&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/mistralai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Ministral/Mistral-Nemo&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;8B/12B&lt;/td&gt;
          &lt;td&gt;ministral&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/mistralai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Mistral/Mixtral&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B/8x7B/8x22B&lt;/td&gt;
          &lt;td&gt;mistral&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/mistralai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Mistral Small&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;24B&lt;/td&gt;
          &lt;td&gt;mistral_small&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/allenai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;OLMo&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;1B/7B&lt;/td&gt;
          &lt;td&gt;-&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/google&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;PaliGemma/PaliGemma2&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;3B/10B/28B&lt;/td&gt;
          &lt;td&gt;paligemma&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/microsoft&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Phi-1.5/Phi-2&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;1.3B/2.7B&lt;/td&gt;
          &lt;td&gt;-&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/microsoft&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Phi-3/Phi-3.5&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;4B/14B&lt;/td&gt;
          &lt;td&gt;phi&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/microsoft&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Phi-3-small&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B&lt;/td&gt;
          &lt;td&gt;phi_small&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/microsoft&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Phi-4&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;14B&lt;/td&gt;
          &lt;td&gt;phi4&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/mistralai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Pixtral&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;12B&lt;/td&gt;
          &lt;td&gt;pixtral&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Qwen&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen (1-2.5) (Code/Math/MoE/QwQ)&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;0.5B/1.5B/3B/7B/14B/32B/72B/110B&lt;/td&gt;
          &lt;td&gt;qwen&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Qwen&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen3 (MoE)&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;0.6B/1.7B/4B/8B/14B/32B/235B&lt;/td&gt;
          &lt;td&gt;qwen3&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Qwen&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen2-Audio&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B&lt;/td&gt;
          &lt;td&gt;qwen2_audio&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Qwen&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen2.5-Omni&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;3B/7B&lt;/td&gt;
          &lt;td&gt;qwen2_omni&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Qwen&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen2-VL/Qwen2.5-VL/QVQ&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;2B/3B/7B/32B/72B&lt;/td&gt;
          &lt;td&gt;qwen2_vl&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/ByteDance-Seed&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Seed Coder&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;8B&lt;/td&gt;
          &lt;td&gt;seed_coder&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Skywork&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Skywork o1&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;8B&lt;/td&gt;
          &lt;td&gt;skywork_o1&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/bigcode&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;StarCoder 2&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;3B/7B/15B&lt;/td&gt;
          &lt;td&gt;-&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Tele-AI&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;TeleChat2&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;3B/7B/35B/115B&lt;/td&gt;
          &lt;td&gt;telechat2&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/xverse&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;XVERSE&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;7B/13B/65B&lt;/td&gt;
          &lt;td&gt;xverse&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/01-ai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Yi/Yi-1.5 (Code)&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;1.5B/6B/9B/34B&lt;/td&gt;
          &lt;td&gt;yi&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/01-ai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Yi-VL&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;6B/34B&lt;/td&gt;
          &lt;td&gt;yi_vl&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/IEITYuan&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Yuan 2&lt;/a&gt;&lt;/td&gt;
          &lt;td&gt;2B/51B/102B&lt;/td&gt;
          &lt;td&gt;yuan&lt;/td&gt;
      &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;
&lt;blockquote&gt;
&lt;p&gt;[!NOTE]
For the &amp;ldquo;base&amp;rdquo; models, the &lt;code&gt;template&lt;/code&gt; argument can be chosen from &lt;code&gt;default&lt;/code&gt;, &lt;code&gt;alpaca&lt;/code&gt;, &lt;code&gt;vicuna&lt;/code&gt; etc. But make sure to use the &lt;strong&gt;corresponding template&lt;/strong&gt; for the &amp;ldquo;instruct/chat&amp;rdquo; models.&lt;/p&gt;
&lt;p&gt;Remember to use the &lt;strong&gt;SAME&lt;/strong&gt; template in training and inference.&lt;/p&gt;
&lt;p&gt;*: You should install the &lt;code&gt;transformers&lt;/code&gt; from main branch and use &lt;code&gt;DISABLE_VERSION_CHECK=1&lt;/code&gt; to skip version check.&lt;/p&gt;
&lt;p&gt;**: You need to install a specific version of &lt;code&gt;transformers&lt;/code&gt; to use the corresponding model.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Please refer to &lt;a class=&#34;link&#34; href=&#34;src/llamafactory/extras/constants.py&#34; &gt;constants.py&lt;/a&gt; for a full list of models we supported.&lt;/p&gt;
&lt;p&gt;You also can add a custom chat template to &lt;a class=&#34;link&#34; href=&#34;src/llamafactory/data/template.py&#34; &gt;template.py&lt;/a&gt;.&lt;/p&gt;
&lt;h2 id=&#34;supported-training-approaches&#34;&gt;Supported Training Approaches
&lt;/h2&gt;&lt;table&gt;
  &lt;thead&gt;
      &lt;tr&gt;
          &lt;th&gt;Approach&lt;/th&gt;
          &lt;th&gt;Full-tuning&lt;/th&gt;
          &lt;th&gt;Freeze-tuning&lt;/th&gt;
          &lt;th&gt;LoRA&lt;/th&gt;
          &lt;th&gt;QLoRA&lt;/th&gt;
      &lt;/tr&gt;
  &lt;/thead&gt;
  &lt;tbody&gt;
      &lt;tr&gt;
          &lt;td&gt;Pre-Training&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;Supervised Fine-Tuning&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;Reward Modeling&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;PPO Training&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;DPO Training&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;KTO Training&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;ORPO Training&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;SimPO Training&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
          &lt;td&gt;:white_check_mark:&lt;/td&gt;
      &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;
&lt;blockquote&gt;
&lt;p&gt;[!TIP]
The implementation details of PPO can be found in &lt;a class=&#34;link&#34; href=&#34;https://newfacade.github.io/notes-on-reinforcement-learning/17-ppo-trl.html&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;this blog&lt;/a&gt;.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;h2 id=&#34;provided-datasets&#34;&gt;Provided Datasets
&lt;/h2&gt;&lt;details&gt;&lt;summary&gt;Pre-training datasets&lt;/summary&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;data/wiki_demo.txt&#34; &gt;Wiki Demo (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/tiiuae/falcon-refinedweb&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;RefinedWeb (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/togethercomputer/RedPajama-Data-V2&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;RedPajama V2 (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/olm/olm-wikipedia-20221220&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Wikipedia (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/pleisto/wikipedia-cn-20230720-filtered&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Wikipedia (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/EleutherAI/pile&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Pile (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/Skywork/SkyPile-150B&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;SkyPile (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/HuggingFaceFW/fineweb&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;FineWeb (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;FineWeb-Edu (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/bigcode/the-stack&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;The Stack (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/bigcode/starcoderdata&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;StarCoder (en)&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/details&gt;
&lt;details&gt;&lt;summary&gt;Supervised fine-tuning datasets&lt;/summary&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;data/identity.json&#34; &gt;Identity (en&amp;amp;zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/tatsu-lab/stanford_alpaca&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Stanford Alpaca (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/ymcui/Chinese-LLaMA-Alpaca-3&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Stanford Alpaca (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/Instruction-Tuning-with-GPT-4/GPT-4-LLM&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Alpaca GPT4 (en&amp;amp;zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/glaiveai/glaive-function-calling-v2&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Glaive Function Calling V2 (en&amp;amp;zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/GAIR/lima&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LIMA (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/JosephusCheung/GuanacoDataset&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Guanaco Dataset (multilingual)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/BelleGroup/train_2M_CN&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;BELLE 2M (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/BelleGroup/train_1M_CN&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;BELLE 1M (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/BelleGroup/train_0.5M_CN&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;BELLE 0.5M (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/BelleGroup/generated_chat_0.4M&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;BELLE Dialogue 0.4M (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/BelleGroup/school_math_0.25M&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;BELLE School Math 0.25M (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/BelleGroup/multiturn_chat_0.8M&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;BELLE Multiturn Chat 0.8M (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/thunlp/UltraChat&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;UltraChat (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/garage-bAInd/Open-Platypus&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;OpenPlatypus (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/sahil2801/CodeAlpaca-20k&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;CodeAlpaca 20k (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/QingyiSi/Alpaca-CoT&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Alpaca CoT (multilingual)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/Open-Orca/OpenOrca&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;OpenOrca (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/Open-Orca/SlimOrca&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;SlimOrca (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/TIGER-Lab/MathInstruct&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;MathInstruct (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/YeungNLP/firefly-train-1.1M&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Firefly 1.1M (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/wiki_qa&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Wiki QA (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/suolyer/webqa&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Web QA (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/zxbsmk/webnovel_cn&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;WebNovel (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/berkeley-nest/Nectar&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Nectar (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://www.modelscope.cn/datasets/deepctrl/deepctrl-sft-data&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;deepctrl (en&amp;amp;zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/HasturOfficial/adgen&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Advertise Generating (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/totally-not-an-llm/sharegpt-hyperfiltered-3k&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;ShareGPT Hyperfiltered (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/shibing624/sharegpt_gpt4&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;ShareGPT4 (en&amp;amp;zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/HuggingFaceH4/ultrachat_200k&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;UltraChat 200k (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/THUDM/AgentInstruct&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;AgentInstruct (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/lmsys/lmsys-chat-1m&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LMSYS Chat 1M (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/WizardLM/WizardLM_evol_instruct_V2_196k&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Evol Instruct V2 (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/HuggingFaceTB/cosmopedia&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Cosmopedia (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/hfl/stem_zh_instruction&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;STEM (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/hfl/ruozhiba_gpt4_turbo&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Ruozhiba (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/m-a-p/neo_sft_phase2&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Neo-sft (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/Magpie-Align/Magpie-Pro-300K-Filtered&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Magpie-Pro-300K-Filtered (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/argilla/magpie-ultra-v0.1&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Magpie-ultra-v0.1 (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/TIGER-Lab/WebInstructSub&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;WebInstructSub (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/O1-OPEN/OpenO1-SFT&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;OpenO1-SFT (en&amp;amp;zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Open-Thoughts (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/open-r1/OpenR1-Math-220k&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Open-R1-Math (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/Congliu/Chinese-DeepSeek-R1-Distill-data-110k-SFT&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Chinese-DeepSeek-R1-Distill (zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/BUAADreamer/llava-en-zh-300k&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LLaVA mixed (en&amp;amp;zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/jugg1024/pokemon-gpt4o-captions&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Pokemon-gpt4o-captions (en&amp;amp;zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/mayflowergmbh/oasst_de&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Open Assistant (de)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/mayflowergmbh/dolly-15k_de&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Dolly 15k (de)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/mayflowergmbh/alpaca-gpt4_de&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Alpaca GPT4 (de)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/mayflowergmbh/openschnabeltier_de&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;OpenSchnabeltier (de)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/mayflowergmbh/evol-instruct_de&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Evol Instruct (de)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/mayflowergmbh/dolphin_de&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Dolphin (de)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/mayflowergmbh/booksum_de&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Booksum (de)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/mayflowergmbh/airoboros-3.0_de&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Airoboros (de)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/mayflowergmbh/ultra-chat_de&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Ultrachat (de)&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/details&gt;
&lt;details&gt;&lt;summary&gt;Preference datasets&lt;/summary&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/hiyouga/DPO-En-Zh-20k&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;DPO mixed (en&amp;amp;zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/HuggingFaceH4/ultrafeedback_binarized&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;UltraFeedback (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/m-a-p/COIG-P&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;COIG-P (en&amp;amp;zh)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/openbmb/RLHF-V-Dataset&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;RLHF-V (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/Zhihui/VLFeedback&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;VLFeedback (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/openbmb/RLAIF-V-Dataset&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;RLAIF-V (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/Intel/orca_dpo_pairs&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Orca DPO Pairs (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/Anthropic/hh-rlhf&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;HH-RLHF (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/berkeley-nest/Nectar&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Nectar (en)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/mayflowergmbh/intel_orca_dpo_pairs_de&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Orca DPO (de)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/datasets/argilla/kto-mix-15k&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;KTO mixed (en)&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;/details&gt;
&lt;p&gt;Some datasets require confirmation before using them, so we recommend logging in with your Hugging Face account using these commands.&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;pip install --upgrade huggingface_hub
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;huggingface-cli login
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;h2 id=&#34;requirement&#34;&gt;Requirement
&lt;/h2&gt;&lt;table&gt;
  &lt;thead&gt;
      &lt;tr&gt;
          &lt;th&gt;Mandatory&lt;/th&gt;
          &lt;th&gt;Minimum&lt;/th&gt;
          &lt;th&gt;Recommend&lt;/th&gt;
      &lt;/tr&gt;
  &lt;/thead&gt;
  &lt;tbody&gt;
      &lt;tr&gt;
          &lt;td&gt;python&lt;/td&gt;
          &lt;td&gt;3.9&lt;/td&gt;
          &lt;td&gt;3.10&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;torch&lt;/td&gt;
          &lt;td&gt;2.0.0&lt;/td&gt;
          &lt;td&gt;2.6.0&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;torchvision&lt;/td&gt;
          &lt;td&gt;0.15.0&lt;/td&gt;
          &lt;td&gt;0.21.0&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;transformers&lt;/td&gt;
          &lt;td&gt;4.45.0&lt;/td&gt;
          &lt;td&gt;4.50.0&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;datasets&lt;/td&gt;
          &lt;td&gt;2.16.0&lt;/td&gt;
          &lt;td&gt;3.2.0&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;accelerate&lt;/td&gt;
          &lt;td&gt;0.34.0&lt;/td&gt;
          &lt;td&gt;1.2.1&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;peft&lt;/td&gt;
          &lt;td&gt;0.14.0&lt;/td&gt;
          &lt;td&gt;0.15.1&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;trl&lt;/td&gt;
          &lt;td&gt;0.8.6&lt;/td&gt;
          &lt;td&gt;0.9.6&lt;/td&gt;
      &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;
&lt;table&gt;
  &lt;thead&gt;
      &lt;tr&gt;
          &lt;th&gt;Optional&lt;/th&gt;
          &lt;th&gt;Minimum&lt;/th&gt;
          &lt;th&gt;Recommend&lt;/th&gt;
      &lt;/tr&gt;
  &lt;/thead&gt;
  &lt;tbody&gt;
      &lt;tr&gt;
          &lt;td&gt;CUDA&lt;/td&gt;
          &lt;td&gt;11.6&lt;/td&gt;
          &lt;td&gt;12.2&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;deepspeed&lt;/td&gt;
          &lt;td&gt;0.10.0&lt;/td&gt;
          &lt;td&gt;0.16.4&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;bitsandbytes&lt;/td&gt;
          &lt;td&gt;0.39.0&lt;/td&gt;
          &lt;td&gt;0.43.1&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;vllm&lt;/td&gt;
          &lt;td&gt;0.4.3&lt;/td&gt;
          &lt;td&gt;0.8.2&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;flash-attn&lt;/td&gt;
          &lt;td&gt;2.5.6&lt;/td&gt;
          &lt;td&gt;2.7.2&lt;/td&gt;
      &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;
&lt;h3 id=&#34;hardware-requirement&#34;&gt;Hardware Requirement
&lt;/h3&gt;&lt;p&gt;* &lt;em&gt;estimated&lt;/em&gt;&lt;/p&gt;
&lt;table&gt;
  &lt;thead&gt;
      &lt;tr&gt;
          &lt;th&gt;Method&lt;/th&gt;
          &lt;th&gt;Bits&lt;/th&gt;
          &lt;th&gt;7B&lt;/th&gt;
          &lt;th&gt;14B&lt;/th&gt;
          &lt;th&gt;30B&lt;/th&gt;
          &lt;th&gt;70B&lt;/th&gt;
          &lt;th&gt;&lt;code&gt;x&lt;/code&gt;B&lt;/th&gt;
      &lt;/tr&gt;
  &lt;/thead&gt;
  &lt;tbody&gt;
      &lt;tr&gt;
          &lt;td&gt;Full (&lt;code&gt;bf16&lt;/code&gt; or &lt;code&gt;fp16&lt;/code&gt;)&lt;/td&gt;
          &lt;td&gt;32&lt;/td&gt;
          &lt;td&gt;120GB&lt;/td&gt;
          &lt;td&gt;240GB&lt;/td&gt;
          &lt;td&gt;600GB&lt;/td&gt;
          &lt;td&gt;1200GB&lt;/td&gt;
          &lt;td&gt;&lt;code&gt;18x&lt;/code&gt;GB&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;Full (&lt;code&gt;pure_bf16&lt;/code&gt;)&lt;/td&gt;
          &lt;td&gt;16&lt;/td&gt;
          &lt;td&gt;60GB&lt;/td&gt;
          &lt;td&gt;120GB&lt;/td&gt;
          &lt;td&gt;300GB&lt;/td&gt;
          &lt;td&gt;600GB&lt;/td&gt;
          &lt;td&gt;&lt;code&gt;8x&lt;/code&gt;GB&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;Freeze/LoRA/GaLore/APOLLO/BAdam&lt;/td&gt;
          &lt;td&gt;16&lt;/td&gt;
          &lt;td&gt;16GB&lt;/td&gt;
          &lt;td&gt;32GB&lt;/td&gt;
          &lt;td&gt;64GB&lt;/td&gt;
          &lt;td&gt;160GB&lt;/td&gt;
          &lt;td&gt;&lt;code&gt;2x&lt;/code&gt;GB&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;QLoRA&lt;/td&gt;
          &lt;td&gt;8&lt;/td&gt;
          &lt;td&gt;10GB&lt;/td&gt;
          &lt;td&gt;20GB&lt;/td&gt;
          &lt;td&gt;40GB&lt;/td&gt;
          &lt;td&gt;80GB&lt;/td&gt;
          &lt;td&gt;&lt;code&gt;x&lt;/code&gt;GB&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;QLoRA&lt;/td&gt;
          &lt;td&gt;4&lt;/td&gt;
          &lt;td&gt;6GB&lt;/td&gt;
          &lt;td&gt;12GB&lt;/td&gt;
          &lt;td&gt;24GB&lt;/td&gt;
          &lt;td&gt;48GB&lt;/td&gt;
          &lt;td&gt;&lt;code&gt;x/2&lt;/code&gt;GB&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;QLoRA&lt;/td&gt;
          &lt;td&gt;2&lt;/td&gt;
          &lt;td&gt;4GB&lt;/td&gt;
          &lt;td&gt;8GB&lt;/td&gt;
          &lt;td&gt;16GB&lt;/td&gt;
          &lt;td&gt;24GB&lt;/td&gt;
          &lt;td&gt;&lt;code&gt;x/4&lt;/code&gt;GB&lt;/td&gt;
      &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;
&lt;h2 id=&#34;getting-started&#34;&gt;Getting Started
&lt;/h2&gt;&lt;h3 id=&#34;installation&#34;&gt;Installation
&lt;/h3&gt;&lt;blockquote&gt;
&lt;p&gt;[!IMPORTANT]
Installation is mandatory.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;3
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;git clone --depth &lt;span class=&#34;m&#34;&gt;1&lt;/span&gt; https://github.com/hiyouga/LLaMA-Factory.git
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nb&#34;&gt;cd&lt;/span&gt; LLaMA-Factory
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;pip install -e &lt;span class=&#34;s2&#34;&gt;&amp;#34;.[torch,metrics]&amp;#34;&lt;/span&gt; --no-build-isolation
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;Extra dependencies available: torch, torch-npu, metrics, deepspeed, liger-kernel, bitsandbytes, hqq, eetq, gptq, aqlm, vllm, sglang, galore, apollo, badam, adam-mini, qwen, minicpm_v, modelscope, openmind, swanlab, quality&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;[!TIP]
Use &lt;code&gt;pip install -e . --no-deps --no-build-isolation&lt;/code&gt; to resolve package conflicts.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;details&gt;&lt;summary&gt;Setting up a virtual environment with &lt;b&gt;uv&lt;/b&gt;&lt;/summary&gt;
&lt;p&gt;Create an isolated Python environment with &lt;a class=&#34;link&#34; href=&#34;https://github.com/astral-sh/uv&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;uv&lt;/a&gt;:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;uv sync --extra torch --extra metrics --prerelease&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;allow
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;Run LLaMA-Factory in the isolated environment:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;uv run --prerelease&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;allow llamafactory-cli train examples/train_lora/llama3_lora_pretrain.yaml
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;/details&gt;
&lt;details&gt;&lt;summary&gt;For Windows users&lt;/summary&gt;
&lt;h4 id=&#34;install-pytorch&#34;&gt;Install PyTorch
&lt;/h4&gt;&lt;p&gt;You need to manually install the GPU version of PyTorch on the Windows platform. Please refer to the &lt;a class=&#34;link&#34; href=&#34;https://pytorch.org/get-started/locally/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;official website&lt;/a&gt; and the following command to install PyTorch with CUDA support:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;3
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;pip uninstall torch torchvision torchaudio
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu126
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;python -c &lt;span class=&#34;s2&#34;&gt;&amp;#34;import torch; print(torch.cuda.is_available())&amp;#34;&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;If you see &lt;code&gt;True&lt;/code&gt; then you have successfully installed PyTorch with CUDA support.&lt;/p&gt;
&lt;p&gt;Try &lt;code&gt;dataloader_num_workers: 0&lt;/code&gt; if you encounter &lt;code&gt;Can&#39;t pickle local object&lt;/code&gt; error.&lt;/p&gt;
&lt;h4 id=&#34;install-bitsandbytes&#34;&gt;Install BitsAndBytes
&lt;/h4&gt;&lt;p&gt;If you want to enable the quantized LoRA (QLoRA) on the Windows platform, you need to install a pre-built version of &lt;code&gt;bitsandbytes&lt;/code&gt; library, which supports CUDA 11.1 to 12.2, please select the appropriate &lt;a class=&#34;link&#34; href=&#34;https://github.com/jllllll/bitsandbytes-windows-webui/releases/tag/wheels&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;release version&lt;/a&gt; based on your CUDA version.&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;pip install https://github.com/jllllll/bitsandbytes-windows-webui/releases/download/wheels/bitsandbytes-0.41.2.post2-py3-none-win_amd64.whl
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;h4 id=&#34;install-flash-attention-2&#34;&gt;Install Flash Attention-2
&lt;/h4&gt;&lt;p&gt;To enable FlashAttention-2 on the Windows platform, please use the script from &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/lldacing/flash-attention-windows-wheel&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;flash-attention-windows-wheel&lt;/a&gt; to compile and install it by yourself.&lt;/p&gt;
&lt;/details&gt;
&lt;details&gt;&lt;summary&gt;For Ascend NPU users&lt;/summary&gt;
&lt;p&gt;To install LLaMA Factory on Ascend NPU devices, please upgrade Python to version 3.10 or higher and specify extra dependencies: &lt;code&gt;pip install -e &amp;quot;.[torch-npu,metrics]&amp;quot;&lt;/code&gt;. Additionally, you need to install the &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://www.hiascend.com/developer/download/community/result?module=cann&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Ascend CANN Toolkit and Kernels&lt;/a&gt;&lt;/strong&gt;. Please follow the &lt;a class=&#34;link&#34; href=&#34;https://www.hiascend.com/document/detail/en/CANNCommunityEdition/600alphaX/softwareinstall/instg/atlasdeploy_03_0031.html&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;installation tutorial&lt;/a&gt; or use the following commands:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt; 1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 3
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 4
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 5
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 6
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 7
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 8
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 9
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;10
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;11
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;c1&#34;&gt;# replace the url according to your CANN version and devices&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;c1&#34;&gt;# install CANN Toolkit&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;wget https://ascend-repo.obs.cn-east-2.myhuaweicloud.com/Milan-ASL/Milan-ASL%20V100R001C20SPC702/Ascend-cann-toolkit_8.0.0.alpha002_linux-&lt;span class=&#34;s2&#34;&gt;&amp;#34;&lt;/span&gt;&lt;span class=&#34;k&#34;&gt;$(&lt;/span&gt;uname -i&lt;span class=&#34;k&#34;&gt;)&lt;/span&gt;&lt;span class=&#34;s2&#34;&gt;&amp;#34;&lt;/span&gt;.run
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;bash Ascend-cann-toolkit_8.0.0.alpha002_linux-&lt;span class=&#34;s2&#34;&gt;&amp;#34;&lt;/span&gt;&lt;span class=&#34;k&#34;&gt;$(&lt;/span&gt;uname -i&lt;span class=&#34;k&#34;&gt;)&lt;/span&gt;&lt;span class=&#34;s2&#34;&gt;&amp;#34;&lt;/span&gt;.run --install
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;c1&#34;&gt;# install CANN Kernels&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;wget https://ascend-repo.obs.cn-east-2.myhuaweicloud.com/Milan-ASL/Milan-ASL%20V100R001C20SPC702/Ascend-cann-kernels-910b_8.0.0.alpha002_linux-&lt;span class=&#34;s2&#34;&gt;&amp;#34;&lt;/span&gt;&lt;span class=&#34;k&#34;&gt;$(&lt;/span&gt;uname -i&lt;span class=&#34;k&#34;&gt;)&lt;/span&gt;&lt;span class=&#34;s2&#34;&gt;&amp;#34;&lt;/span&gt;.run
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;bash Ascend-cann-kernels-910b_8.0.0.alpha002_linux-&lt;span class=&#34;s2&#34;&gt;&amp;#34;&lt;/span&gt;&lt;span class=&#34;k&#34;&gt;$(&lt;/span&gt;uname -i&lt;span class=&#34;k&#34;&gt;)&lt;/span&gt;&lt;span class=&#34;s2&#34;&gt;&amp;#34;&lt;/span&gt;.run --install
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;c1&#34;&gt;# set env variables&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nb&#34;&gt;source&lt;/span&gt; /usr/local/Ascend/ascend-toolkit/set_env.sh
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;table&gt;
  &lt;thead&gt;
      &lt;tr&gt;
          &lt;th&gt;Requirement&lt;/th&gt;
          &lt;th&gt;Minimum&lt;/th&gt;
          &lt;th&gt;Recommend&lt;/th&gt;
      &lt;/tr&gt;
  &lt;/thead&gt;
  &lt;tbody&gt;
      &lt;tr&gt;
          &lt;td&gt;CANN&lt;/td&gt;
          &lt;td&gt;8.0.RC1&lt;/td&gt;
          &lt;td&gt;8.0.0.alpha002&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;torch&lt;/td&gt;
          &lt;td&gt;2.1.0&lt;/td&gt;
          &lt;td&gt;2.4.0&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;torch-npu&lt;/td&gt;
          &lt;td&gt;2.1.0&lt;/td&gt;
          &lt;td&gt;2.4.0.post2&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;deepspeed&lt;/td&gt;
          &lt;td&gt;0.13.2&lt;/td&gt;
          &lt;td&gt;0.13.2&lt;/td&gt;
      &lt;/tr&gt;
      &lt;tr&gt;
          &lt;td&gt;vllm-ascend&lt;/td&gt;
          &lt;td&gt;-&lt;/td&gt;
          &lt;td&gt;0.7.3&lt;/td&gt;
      &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;Remember to use &lt;code&gt;ASCEND_RT_VISIBLE_DEVICES&lt;/code&gt; instead of &lt;code&gt;CUDA_VISIBLE_DEVICES&lt;/code&gt; to specify the device to use.&lt;/p&gt;
&lt;p&gt;If you cannot infer model on NPU devices, try setting &lt;code&gt;do_sample: false&lt;/code&gt; in the configurations.&lt;/p&gt;
&lt;p&gt;Download the pre-built Docker images: &lt;a class=&#34;link&#34; href=&#34;http://mirrors.cn-central-221.ovaijisuan.com/detail/130.html&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;32GB&lt;/a&gt; | &lt;a class=&#34;link&#34; href=&#34;http://mirrors.cn-central-221.ovaijisuan.com/detail/131.html&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;64GB&lt;/a&gt;&lt;/p&gt;
&lt;h4 id=&#34;install-bitsandbytes-1&#34;&gt;Install BitsAndBytes
&lt;/h4&gt;&lt;p&gt;To use QLoRA based on bitsandbytes on Ascend NPU, please follow these 3 steps:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;Manually compile bitsandbytes: Refer to &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/docs/bitsandbytes/installation?backend=Ascend&amp;#43;NPU&amp;amp;platform=Ascend&amp;#43;NPU&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;the installation documentation&lt;/a&gt; for the NPU version of bitsandbytes to complete the compilation and installation. The compilation requires a cmake version of at least 3.22.1 and a g++ version of at least 12.x.&lt;/li&gt;
&lt;/ol&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt; 1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 3
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 4
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 5
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 6
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 7
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 8
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 9
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;10
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;11
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;12
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;13
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;14
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;15
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;c1&#34;&gt;# Install bitsandbytes from source&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;c1&#34;&gt;# Clone bitsandbytes repo, Ascend NPU backend is currently enabled on multi-backend-refactor branch&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;git clone -b multi-backend-refactor https://github.com/bitsandbytes-foundation/bitsandbytes.git
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nb&#34;&gt;cd&lt;/span&gt; bitsandbytes/
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;c1&#34;&gt;# Install dependencies&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;pip install -r requirements-dev.txt
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;c1&#34;&gt;# Install the dependencies for the compilation tools. Note that the commands for this step may vary depending on the operating system. The following are provided for reference&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;apt-get install -y build-essential cmake
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;c1&#34;&gt;# Compile &amp;amp; install  &lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;cmake -DCOMPUTE_BACKEND&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;npu -S .
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;make
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;pip install .
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;ol start=&#34;2&#34;&gt;
&lt;li&gt;Install transformers from the main branch.&lt;/li&gt;
&lt;/ol&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;3
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;git clone -b main https://github.com/huggingface/transformers.git
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nb&#34;&gt;cd&lt;/span&gt; transformers
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;pip install .
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;ol start=&#34;3&#34;&gt;
&lt;li&gt;Set &lt;code&gt;double_quantization: false&lt;/code&gt; in the configuration. You can refer to the &lt;a class=&#34;link&#34; href=&#34;examples/train_qlora/llama3_lora_sft_bnb_npu.yaml&#34; &gt;example&lt;/a&gt;.&lt;/li&gt;
&lt;/ol&gt;
&lt;/details&gt;
&lt;h3 id=&#34;data-preparation&#34;&gt;Data Preparation
&lt;/h3&gt;&lt;p&gt;Please refer to &lt;a class=&#34;link&#34; href=&#34;data/README.md&#34; &gt;data/README.md&lt;/a&gt; for checking the details about the format of dataset files. You can use datasets on HuggingFace / ModelScope / Modelers hub, load the dataset in local disk, or specify a path to s3/gcs cloud storage.&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;[!NOTE]
Please update &lt;code&gt;data/dataset_info.json&lt;/code&gt; to use your custom dataset.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;You can also use &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/ConardLi/easy-dataset&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Easy Dataset&lt;/a&gt;&lt;/strong&gt; or &lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/open-sciencelab/GraphGen&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;GraphGen&lt;/a&gt;&lt;/strong&gt; to create synthetic data for fine-tuning.&lt;/p&gt;
&lt;h3 id=&#34;quickstart&#34;&gt;Quickstart
&lt;/h3&gt;&lt;p&gt;Use the following 3 commands to run LoRA &lt;strong&gt;fine-tuning&lt;/strong&gt;, &lt;strong&gt;inference&lt;/strong&gt; and &lt;strong&gt;merging&lt;/strong&gt; of the Llama3-8B-Instruct model, respectively.&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;3
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;llamafactory-cli train examples/train_lora/llama3_lora_sft.yaml
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;llamafactory-cli chat examples/inference/llama3_lora_sft.yaml
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;llamafactory-cli &lt;span class=&#34;nb&#34;&gt;export&lt;/span&gt; examples/merge_lora/llama3_lora_sft.yaml
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;See &lt;a class=&#34;link&#34; href=&#34;examples/README.md&#34; &gt;examples/README.md&lt;/a&gt; for advanced usage (including distributed training).&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;[!TIP]
Use &lt;code&gt;llamafactory-cli help&lt;/code&gt; to show help information.&lt;/p&gt;
&lt;p&gt;Read &lt;a class=&#34;link&#34; href=&#34;https://github.com/hiyouga/LLaMA-Factory/issues/4614&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;FAQs&lt;/a&gt; first if you encounter any problems.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;h3 id=&#34;fine-tuning-with-llama-board-gui-powered-by-gradio&#34;&gt;Fine-Tuning with LLaMA Board GUI (powered by &lt;a class=&#34;link&#34; href=&#34;https://github.com/gradio-app/gradio&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Gradio&lt;/a&gt;)
&lt;/h3&gt;&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;llamafactory-cli webui
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;h3 id=&#34;build-docker&#34;&gt;Build Docker
&lt;/h3&gt;&lt;p&gt;For CUDA users:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;3
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nb&#34;&gt;cd&lt;/span&gt; docker/docker-cuda/
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker compose up -d
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker compose &lt;span class=&#34;nb&#34;&gt;exec&lt;/span&gt; llamafactory bash
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;For Ascend NPU users:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;3
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nb&#34;&gt;cd&lt;/span&gt; docker/docker-npu/
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker compose up -d
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker compose &lt;span class=&#34;nb&#34;&gt;exec&lt;/span&gt; llamafactory bash
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;For AMD ROCm users:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;3
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nb&#34;&gt;cd&lt;/span&gt; docker/docker-rocm/
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker compose up -d
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker compose &lt;span class=&#34;nb&#34;&gt;exec&lt;/span&gt; llamafactory bash
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;details&gt;&lt;summary&gt;Build without Docker Compose&lt;/summary&gt;
&lt;p&gt;For CUDA users:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt; 1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 3
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 4
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 5
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 6
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 7
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 8
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 9
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;10
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;11
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;12
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;13
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;14
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;15
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;16
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;17
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;18
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;19
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;20
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;21
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker build -f ./docker/docker-cuda/Dockerfile &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;INSTALL_BNB&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;nb&#34;&gt;false&lt;/span&gt; &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;INSTALL_VLLM&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;nb&#34;&gt;false&lt;/span&gt; &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;INSTALL_DEEPSPEED&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;nb&#34;&gt;false&lt;/span&gt; &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;INSTALL_FLASHATTN&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;nb&#34;&gt;false&lt;/span&gt; &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;PIP_INDEX&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;https://pypi.org/simple &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -t llamafactory:latest .
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker run -dit --gpus&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;all &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./hf_cache:/root/.cache/huggingface &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./ms_cache:/root/.cache/modelscope &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./om_cache:/root/.cache/openmind &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./data:/app/data &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./output:/app/output &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -p 7860:7860 &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -p 8000:8000 &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --shm-size 16G &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --name llamafactory &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    llamafactory:latest
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker &lt;span class=&#34;nb&#34;&gt;exec&lt;/span&gt; -it llamafactory bash
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;For Ascend NPU users:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt; 1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 3
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 4
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 5
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 6
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 7
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 8
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 9
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;10
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;11
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;12
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;13
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;14
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;15
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;16
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;17
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;18
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;19
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;20
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;21
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;22
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;23
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;24
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;25
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;26
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;27
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;28
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;c1&#34;&gt;# Choose docker image upon your environment&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker build -f ./docker/docker-npu/Dockerfile &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;INSTALL_DEEPSPEED&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;nb&#34;&gt;false&lt;/span&gt; &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;PIP_INDEX&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;https://pypi.org/simple &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -t llamafactory:latest .
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;c1&#34;&gt;# Change `device` upon your resources&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker run -dit &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./hf_cache:/root/.cache/huggingface &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./ms_cache:/root/.cache/modelscope &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./om_cache:/root/.cache/openmind &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./data:/app/data &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./output:/app/output &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v /usr/local/dcmi:/usr/local/dcmi &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v /usr/local/bin/npu-smi:/usr/local/bin/npu-smi &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v /usr/local/Ascend/driver:/usr/local/Ascend/driver &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v /etc/ascend_install.info:/etc/ascend_install.info &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -p 7860:7860 &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -p 8000:8000 &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --device /dev/davinci0 &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --device /dev/davinci_manager &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --device /dev/devmm_svm &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --device /dev/hisi_hdc &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --shm-size 16G &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --name llamafactory &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    llamafactory:latest
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker &lt;span class=&#34;nb&#34;&gt;exec&lt;/span&gt; -it llamafactory bash
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;For AMD ROCm users:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt; 1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 3
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 4
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 5
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 6
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 7
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 8
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt; 9
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;10
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;11
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;12
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;13
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;14
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;15
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;16
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;17
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;18
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;19
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;20
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;21
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;22
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;23
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;24
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker build -f ./docker/docker-rocm/Dockerfile &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;INSTALL_BNB&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;nb&#34;&gt;false&lt;/span&gt; &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;INSTALL_VLLM&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;nb&#34;&gt;false&lt;/span&gt; &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;INSTALL_DEEPSPEED&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;nb&#34;&gt;false&lt;/span&gt; &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;INSTALL_FLASHATTN&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;nb&#34;&gt;false&lt;/span&gt; &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --build-arg &lt;span class=&#34;nv&#34;&gt;PIP_INDEX&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;https://pypi.org/simple &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -t llamafactory:latest .
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker run -dit &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./hf_cache:/root/.cache/huggingface &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./ms_cache:/root/.cache/modelscope &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./om_cache:/root/.cache/openmind &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./data:/app/data &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./output:/app/output &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -v ./saves:/app/saves &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -p 7860:7860 &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    -p 8000:8000 &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --device /dev/kfd &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --device /dev/dri &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --shm-size 16G &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    --name llamafactory &lt;span class=&#34;se&#34;&gt;\
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;    llamafactory:latest
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;docker &lt;span class=&#34;nb&#34;&gt;exec&lt;/span&gt; -it llamafactory bash
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;/details&gt;
&lt;details&gt;&lt;summary&gt;Details about volume&lt;/summary&gt;
&lt;ul&gt;
&lt;li&gt;&lt;code&gt;hf_cache&lt;/code&gt;: Utilize Hugging Face cache on the host machine. Reassignable if a cache already exists in a different directory.&lt;/li&gt;
&lt;li&gt;&lt;code&gt;ms_cache&lt;/code&gt;: Similar to Hugging Face cache but for ModelScope users.&lt;/li&gt;
&lt;li&gt;&lt;code&gt;om_cache&lt;/code&gt;: Similar to Hugging Face cache but for Modelers users.&lt;/li&gt;
&lt;li&gt;&lt;code&gt;data&lt;/code&gt;: Place datasets on this dir of the host machine so that they can be selected on LLaMA Board GUI.&lt;/li&gt;
&lt;li&gt;&lt;code&gt;output&lt;/code&gt;: Set export dir to this location so that the merged result can be accessed directly on the host machine.&lt;/li&gt;
&lt;/ul&gt;
&lt;/details&gt;
&lt;h3 id=&#34;deploy-with-openai-style-api-and-vllm&#34;&gt;Deploy with OpenAI-style API and vLLM
&lt;/h3&gt;&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nv&#34;&gt;API_PORT&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;m&#34;&gt;8000&lt;/span&gt; llamafactory-cli api examples/inference/llama3.yaml &lt;span class=&#34;nv&#34;&gt;infer_backend&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;vllm &lt;span class=&#34;nv&#34;&gt;vllm_enforce_eager&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;nb&#34;&gt;true&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;blockquote&gt;
&lt;p&gt;[!TIP]
Visit &lt;a class=&#34;link&#34; href=&#34;https://platform.openai.com/docs/api-reference/chat/create&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;this page&lt;/a&gt; for API document.&lt;/p&gt;
&lt;p&gt;Examples: &lt;a class=&#34;link&#34; href=&#34;scripts/api_example/test_image.py&#34; &gt;Image understanding&lt;/a&gt; | &lt;a class=&#34;link&#34; href=&#34;scripts/api_example/test_toolcall.py&#34; &gt;Function calling&lt;/a&gt;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;h3 id=&#34;download-from-modelscope-hub&#34;&gt;Download from ModelScope Hub
&lt;/h3&gt;&lt;p&gt;If you have trouble with downloading models and datasets from Hugging Face, you can use ModelScope.&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nb&#34;&gt;export&lt;/span&gt; &lt;span class=&#34;nv&#34;&gt;USE_MODELSCOPE_HUB&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;m&#34;&gt;1&lt;/span&gt; &lt;span class=&#34;c1&#34;&gt;# `set USE_MODELSCOPE_HUB=1` for Windows&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;Train the model by specifying a model ID of the ModelScope Hub as the &lt;code&gt;model_name_or_path&lt;/code&gt;. You can find a full list of model IDs at &lt;a class=&#34;link&#34; href=&#34;https://modelscope.cn/models&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;ModelScope Hub&lt;/a&gt;, e.g., &lt;code&gt;LLM-Research/Meta-Llama-3-8B-Instruct&lt;/code&gt;.&lt;/p&gt;
&lt;h3 id=&#34;download-from-modelers-hub&#34;&gt;Download from Modelers Hub
&lt;/h3&gt;&lt;p&gt;You can also use Modelers Hub to download models and datasets.&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nb&#34;&gt;export&lt;/span&gt; &lt;span class=&#34;nv&#34;&gt;USE_OPENMIND_HUB&lt;/span&gt;&lt;span class=&#34;o&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;m&#34;&gt;1&lt;/span&gt; &lt;span class=&#34;c1&#34;&gt;# `set USE_OPENMIND_HUB=1` for Windows&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;Train the model by specifying a model ID of the Modelers Hub as the &lt;code&gt;model_name_or_path&lt;/code&gt;. You can find a full list of model IDs at &lt;a class=&#34;link&#34; href=&#34;https://modelers.cn/models&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Modelers Hub&lt;/a&gt;, e.g., &lt;code&gt;TeleAI/TeleChat-7B-pt&lt;/code&gt;.&lt;/p&gt;
&lt;h3 id=&#34;use-wb-logger&#34;&gt;Use W&amp;amp;B Logger
&lt;/h3&gt;&lt;p&gt;To use &lt;a class=&#34;link&#34; href=&#34;https://wandb.ai&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Weights &amp;amp; Biases&lt;/a&gt; for logging experimental results, you need to add the following arguments to yaml files.&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-yaml&#34; data-lang=&#34;yaml&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nt&#34;&gt;report_to&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;:&lt;/span&gt;&lt;span class=&#34;w&#34;&gt; &lt;/span&gt;&lt;span class=&#34;l&#34;&gt;wandb&lt;/span&gt;&lt;span class=&#34;w&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nt&#34;&gt;run_name&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;:&lt;/span&gt;&lt;span class=&#34;w&#34;&gt; &lt;/span&gt;&lt;span class=&#34;l&#34;&gt;test_run&lt;/span&gt;&lt;span class=&#34;w&#34;&gt; &lt;/span&gt;&lt;span class=&#34;c&#34;&gt;# optional&lt;/span&gt;&lt;span class=&#34;w&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;Set &lt;code&gt;WANDB_API_KEY&lt;/code&gt; to &lt;a class=&#34;link&#34; href=&#34;https://wandb.ai/authorize&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;your key&lt;/a&gt; when launching training tasks to log in with your W&amp;amp;B account.&lt;/p&gt;
&lt;h3 id=&#34;use-swanlab-logger&#34;&gt;Use SwanLab Logger
&lt;/h3&gt;&lt;p&gt;To use &lt;a class=&#34;link&#34; href=&#34;https://github.com/SwanHubX/SwanLab&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;SwanLab&lt;/a&gt; for logging experimental results, you need to add the following arguments to yaml files.&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-yaml&#34; data-lang=&#34;yaml&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nt&#34;&gt;use_swanlab&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;:&lt;/span&gt;&lt;span class=&#34;w&#34;&gt; &lt;/span&gt;&lt;span class=&#34;kc&#34;&gt;true&lt;/span&gt;&lt;span class=&#34;w&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nt&#34;&gt;swanlab_run_name&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;:&lt;/span&gt;&lt;span class=&#34;w&#34;&gt; &lt;/span&gt;&lt;span class=&#34;l&#34;&gt;test_run&lt;/span&gt;&lt;span class=&#34;w&#34;&gt; &lt;/span&gt;&lt;span class=&#34;c&#34;&gt;# optional&lt;/span&gt;&lt;span class=&#34;w&#34;&gt;
&lt;/span&gt;&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;When launching training tasks, you can log in to SwanLab in three ways:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;Add &lt;code&gt;swanlab_api_key=&amp;lt;your_api_key&amp;gt;&lt;/code&gt; to the yaml file, and set it to your &lt;a class=&#34;link&#34; href=&#34;https://swanlab.cn/settings&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;API key&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;Set the environment variable &lt;code&gt;SWANLAB_API_KEY&lt;/code&gt; to your &lt;a class=&#34;link&#34; href=&#34;https://swanlab.cn/settings&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;API key&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;Use the &lt;code&gt;swanlab login&lt;/code&gt; command to complete the login.&lt;/li&gt;
&lt;/ol&gt;
&lt;h2 id=&#34;projects-using-llama-factory&#34;&gt;Projects using LLaMA Factory
&lt;/h2&gt;&lt;p&gt;If you have a project that should be incorporated, please contact via email or create a pull request.&lt;/p&gt;
&lt;details&gt;&lt;summary&gt;Click to show&lt;/summary&gt;
&lt;ol&gt;
&lt;li&gt;Wang et al. ESRL: Efficient Sampling-based Reinforcement Learning for Sequence Generation. 2023. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2308.02223&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Yu et al. Open, Closed, or Small Language Models for Text Classification? 2023. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2308.10092&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Wang et al. UbiPhysio: Support Daily Functioning, Fitness, and Rehabilitation with Action Understanding and Feedback in Natural Language. 2023. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2308.10526&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Luceri et al. Leveraging Large Language Models to Detect Influence Campaigns in Social Media. 2023. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2311.07816&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zhang et al. Alleviating Hallucinations of Large Language Models through Induced Hallucinations. 2023. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2312.15710&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Wang et al. Know Your Needs Better: Towards Structured Understanding of Marketer Demands with Analogical Reasoning Augmented LLMs. KDD 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2401.04319&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Wang et al. CANDLE: Iterative Conceptualization and Instantiation Distillation from Large Language Models for Commonsense Reasoning. ACL 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2401.07286&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Choi et al. FACT-GPT: Fact-Checking Augmentation via Claim Matching with LLMs. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.05904&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zhang et al. AutoMathText: Autonomous Data Selection with Language Models for Mathematical Texts. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.07625&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Lyu et al. KnowTuning: Knowledge-aware Fine-tuning for Large Language Models. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.11176&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Yang et al. LaCo: Large Language Model Pruning via Layer Collaps. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.11187&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Bhardwaj et al. Language Models are Homer Simpson! Safety Re-Alignment of Fine-tuned Language Models through Task Arithmetic. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.11746&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Yang et al. Enhancing Empathetic Response Generation by Augmenting LLMs with Small-scale Empathetic Models. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.11801&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Yi et al. Generation Meets Verification: Accelerating Large Language Model Inference with Smart Parallel Auto-Correct Decoding. ACL 2024 Findings. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.11809&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Cao et al. Head-wise Shareable Attention for Large Language Models. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.11819&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zhang et al. Enhancing Multilingual Capabilities of Large Language Models through Self-Distillation from Resource-Rich Languages. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.12204&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Kim et al. Efficient and Effective Vocabulary Expansion Towards Multilingual Large Language Models. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.14714&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Yu et al. KIEval: A Knowledge-grounded Interactive Evaluation Framework for Large Language Models. ACL 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2402.15043&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Huang et al. Key-Point-Driven Data Synthesis with its Enhancement on Mathematical Reasoning. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2403.02333&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Duan et al. Negating Negatives: Alignment without Human Positive Samples via Distributional Dispreference Optimization. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2403.03419&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Xie and Schwertfeger. Empowering Robotics with Large Language Models: osmAG Map Comprehension with LLMs. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2403.08228&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Wu et al. Large Language Models are Parallel Multilingual Learners. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2403.09073&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zhang et al. EDT: Improving Large Language Models&amp;rsquo; Generation by Entropy-based Dynamic Temperature Sampling. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2403.14541&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Weller et al. FollowIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2403.15246&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Hongbin Na. CBT-LLM: A Chinese Large Language Model for Cognitive Behavioral Therapy-based Mental Health Question Answering. COLING 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2403.16008&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zan et al. CodeS: Natural Language to Code Repository via Multi-Layer Sketch. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2403.16443&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Liu et al. Extensive Self-Contrast Enables Feedback-Free Language Model Alignment. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.00604&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Luo et al. BAdam: A Memory Efficient Full Parameter Training Method for Large Language Models. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.02827&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Du et al. Chinese Tiny LLM: Pretraining a Chinese-Centric Large Language Model. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.04167&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Ma et al. Parameter Efficient Quasi-Orthogonal Fine-Tuning via Givens Rotation. ICML 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.04316&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Liu et al. Dynamic Generation of Personalities with Large Language Models. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.07084&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Shang et al. How Far Have We Gone in Stripped Binary Code Understanding Using Large Language Models. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.09836&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Huang et al. LLMTune: Accelerate Database Knob Tuning with Large Language Models. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.11581&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Deng et al. Text-Tuple-Table: Towards Information Integration in Text-to-Table Generation via Global Tuple Extraction. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.14215&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Acikgoz et al. Hippocrates: An Open-Source Framework for Advancing Large Language Models in Healthcare. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.16621&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zhang et al. Small Language Models Need Strong Verifiers to Self-Correct Reasoning. ACL 2024 Findings. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.17140&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zhou et al. FREB-TQA: A Fine-Grained Robustness Evaluation Benchmark for Table Question Answering. NAACL 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2404.18585&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Xu et al. Large Language Models for Cyber Security: A Systematic Literature Review. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2405.04760&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Dammu et al. &amp;ldquo;They are uncultured&amp;rdquo;: Unveiling Covert Harms and Social Threats in LLM Generated Conversations. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2405.05378&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Yi et al. A safety realignment framework via subspace-oriented model fusion for large language models. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2405.09055&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Lou et al. SPO: Multi-Dimensional Preference Sequential Alignment With Implicit Reward Modeling. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2405.12739&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zhang et al. Getting More from Less: Large Language Models are Good Spontaneous Multilingual Learners. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2405.13816&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zhang et al. TS-Align: A Teacher-Student Collaborative Framework for Scalable Iterative Finetuning of Large Language Models. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2405.20215&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zihong Chen. Sentence Segmentation and Sentence Punctuation Based on XunziALLM. 2024. &lt;a class=&#34;link&#34; href=&#34;https://aclanthology.org/2024.lt4hala-1.30&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[paper]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Gao et al. The Best of Both Worlds: Toward an Honest and Helpful Large Language Model. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.00380&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Wang and Song. MARS: Benchmarking the Metaphysical Reasoning Abilities of Language Models with a Multi-task Evaluation Dataset. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.02106&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Hu et al. Computational Limits of Low-Rank Adaptation (LoRA) for Transformer-Based Models. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.03136&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Ge et al. Time Sensitive Knowledge Editing through Efficient Finetuning. ACL 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.04496&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Tan et al. Peer Review as A Multi-Turn and Long-Context Dialogue with Role-Based Interactions. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.05688&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Song et al. Turbo Sparse: Achieving LLM SOTA Performance with Minimal Activated Parameters. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.05955&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Gu et al. RWKV-CLIP: A Robust Vision-Language Representation Learner. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.06973&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Chen et al. Advancing Tool-Augmented Large Language Models: Integrating Insights from Errors in Inference Trees. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.07115&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zhu et al. Are Large Language Models Good Statisticians?. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.07815&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Li et al. Know the Unknown: An Uncertainty-Sensitive Method for LLM Instruction Tuning. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.10099&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Ding et al. IntentionQA: A Benchmark for Evaluating Purchase Intention Comprehension Abilities of Language Models in E-commerce. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.10173&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;He et al. COMMUNITY-CROSS-INSTRUCT: Unsupervised Instruction Generation for Aligning Large Language Models to Online Communities. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.12074&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Lin et al. FVEL: Interactive Formal Verification Environment with Large Language Models via Theorem Proving. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.14408&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Treutlein et al. Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.14546&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Feng et al. SS-Bench: A Benchmark for Social Story Generation and Evaluation. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.15695&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Feng et al. Self-Constructed Context Decompilation with Fined-grained Alignment Enhancement. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.17233&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Liu et al. Large Language Models for Cuffless Blood Pressure Measurement From Wearable Biosignals. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.18069&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Iyer et al. Exploring Very Low-Resource Translation with LLMs: The University of Edinburgh&amp;rsquo;s Submission to AmericasNLP 2024 Translation Task. AmericasNLP 2024. &lt;a class=&#34;link&#34; href=&#34;https://aclanthology.org/2024.americasnlp-1.25&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[paper]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Li et al. Calibrating LLMs with Preference Optimization on Thought Trees for Generating Rationale in Science Question Scoring. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2406.19949&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Yang et al. Financial Knowledge Large Language Model. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2407.00365&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Lin et al. DogeRM: Equipping Reward Models with Domain Knowledge through Model Merging. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2407.01470&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Bako et al. Evaluating the Semantic Profiling Abilities of LLMs for Natural Language Utterances in Data Visualization. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2407.06129&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Huang et al. RoLoRA: Fine-tuning Rotated Outlier-free LLMs for Effective Weight-Activation Quantization. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2407.08044&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Jiang et al. LLM-Collaboration on Automatic Science Journalism for the General Audience. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2407.09756&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Inouye et al. Applied Auto-tuning on LoRA Hyperparameters. 2024. &lt;a class=&#34;link&#34; href=&#34;https://scholarcommons.scu.edu/cseng_senior/272/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[paper]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Qi et al. Research on Tibetan Tourism Viewpoints information generation system based on LLM. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2407.13561&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Xu et al. Course-Correction: Safety Alignment Using Synthetic Preferences. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2407.16637&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Sun et al. LAMBDA: A Large Model Based Data Agent. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2407.17535&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zhu et al. CollectiveSFT: Scaling Large Language Models for Chinese Medical Benchmark with Collective Instructions in Healthcare. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2407.19705&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Yu et al. Correcting Negative Bias in Large Language Models through Negative Attention Score Alignment. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2408.00137&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Xie et al. The Power of Personalized Datasets: Advancing Chinese Composition Writing for Elementary School through Targeted Model Fine-Tuning. IALP 2024. &lt;a class=&#34;link&#34; href=&#34;https://www.asianlp.sg/conferences/ialp2024/proceedings/papers/IALP2024_P055.pdf&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[paper]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Liu et al. Instruct-Code-Llama: Improving Capabilities of Language Model in Competition Level Code Generation by Online Judge Feedback. ICIC 2024. &lt;a class=&#34;link&#34; href=&#34;https://link.springer.com/chapter/10.1007/978-981-97-5669-8_11&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[paper]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Wang et al. Cybernetic Sentinels: Unveiling the Impact of Safety Data Selection on Model Security in Supervised Fine-Tuning. ICIC 2024. &lt;a class=&#34;link&#34; href=&#34;https://link.springer.com/chapter/10.1007/978-981-97-5669-8_23&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[paper]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Xia et al. Understanding the Performance and Estimating the Cost of LLM Fine-Tuning. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2408.04693&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Zeng et al. Perceive, Reflect, and Plan: Designing LLM Agent for Goal-Directed City Navigation without Instructions. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2408.04168&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Xia et al. Using Pre-trained Language Model for Accurate ESG Prediction. FinNLP 2024. &lt;a class=&#34;link&#34; href=&#34;https://aclanthology.org/2024.finnlp-2.1/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[paper]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Liang et al. I-SHEEP: Self-Alignment of LLM from Scratch through an Iterative Self-Enhancement Paradigm. 2024. &lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2408.08072&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[arxiv]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Bai et al. Aligning Large Language Model with Direct Multi-Preference Optimization for Recommendation. CIKM 2024. &lt;a class=&#34;link&#34; href=&#34;https://dl.acm.org/doi/10.1145/3627673.3679611&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[paper]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/Yu-Yang-Li/StarWhisper&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;StarWhisper&lt;/a&gt;&lt;/strong&gt;: A large language model for Astronomy, based on ChatGLM2-6B and Qwen-14B.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/FudanDISC/DISC-LawLLM&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;DISC-LawLLM&lt;/a&gt;&lt;/strong&gt;: A large language model specialized in Chinese legal domain, based on Baichuan-13B, is capable of retrieving and reasoning on legal knowledge.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/X-D-Lab/Sunsimiao&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Sunsimiao&lt;/a&gt;&lt;/strong&gt;: A large language model specialized in Chinese medical domain, based on Baichuan-7B and ChatGLM-6B.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/WangRongsheng/CareGPT&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;CareGPT&lt;/a&gt;&lt;/strong&gt;: A series of large language models for Chinese medical domain, based on LLaMA2-7B and Baichuan-13B.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/PKU-YuanGroup/Machine-Mindset/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;MachineMindset&lt;/a&gt;&lt;/strong&gt;: A series of MBTI Personality large language models, capable of giving any LLM 16 different personality types based on different datasets and training methods.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Nekochu/Luminia-13B-v3&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Luminia-13B-v3&lt;/a&gt;&lt;/strong&gt;: A large language model specialized in generate metadata for stable diffusion. &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/spaces/Nekochu/Luminia-13B_SD_Prompt&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[demo]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/BUAADreamer/Chinese-LLaVA-Med&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Chinese-LLaVA-Med&lt;/a&gt;&lt;/strong&gt;: A multimodal large language model specialized in Chinese medical domain, based on LLaVA-1.5-7B.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/THUDM/AutoRE&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;AutoRE&lt;/a&gt;&lt;/strong&gt;: A document-level relation extraction system based on large language models.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/NVIDIA/RTX-AI-Toolkit&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;NVIDIA RTX AI Toolkit&lt;/a&gt;&lt;/strong&gt;: SDKs for fine-tuning LLMs on Windows PC for NVIDIA RTX.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/LazyAGI/LazyLLM&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LazyLLM&lt;/a&gt;&lt;/strong&gt;: An easy and lazy way for building multi-agent LLMs applications and supports model fine-tuning via LLaMA Factory.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/NLPJCL/RAG-Retrieval&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;RAG-Retrieval&lt;/a&gt;&lt;/strong&gt;: A full pipeline for RAG retrieval model fine-tuning, inference, and distillation. &lt;a class=&#34;link&#34; href=&#34;https://zhuanlan.zhihu.com/p/987727357&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;[blog]&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/Qihoo360/360-LLaMA-Factory&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;360-LLaMA-Factory&lt;/a&gt;&lt;/strong&gt;: A modified library that supports long sequence SFT &amp;amp; DPO using ring attention.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://novasky-ai.github.io/posts/sky-t1/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Sky-T1&lt;/a&gt;&lt;/strong&gt;: An o1-like model fine-tuned by NovaSky AI with very small cost.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/xming521/WeClone&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;WeClone&lt;/a&gt;&lt;/strong&gt;: One-stop solution for creating your digital avatar from chat logs.&lt;/li&gt;
&lt;/ol&gt;
&lt;/details&gt;
&lt;h2 id=&#34;license&#34;&gt;License
&lt;/h2&gt;&lt;p&gt;This repository is licensed under the &lt;a class=&#34;link&#34; href=&#34;LICENSE&#34; &gt;Apache-2.0 License&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;Please follow the model licenses to use the corresponding model weights: &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/baichuan-inc/Baichuan2-7B-Base/blob/main/Community%20License%20for%20Baichuan%202%20Model.pdf&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Baichuan 2&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/spaces/bigscience/license&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;BLOOM&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://github.com/THUDM/ChatGLM3/blob/main/MODEL_LICENSE&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;ChatGLM3&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://cohere.com/c4ai-cc-by-nc-license&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Command R&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://github.com/deepseek-ai/DeepSeek-LLM/blob/main/LICENSE-MODEL&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;DeepSeek&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/tiiuae/falcon-180B/blob/main/LICENSE.txt&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Falcon&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://ai.google.dev/gemma/terms&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Gemma&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/THUDM/glm-4-9b/blob/main/LICENSE&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;GLM-4&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://github.com/openai/gpt-2/blob/master/LICENSE&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;GPT-2&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;LICENSE&#34; &gt;Granite&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/IndexTeam/Index-1.9B/blob/main/LICENSE&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Index&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://github.com/InternLM/InternLM#license&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;InternLM&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://github.com/facebookresearch/llama/blob/main/MODEL_CARD.md&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://ai.meta.com/llama/license/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama 2&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://llama.meta.com/llama3/license/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama 3&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://github.com/meta-llama/llama-models/blob/main/models/llama4/LICENSE&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Llama 4&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://github.com/OpenBMB/MiniCPM/blob/main/MiniCPM%20Model%20License.md&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;MiniCPM&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;LICENSE&#34; &gt;Mistral/Mixtral/Pixtral&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;LICENSE&#34; &gt;OLMo&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/microsoft/phi-1_5/resolve/main/Research%20License.docx&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Phi-1.5/Phi-2&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/microsoft/Phi-3-mini-4k-instruct/blob/main/LICENSE&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Phi-3/Phi-4&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://github.com/QwenLM/Qwen/blob/main/Tongyi%20Qianwen%20LICENSE%20AGREEMENT&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Qwen&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Skywork/Skywork-13B-base/blob/main/Skywork%20Community%20License.pdf&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Skywork&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/spaces/bigcode/bigcode-model-license-agreement&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;StarCoder 2&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/Tele-AI/telechat-7B/blob/main/TeleChat%E6%A8%A1%E5%9E%8B%E7%A4%BE%E5%8C%BA%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE.pdf&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;TeleChat2&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://github.com/xverse-ai/XVERSE-13B/blob/main/MODEL_LICENSE.pdf&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;XVERSE&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://huggingface.co/01-ai/Yi-6B/blob/main/LICENSE&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Yi&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;LICENSE&#34; &gt;Yi-1.5&lt;/a&gt; / &lt;a class=&#34;link&#34; href=&#34;https://github.com/IEIT-Yuan/Yuan-2.0/blob/main/LICENSE-Yuan&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Yuan 2&lt;/a&gt;&lt;/p&gt;
&lt;h2 id=&#34;citation&#34;&gt;Citation
&lt;/h2&gt;&lt;p&gt;If this work is helpful, please kindly cite as:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;3
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;4
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;5
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;6
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;7
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;8
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;9
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bibtex&#34; data-lang=&#34;bibtex&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nc&#34;&gt;@inproceedings&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;{&lt;/span&gt;&lt;span class=&#34;nl&#34;&gt;zheng2024llamafactory&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;,&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;  &lt;span class=&#34;na&#34;&gt;title&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;s&#34;&gt;{LlamaFactory: Unified Efficient Fine-Tuning of 100+ Language Models}&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;,&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;  &lt;span class=&#34;na&#34;&gt;author&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;s&#34;&gt;{Yaowei Zheng and Richong Zhang and Junhao Zhang and Yanhan Ye and Zheyan Luo and Zhangchi Feng and Yongqiang Ma}&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;,&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;  &lt;span class=&#34;na&#34;&gt;booktitle&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;s&#34;&gt;{Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 3: System Demonstrations)}&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;,&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;  &lt;span class=&#34;na&#34;&gt;address&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;s&#34;&gt;{Bangkok, Thailand}&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;,&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;  &lt;span class=&#34;na&#34;&gt;publisher&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;s&#34;&gt;{Association for Computational Linguistics}&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;,&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;  &lt;span class=&#34;na&#34;&gt;year&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;s&#34;&gt;{2024}&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;,&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;  &lt;span class=&#34;na&#34;&gt;url&lt;/span&gt;&lt;span class=&#34;p&#34;&gt;=&lt;/span&gt;&lt;span class=&#34;s&#34;&gt;{http://arxiv.org/abs/2403.13372}&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;p&#34;&gt;}&lt;/span&gt;
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;h2 id=&#34;acknowledgement&#34;&gt;Acknowledgement
&lt;/h2&gt;&lt;p&gt;This repo benefits from &lt;a class=&#34;link&#34; href=&#34;https://github.com/huggingface/peft&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;PEFT&lt;/a&gt;, &lt;a class=&#34;link&#34; href=&#34;https://github.com/huggingface/trl&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;TRL&lt;/a&gt;, &lt;a class=&#34;link&#34; href=&#34;https://github.com/artidoro/qlora&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;QLoRA&lt;/a&gt; and &lt;a class=&#34;link&#34; href=&#34;https://github.com/lm-sys/FastChat&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;FastChat&lt;/a&gt;. Thanks for their wonderful works.&lt;/p&gt;
&lt;h2 id=&#34;star-history&#34;&gt;Star History
&lt;/h2&gt;&lt;p&gt;&lt;img src=&#34;https://api.star-history.com/svg?repos=hiyouga/LLaMA-Factory&amp;amp;type=Date&#34;
	
	
	
	loading=&#34;lazy&#34;
	
		alt=&#34;Star History Chart&#34;
	
	
&gt;&lt;/p&gt;
</description>
        </item>
        
    </channel>
</rss>
