Llama 3.1 is out! 23rd of July, facebook announced the next iteration of the Llama family to Hugging Face. It’s exciting for hugging face to collaborate with Meta to ensure the best integration in the Hugging Face ecosystem. Eight open-weight models (3 base models and 5 fine-tuned ones) are available on the Hub. In this blog, we will keep you informed on everything you need to know concerning the collaboration of these giants in the Natural language processing domain.
Llama 3.1 comes in three sizes: 8B for efficient deployment and development on consumer-size GPU, 70B for large-scale AI native applications, and 405B for synthetic data, LLM as a Judge or distillation. All three come in base and instruction-tuned variants.
In addition to the six generative models, Meta released two new models: Llama Guard 3 and Prompt Guard. Prompt Guard is a small classifier that detects prompt injections and jailbreaks. Llama Guard 3 is a safeguard model that can classify LLM inputs and generations.
Among the features and integrations being released, we have:
- Models on the Hub
- Hugging Face Transformers and TGI integration
- Hugging Chat integration for Meta Llama 3.1 405B Instruct
- Inference & Deployment Integration with Inference Endpoints, Google Cloud, Amazon SageMaker & DELL Enterprise Hub
- Quantization for FP8, AWQ and GPTQ for easier inference
- Fine-tuning Llama 3.1 8B on a single GPU with 🤗 TRL
- Generate synthetic data using Llama 3.1 70B and 405B with Distilabel
Table of contents
- What’s new with Llama 3.1?
- How much memory does Llama 3.1 need?
- Llama 3.1 evaluation
- Using Hugging Face Transformers
- How to prompt Llama 3.1
- Custom Tool calling
- Demo
- Llama 3.1 405B quantization with FP8, AWQ, and GPTQ
- Inference Integrations
- Hugging Face Partner Integrations
- Fine-tuning with Hugging Face TRL
- Synthetic data generation with distilabel
- Additional Resources
- Acknowledgments
What’s new with Llama 3.1?
Llama 3.1 is an exciting update that builds upon the features of its predecessor, introducing several key enhancements:
- Large Context Length: An impressive increase to 128K tokens (up from 8K) allowing for much longer interactions and document processing.
- Multilingual Capabilities: Support for 8 languages, including English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.
- Tool Usage Capabilities: Enhanced functionality with tool integration for more complex tasks.
- Large Dense Model: A massive model with 405 billion parameters.
- More Permissive License: An updated license that allows for greater flexibility in how the models can be used.
Let’s dive into these!
The Llama 3.1 release introduces six new open LLM models based on the Llama 3 architecture, available in three sizes: 8B, 70B, and 405B parameters, each with base (pre-trained) and instruct-tuned versions. All the variants support a context length of 128K tokens and 8 languages, including English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai. Llama 3.1 continues to use Grouped-Query Attention (GQA), an efficient representation that should help with longer contexts.
- Meta-Llama-3.1-8B: Base 8B model
- Meta-Llama-3.1-8B-Instruct: Instruct fine-tuned version of the base 8B model
- Meta-Llama-3.1-70B: Base 70B model
- Meta-Llama-3.1-70B-Instruct: Instruct fine-tuned version of the base 70B model
- Meta-Llama-3.1-405B: Base 405B model
- Meta-Llama-3.1-405B-Instruct: Instruct fine-tuned version of the base 405B model
Additional Releases
In addition to these six models, two new tools have been introduced:
- Llama Guard 3 is the latest iteration in the Llama Guard family, fine-tuned on Llama 3.1 8B. It is designed for production use cases, with a 128k context length and multilingual capabilities. Llama Guard 3 can classify LLM inputs (prompts) and responses to identify content that would be considered unsafe in a risk taxonomy.
- Prompt Guard, on the other hand, is a small 279M parameter BERT-based classifier engineered to detect prompt injection and jailbreaking attempts. It was trained on a large corpus of attacks and it is suggested to be fine-tuned with application-specific datasets for enhanced accuracy.
New in Llama 3.1 compared to Llama 3 is that the instruct models are optimized for tool calling, catering to agentic use cases. It includes two built-in tools (search functionality, mathematical reasoning with Wolfram Alpha) that can be expanded with custom JSON functions.
The Llama 3.1 models were trained on over 15 trillion tokens using a custom-built GPU cluster with a total of 39.3M GPU hours (1.46M for 8B, 7.0M for 70B, 30.84M for 405B). While details about the training dataset mix are not fully disclosed, it is speculated that the dataset has a more diverse curation to enhance multilingual performance. Llama 3.1 Instruct has been specifically optimized for instruction following with data drawn from publicly available instruction datasets and over 25M synthetically generated examples through supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF). Meta developed LLM-based classifiers to filter and curate high-quality prompts and responses during the creation of the data mix.
Regarding the licensing terms, Llama 3.1 comes with a similar license to its predecessor, with one significant change: it enables using model outputs that can be used to improve other LLMs. This means that synthetic data generation and distillation are allowed, even with different models! This is particularly important for the 405B model, as discussed later. The license also provides for redistribution, fine-tuning, and the creation of derivative work and still requires derived models to include “Llama” at the beginning of their names, and any derivative works or services must mention “Built with Llama”. For full details, it is essential to read the official license.
How much memory does Llama 3.1 need?
Llama 3.1 introduces exciting advancements, but it’s important to consider your hardware resources when running it. We broke down the memory requirements for both training and inference across the three model sizes.
Inference Memory Requirements
For inference, the memory requirements depend on the model size and the precision of the weights. Below is a table outlining the approximate memory needed for various configurations:
Model Size | FP16 | FP8 | INT4 |
---|---|---|---|
8B | 16 GB | 8 GB | 4 GB |
70B | 140 GB | 70 GB | 35 GB |
405B | 810 GB | 405 GB | 203 GB |
Note: The above-quoted numbers indicate the GPU VRAM required just to load the model checkpoint. They don’t include torch reserved space for kernels or CUDA graphs.
For instance, an H100 node (of 8x H100) has ~640GB of VRAM, so the 405B model would need to be run in a multi-node setup or run at a lower precision (e.g. FP8), which would be the recommended approach.
It’s important to remember that lower precision (e.g., INT4) might result in some loss of accuracy but it can significantly reduce memory requirements and increase inference speed. In addition to the model weights, you will also need to keep the KV Cache in memory, which holds the keys and values of all the tokens in the model’s context such that they don’t need to be recomputed when generating a new token. Especially when making use of the long available context length, it becomes a significant factor. In FP16, the KV cache memory requirements are:
Model Size | 1k tokens | 16k tokens | 128k tokens |
---|---|---|---|
8B | 0.125 GB | 1.95 GB | 15.62 GB |
70B | 0.313 GB | 4.88 GB | 39.06 GB |
405B | 0.984 GB | 15.38 | 123.05 GB |
Especially for the small model the cache uses as much memory as the weights when approaching the context length maximum.
Training Memory Requirements
The table below outlines the approximate memory requirements for training Llama 3.1 models using various techniques:
Model Size | Full Fine-tuning | LoRA | Q-LoRA |
---|---|---|---|
8B | 60 GB | 16 GB | 6 GB |
70B | 300 GB | 160 GB | 48 GB |
405B | 3.25 TB | 950 GB | 250 GB |
Note: These are estimated values and may vary based on specific implementation details and optimizations.
Llama 3.1 evaluation
Note: We are currently evaluating Llama 3.1 individually on the new Open LLM Leaderboard and will update this section later today. Below is an excerpt from the official evaluation from Meta.
Category | Benchmark | # Shots | Metric | Llama 3 8B | Llama 3.1 8B | Llama 3 70B | Llama 3.1 70B | Llama 3.1 405B |
---|---|---|---|---|---|---|---|---|
General | MMLU | 5 | macro_avg/acc_char | 66.7 | 66.7 | 79.5 | 79.3 | 85.2 |
MMLU PRO (CoT) | 5 | macro_avg/acc_char | 36.2 | 37.1 | 55.0 | 53.8 | 61.6 | |
AGIEval English | 3-5 | average/acc_char | 47.1 | 47.8 | 63.0 | 64.6 | 71.6 | |
CommonSenseQA | 7 | acc_char | 72.6 | 75.0 | 83.8 | 84.1 | 85.8 | |
Winogrande | 5 | acc_char | – | 60.5 | – | 83.3 | 86.7 | |
BIG-Bench Hard (CoT) | 3 | average/em | 61.1 | 64.2 | 81.3 | 81.6 | 85.9 | |
ARC-Challenge | 25 | acc_char | 79.4 | 79.7 | 93.1 | 92.9 | 96.1 | |
Knowledge reasoning | TriviaQA-Wiki | 5 | em | 78.5 | 77.6 | 89.7 | 89.8 | 91.8 |
SQuAD | 1 | em | 76.4 | 77.0 | 85.6 | 81.8 | 89.3 | |
Reading comprehension | QuAC (F1) | 1 | f1 | 44.4 | 44.9 | 51.1 | 51.1 | 53.6 |
BoolQ | 0 | acc_char | 75.7 | 75.0 | 79.0 | 79.4 | 80.0 | |
DROP (F1) | 3 | f1 | 58.4 | 59.5 | 79.7 | 79.6 | 84.8 |
Using Hugging Face Transformers
Llama 3.1 requires a minor adjustment in modeling to effectively handle RoPE scaling. With Transformers release 4.43, you can utilize the new Llama 3.1 models and take advantage of all the tools within the Hugging Face ecosystem. Ensure to use the latest transformers release:
pip install “transformers>=4.43” –upgrade |
---|
A couple of details:
- Transformers loads the model in bfloat16 by default. This is the format used by the original checkpoint published by Meta, so it’s the recommended way to ensure optimal precision for evaluations.
- Assistant responses may end with the special token <|eot_id|>, but generation should also stop if the regular EOS token is encountered. We can stop generation early by providing a list of terminators in the eos_token_id parameter.
- We used the default sampling parameters (temperature and top_p) taken from the original meta codebase. While extensive testing hasn’t been conducted yet, feel free to experiment!
The following snippet shows how to use meta-llama/Meta-Llama-3.1-8B-Instruct. It requires about 16 GB of VRAM, which fits many consumer GPUs. The same snippet works for meta-llama/Meta-Llama-3.1-70B-Instruct“, which, at 140GB of VRAM & meta-llama/Meta-Llama-3.1-405B-Instruct“ (requiring 810GB VRAM), makes it a very interesting model for production use cases.You can further reduce memory consumption by loading the model in 8-bit or 4-bit mode.
from transformers import pipelineimport torchmodel_id = “meta-llama/Meta-Llama-3.1-8B-Instruct”
pipe = pipeline( “text-generation”, model=model_id, model_kwargs={“torch_dtype”: torch.bfloat16}, device=”cuda”, ) messages = [ {“role”: “user”, “content”: “Who are you? Please, answer in pirate-speak.”}, ] outputs = pipe( messages, max_new_tokens=256, do_sample=False, ) assistant_response = outputs[0][“generated_text”][-1][“content”] print(assistant_response) # Arrrr, me hearty! Yer lookin’ fer a bit o’ information about meself, eh? Alright then, matey! I be a language-generatin’ swashbuckler, a digital buccaneer with a penchant fer spinnin’ words into gold doubloons o’ knowledge! Me name be… (dramatic pause)…Assistant! Aye, that be me name, and I be here to help ye navigate the seven seas o’ questions and find the hidden treasure o’ answers! So hoist the sails and set course fer adventure, me hearty! What be yer first question? |
---|
You can also automatically quantize the model, loading it in 8-bit or even 4-bit mode with bitsandbytes. 4-bit loading of the large 70B version takes about 34 GB of memory to run. This is how you’d load the generation pipeline in 4-bit:
pipeline = pipeline(“text-generation”,model=model_id,
model_kwargs={ “torch_dtype”: torch.bfloat16, “quantization_config”: {“load_in_4bit”: True} }, ) |
---|
For further information on using the models with transformers, please check the model cards.
Note: Transformers takes care of all pesky prompt template issues and more, if you want to know more about prompting then check out the next section.
How to prompt Llama 3.1
The base models have no prompt format. Similar to other base models, they can be utilized to extend an input sequence with a plausible continuation or for zero-shot/few-shot inference. Moreover, they serve as an excellent foundation for fine-tuning to suit your specific use cases.
The Instruct versions support conversational format with 4 distinct roles:
- system: This role sets the context for the conversation. It allows you to include rules, guidelines, or any necessary information that helps the model respond effectively. It’s also used to enable tool usage when appropriate.
- user: This role represents the input, commands, and questions from the user directed to the model.
- assistant: This role represents the assistant’s response based on the context provided in the ‘system’ and ‘user’ prompts.
- ipython: A new role introduced in Llama 3.1. This role is used as the output of a tool call when sent back to the LLM.
The Instruct versions use the following conversation structure for simple conversations:
<|begin_of_text|><|start_header_id|>system<|end_header_id|>{{ system_prompt }}<|eot_id|><|start_header_id|>user<|end_header_id|>{{ user_msg_1 }}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
{{ model_answer_1 }}<|eot_id|> |
---|
Llama 3.1 Instruct models now support tool calling, including three built-in tools (brave_search, wolfram_alpha, and code_interpreter) and custom tool calling via JSON function calling. The built-in tools use Python syntax. The ability to output Python code for function calling is part of the code interpreter tool, which must be enabled in the system prompt using the Environment keyword, as shown below
Built-in Tool calling
Including “Environment: ipython” turns on the code interpreter mode, and the model can generate Python code that it expects to be executed. The message body of the assistant response starts with a special tag <|python_tag|> and ends with <|eom_id|> instead of just the standard <|eot_id|>. The latter indicates the turn is finished, while the former indicates continued multi-step reasoning.
Built-in tool calling example
<|begin_of_text|><|start_header_id|>system<|end_header_id|>Environment: ipythonTools: brave_search, wolfram_alpha
Cutting Knowledge Date: 01 March 2023 Today’s Date: 13 July 2024 You are a helpful Assistant.<|eot_id|><|start_header_id|>user<|end_header_id|> Weather in Menlo Park, California<|eot_id|><|start_header_id|>assistant<|end_header_id|> |
---|
The response from the model at this point would include Python code to call one of the supported tools (brave_search in this case):
<|python_tag|>brave_search.call(query=”current weather in Menlo Park, California”)<|eom_id|> |
---|
The response from executing the call is then sent back to the model to retrieve the final response. For brevity, the following would be appended to the message shown in the previous snippet:
<|python_tag|>brave_search.call(query=”Menlo Park California weather”)<|eom_id|><|start_header_id|>ipython<|end_header_id|>{“query”: “Menlo Park California weather”, “top_k”: [{“title”: “10-Day Weather Forecast for West Menlo Park, CA – The Weather Channel | weather.com”, “url”: “https://weather.com/weather/tenday/l/West+Menlo+Park+CA?canonicalCityId=b2375713aa1943aad7d1a13a85e1c0adad13c1b10563b2bbaad70734dc61cf11”, “description”: “Be prepared with the most accurate 10-day forecast for West <strong>Menlo</strong> <strong>Park</strong>, CA with highs, lows, chance of precipitation from The <strong>Weather</strong> Channel and <strong>Weather</strong>.com”, “type”: “search_result”},….}<|eot_id|><|start_header_id|>assistant<|end_header_id|> |
---|
The final response from the LLM would then be:
The current weather in Menlo Park, California is mostly sunny with a high of 77°F and a low of 56°F.<|eot_id|> |
---|
Custom Tool calling
Llama 3.1 Instruct supports custom function calls from a single user message. The following prompts provide an example of how custom functions can be called from the output of the model. In custom function calling, the model outputs <|eot_id|> instead of <|eom_id|>. The system prompt needs to be adjusted to inform the model how to deal with function call outputs.
Custom Tool Calling JSON Functions
<|begin_of_text|><|start_header_id|>system<|end_header_id|>You are a helpful assistant with tool calling capabilities. When you receive a tool call response, use the output to format an answer to the orginal user question.<|eot_id|><|start_header_id|>user<|end_header_id|>Given the following functions, please respond with a JSON for a function call with its proper arguments that best answers the given prompt.
Respond in the format {“name”: function name, “parameters”: dictionary of argument name and its value}. Do not use variables. { “type”: “function”, “function”: { “name”: “get_current_conditions”, “description”: “Get the current weather conditions for a specific location”, “parameters”: { “type”: “object”, “properties”: { “location”: { “type”: “string”, “description”: “The city and state, e.g., San Francisco, CA” }, “unit”: { “type”: “string”, “enum”: [“Celsius”, “Fahrenheit”], “description”: “The temperature unit to use. Infer this from the user’s location.” } }, “required”: [“location”, “unit”] } } } Question: what is the weather like in Menlo Park?<|eot_id|><|start_header_id|>assitant<|end_header_id|> {“name”: “get_current_conditions”, “parameters”: {“location”: “Menlo Park, CA”, “unit”: “Fahrenheit”}}<|eot_id|><|start_header_id|>ipython<|end_header_id|> |
---|
When we retrieve the output from the selected tool, we pass it back to the model using the same <|python_tag|> delimiter. <|python_tag|> does not imply Python use. It’s only meant to signal the beginning of outputs from any tool.
<|python_tag|>{“tool_call_id”: “get_current_conditions””output”: “Clouds giving way to sun Hi: 76° Tonight: Mainly clear early, then areas of low clouds forming Lo: 56°”
}<|eot_id|><|start_header_id|>assistant<|end_header_id|> The weather in Menlo Park is currently cloudy with a high of 76° and a low of 56°, with clear skies expected tonight.<|eot_id|> |
---|
This format has to be exactly reproduced for effective use. The chat template available in transformers makes it straightforward to format the prompt correctly.
Demo
You can experiment with the three Instruct models of Llama using the following demos:
- Hugging Chat with Llama 3.1 405B https://huggingface.co/chat/models/meta-llama/Meta-Llama-3.1-405b-instruct/
- Hugging Chat with Llama 3.1 70B https://huggingface.co/chat/models/meta-llama/Meta-Llama-3.1-70b-instruct/
- Gradio-powered Space with Llama 3.1 8B demo https://huggingface.co/spaces/ysharma/Chat_with_Meta_llama3_1_8b
The whole stack is open-source. Hugging Chat is powered by chat-ui and text-generation-inference.
Llama 3.1 405B quantization with FP8, AWQ, and GPTQ
Meta created an official FP8 quantized version of Llama 3.1 405B with minimal accuracy degradation. To achieve this, FP8 quantization was only applied to the major linear operators of the model, such as the gate and up and down projections for the FFNs (covering 75% of the inference FLOPs). We worked together to ensure that this FP8 quantization checkpoint is compatible across the community (transformers, TGI, VLLM).
Additionally, we created AWQ and GPTQ quantized variants in INT4 with AutoAWQ and AutoGPTQ, respectively. For AWQ, all the linear layers were quantized using the GEMM kernels performing zero-point quantization down to 4 bits with a group size of 128; and for GPTQ the same setting only using the GPTQ kernels instead. We ensured that the INT4 checkpoints are compatible with transformers and TGI, including Marlin kernel support to speed up inference in TGI for the GPTQ quants.
Available quantized weights for Llama 3.1 405B:
- meta-llama/Meta-Llama-3.1-405B-Base-FP8: Official FP8 quantized weights, can be run on 8xH100
- meta-llama/Meta-Llama-3.1-405B-Instruct-FP8: Official FP8 quantized weights, can be run on 8xH100
- hugging-quants/Meta-Llama-3.1-405B-Instruct-AWQ-INT4: Hugging Face quantized weights, can run on 8xA100 80GB, 8xH100 80GB & 8xA100 40GB (with a reduced KV-cache and without CUDA graphs)
- hugging-quants/Meta-Llama-3.1-405B-Instruct-GPTQ-INT4: Hugging Face quantized weights, can run on 8xA100 80GB, 8xH100 80GB & 8xA100 40GB (with a reduced KV-cache and without CUDA graphs)
- hugging-quants/Meta-Llama-3.1-405B-BNB-NF4: Hugging Face quantized weights, suitable for QLoRA finetuning
- hugging-quants/Meta-Llama-3.1-405B-Instruct-BNB-NF4: Hugging Face quantized weights, suitable for inference on 8xA100 & 4xH100
The Hugging Quants organization contains quantized checkpoints for the 70B and 8B version as well.
Inference Integrations
Hugging Face Inference API
Hugging Face PRO users now have access to exclusive API endpoints hosting Llama 3.1 8B Instruct, Llama 3.1 70B Instruct and Llama 3.1 405B Instruct AWQ powered by text-generation-inference. All versions support the Messages API, so they are compatible with OpenAI client libraries, including LangChain and LlamaIndex.
Note: Update to the latest huggingface_hub version with pip install “huggingface_hub>=0.24.1
from huggingface_hub import InferenceClient# Initialize the client, pointing it to one of the available modelsclient = InferenceClient()
chat_completion = client.chat.completions.create( model=”meta-llama/Meta-Llama-3.1-405B-Instruct-FP8″, messages=[ {“role”: “system”, “content”: “You are a helpful an honest programming assistant.”}, {“role”: “user”, “content”: “Is Rust better than Python?”}, ], stream=True, max_tokens=500 ) # iterate and print stream for message in chat_completion: print(message.choices[0].delta.content, end=””) |
---|
For more details about the use of the Messages API, please check this post.
Hugging Face Inference Endpoints
You can deploy Llama 3.1 on Hugging Face’s Inference Endpoints, which uses Text Generation Inference as the backend. Text Generation Inference is a production-ready inference container developed by Hugging Face with support for FP8, continuous batching, token streaming, tensor parallelism for fast inference on multiple GPUs. To deploy Llama 3.1, go to the model page and click on the Deploy -> Inference Endpoints widget:
- Meta-Llama-3.1-8B-Instruct is recommended on 1x NVIDIA A10G or L4 GPUs
- Meta-Llama-3.1-70B-Instruct is recommended on 4x NVIDIA A100 or as AWQ/GPTQ quantized on 2x A100s
- Meta-Llama-3.1-405B-Instruct-FP8 is recommended on 8x NVIDIA H100 in FP or as AWQ/GPTQ quantized on 8x A100s
from huggingface_hub import InferenceClient# Initialize the client, pointing it to one of the available modelsclient = InferenceClient(
base_url=”<ENDPOINT_URL>”, ) # Create a chat completion chat_completion = client.chat.completions.create( model=”ENDPOINT”, messages=[ {“role”: “system”, “content”: “You are a helpful an honest programming assistant.”}, {“role”: “user”, “content”: “Is Rust better than Python?”}, ], stream=True, max_tokens=500 ) # iterate and print stream for message in chat_completion: print(message.choices[0].delta.content, end=””) |
---|
Hugging Face Partner Integrations
Note: We are currently working with our partners at AWS, Google Cloud, Microsoft Azure and DELL on adding Llama 3.1 8B, 70B, and 405B to Amazon SageMaker, Google Kubernetes Engine, Vertex AI Model Catalog, Azure AI Studio, DELL Enterprise Hub. We will update this section as soon as the containers are available – you can subscribe to Hugging Squad for email updates.
Fine-tuning with Hugging Face TRL
In this section, we’ll look at the tools available in the Hugging Face ecosystem to efficiently train Llama 3.1 on consumer-size GPUs. An example command to fine-tune Llama 3.1 8B on OpenAssistant’s chat dataset can be found below. We use 4-bit quantization and QLoRA to conserve memory to target all the attention blocks’ linear layers.
Fine-Tuning Example with Hugging Face TRL
First, install the nightly version of 🤗 TRL and clone the repo to access the training script:
pip install “transformers>=4.43.2” –upgradepip install –upgrade bitsandbytespip install –ugprade peft
pip install git+https://github.com/huggingface/trl git clone https://github.com/huggingface/trl cd trl |
---|
Then you can run the script:
python \examples/scripts/sft.py \–model_name meta-llama/Meta-Llama-3.1-8B \
–dataset_name OpenAssistant/oasst_top1_2023-08-25 \ –dataset_text_field=”text” \ –per_device_train_batch_size 1 \ –per_device_eval_batch_size 1 \ –gradient_accumulation_steps 4 \ –learning_rate 2e-4 \ –report_to “none” \ –bf16 \ –max_seq_length 1024 \ –lora_r 16 –lora_alpha 32 \ –lora_target_modules q_proj k_proj v_proj o_proj \ –load_in_4bit \ –use_peft \ –attn_implementation “flash_attention_2” \ –logging_steps=10 \ –gradient_checkpointing \ –output_dir llama31 |
---|
If you have more GPUs to spare, you can run training with DeepSpeed and ZeRO Stage 3:
accelerate launch –config_file=examples/accelerate_configs/deepspeed_zero3.yaml \examples/scripts/sft.py \–model_name meta-llama/Meta-Llama-3.1-8B \
–dataset_name OpenAssistant/oasst_top1_2023-08-25 \ –dataset_text_field=”text” \ –per_device_train_batch_size 1 \ –per_device_eval_batch_size 1 \ –gradient_accumulation_steps 4 \ –learning_rate 2e-5 \ –report_to wandb \ –bf16 \ –max_seq_length 1024 \ –attn_implementation eager \ –logging_steps=10 \ –gradient_checkpointing \ –output_dir models/llama |
---|
Synthetic data generation with distilabel
A significant update in Llama 3.1’s license permits the use of model outputs to improve other LLMs, which means you can generate synthetic datasets with Llama 3.1 models and use them to fine-tune smaller, more specialized models.
Let’s look at an example of how to generate a preference dataset with distilabel, an open-source framework for synthetic data generation. This dataset can be used to fine-tune models with the preference optimization methods offered by TRL like DPO or KTO.
First install the latest distilabel release including the hf-inference-endpoints extra with pip as follows:
pip install “distilabel[hf-inference-endpoints]” –upgrade |
---|
Then define a pipeline that:
- loads a dataset with instructions from the Hugging Face Hub.
- generates a response with Llama 3.1 70B Instruct and Llama 3.1 405B Instruct via Hugging Face Inference Endpoints.
- finally, uses Llama 3.1 405B Instruct as a judge to rate the responses using UltraFeedback prompts. From these ratings, chosen and rejected responses can be selected and used to fine-tune a model with preference optimization methods.
See the code below to define the pipeline or run it yourself using this Colab notebook and
explore the generated dataset in the Hub.from distilabel.llms import InferenceEndpointsLLMfrom distilabel.pipeline import Pipeline
from distilabel.steps import LoadDataFromHub, CombineColumns from distilabel.steps.tasks import TextGeneration, UltraFeedback llama70B = InferenceEndpointsLLM( model_id=”meta-llama/Meta-Llama-3.1-70B-Instruct” ) llama405B = InferenceEndpointsLLM( model_id=”meta-llama/Meta-Llama-3.1-405B-Instruct-FP8″ ) with Pipeline(name=”synthetic-data-with-llama3″) as pipeline: # load dataset with prompts load_dataset = LoadDataFromHub( repo_id=”argilla/10Kprompts-mini” ) # generate two responses for each prompt generate = [ TextGeneration(llm=llama70B), TextGeneration(llm=llama405B) ] # combine responses into one column combine = CombineColumns( columns=[“generation”, “model_name”], output_columns=[“generations”, “model_names”] ) # rate responses with 405B LLM-as-a-judge rate = UltraFeedback(aspect=”overall-rating”, llm=llama405B) # define the pipeline load_dataset >> generate >> combine >> rate if __name__ == “__main__”: distiset = pipeline.run() |
---|
What’s next? Besides the example above, distilabel comes with exciting approaches for synthetic data generation with LLMs in a wide range of scenarios and topics. It includes implementations from the current SOTA literature for tasks like evaluating outputs with LLM-as-a-judge methods, evolving instructions, data filtering, as well as defining custom components.