| | --- |
| | license: llama3.1 |
| | language: |
| | - en |
| | - de |
| | - fr |
| | - it |
| | - pt |
| | - hi |
| | - es |
| | - th |
| | pipeline_tag: text-generation |
| | library_name: transformers |
| | tags: |
| | - llama3.1-5B |
| | - llama-3 |
| | - Base_Ft |
| | - facebook |
| | - text-generation-inference |
| | - meta |
| | - ollama |
| | --- |
| | # **Llama-3.1-5B-Instruct** |
| |
|
| | Llama-3.1 is a collection of multilingual large language models (LLMs) that includes pretrained and instruction-tuned generative models in various sizes. The **Llama-3.1-5B-Instruct** model is part of the series optimized for multilingual dialogue use cases, offering powerful conversational abilities and outperforming many open-source and closed chat models on key industry benchmarks. |
| |
|
| | ## Model Overview |
| |
|
| | - **Size**: 5B parameters |
| | - **Model Architecture**: Llama-3.1 is an auto-regressive language model using an optimized transformer architecture. |
| | - **Training**: The model is fine-tuned using Supervised Fine-Tuning (SFT) and Reinforcement Learning with Human Feedback (RLHF) to align with human preferences, ensuring helpfulness, safety, and natural conversations. |
| |
|
| | The **Llama-3.1-5B-Instruct** model is optimized for multilingual text generation and excels in a variety of dialog-based use cases. It is designed to handle a wide array of tasks, including question answering, translation, and instruction following. |
| |
|
| | ## How to Use |
| |
|
| | ### Requirements |
| |
|
| | - Install the latest version of **Transformers**: |
| | ```bash |
| | pip install --upgrade transformers |
| | ``` |
| |
|
| | - Ensure you have **PyTorch** installed with support for `bfloat16`: |
| | ```bash |
| | pip install torch |
| | ``` |
| |
|
| | ### Example Code |
| |
|
| | Below is an example of how to use the **Llama-3.1-5B-Instruct** model for conversational inference: |
| |
|
| | ```python |
| | import transformers |
| | import torch |
| | |
| | # Define the model ID |
| | model_id = "prithivMLmods/Llama-3.1-5B-Instruct" |
| | |
| | # Set up the pipeline for text generation |
| | pipeline = transformers.pipeline( |
| | "text-generation", |
| | model=model_id, |
| | model_kwargs={"torch_dtype": torch.bfloat16}, |
| | device_map="auto", # Use the best device available |
| | ) |
| | |
| | # Define conversation messages |
| | messages = [ |
| | {"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"}, |
| | {"role": "user", "content": "Who are you?"}, |
| | ] |
| | |
| | # Generate a response |
| | outputs = pipeline( |
| | messages, |
| | max_new_tokens=256, |
| | ) |
| | |
| | # Print the generated response |
| | print(outputs[0]["generated_text"][-1]) |
| | ``` |
| |
|
| | ### Model Details |
| |
|
| | - **Model Type**: Instruction-Tuned Large Language Model (LLM) |
| | - **Training**: Trained using supervised fine-tuning and reinforcement learning with human feedback. |
| | - **Supported Tasks**: Dialogue generation, question answering, translation, and other text-based tasks. |
| |
|
| | ### Performance |
| |
|
| | The **Llama-3.1-5B-Instruct** model outperforms many existing models on several benchmarks, making it a reliable choice for conversational AI tasks in multilingual environments. |
| |
|
| | ### Notes |
| |
|
| | - This model is optimized for safety and helpfulness, ensuring a positive user experience. |
| | - The **torch_dtype** is set to `bfloat16` to optimize memory usage and performance. |
| | --- |