| | --- |
| | library_name: transformers |
| | base_model: codellama/CodeLlama-7b-Instruct-hf |
| | license: llama2 |
| | datasets: |
| | - semantixai/LloroV3 |
| | language: |
| | - pt |
| | tags: |
| | - code |
| | - analytics |
| | - analise-dados |
| | - portugues-BR |
| |
|
| | co2_eq_emissions: |
| | emissions: 1320 |
| | source: "Lacoste, Alexandre, et al. “Quantifying the Carbon Emissions of Machine Learning.” ArXiv (Cornell University), 21 Oct. 2019, https://doi.org/10.48550/arxiv.1910.09700." |
| | training_type: "fine-tuning" |
| | geographical_location: "Council Bluffs, Iowa, USA." |
| | hardware_used: "1 A100 40GB GPU" |
| | --- |
| | |
| | **Lloro 7B** |
| |
|
| | <img src="https://cdn-uploads.huggingface.co/production/uploads/653176dc69fffcfe1543860a/h0kNd9OTEu1QdGNjHKXoq.png" width="300" alt="Lloro-7b Logo"/> |
| |
|
| | Lloro, developed by Semantix Research Labs , is a language Model that was trained to effectively perform Portuguese Data Analysis in Python. It is a fine-tuned version of codellama/CodeLlama-7b-Instruct-hf, that was trained on synthetic datasets. The fine-tuning process was performed using the QLORA metodology on a GPU A100 with 40 GB of RAM. |
| |
|
| | **Model description** |
| |
|
| | Model type: A 7B parameter fine-tuned on synthetic datasets. |
| |
|
| | Language(s) (NLP): Primarily Portuguese, but the model is capable to understand English as well |
| |
|
| | Finetuned from model: codellama/CodeLlama-7b-Instruct-hf |
| |
|
| | **What is Lloro's intended use(s)?** |
| |
|
| | Lloro is built for data analysis in Portuguese contexts . |
| |
|
| | Input : Text |
| |
|
| | Output : Text (Code) |
| |
|
| | **Usage** |
| |
|
| | Using Transformers |
| |
|
| | ```python |
| | #Import required libraries |
| | import torch |
| | from transformers import ( |
| | AutoModelForCausalLM, |
| | AutoTokenizer |
| | ) |
| | |
| | #Load Model |
| | model_name = "semantixai/LloroV2" |
| | base_model = AutoModelForCausalLM.from_pretrained( |
| | model_name, |
| | return_dict=True, |
| | torch_dtype=torch.float16, |
| | device_map="auto", |
| | ) |
| | |
| | #Load Tokenizer |
| | tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True) |
| | |
| | |
| | #Define Prompt |
| | user_prompt = "Desenvolva um algoritmo em Python para calcular a média e a mediana dos preços de vendas por tipo de material do produto." |
| | system = "Provide answers in Python without explanations, only the code" |
| | prompt_template = f"[INST] <<SYS>>\\n{system}\\n<</SYS>>\\n\\n{user_prompt}[/INST]" |
| | |
| | #Call the model |
| | input_ids = tokenizer([prompt_template], return_tensors="pt")["input_ids"].to("cuda") |
| | |
| | |
| | outputs = base_model.generate( |
| | input_ids, |
| | do_sample=True, |
| | top_p=0.95, |
| | max_new_tokens=1024, |
| | temperature=0.1, |
| | ) |
| | |
| | #Decode and retrieve Output |
| | output_text = tokenizer.batch_decode(outputs, skip_prompt=True, skip_special_tokens=False) |
| | display(output_text) |
| | ``` |
| |
|
| | Using an OpenAI compatible inference server (like [vLLM](https://docs.vllm.ai/en/latest/index.html)) |
| |
|
| | ```python |
| | from openai import OpenAI |
| | |
| | client = OpenAI( |
| | api_key="EMPTY", |
| | base_url="http://localhost:8000/v1", |
| | ) |
| | user_prompt = "Desenvolva um algoritmo em Python para calcular a média e a mediana dos preços de vendas por tipo de material do produto." |
| | completion = client.chat.completions.create(temperature=0.1,frequency_penalty=0.1,model="semantixai/Lloro",messages=[{"role":"system","content":"Provide answers in Python without explanations, only the code"},{"role":"user","content":user_prompt}]) |
| | ``` |
| |
|
| | **Params** |
| | Training Parameters |
| | | Params | Training Data | Examples | Tokens | LR | |
| | |----------------------------------|-----------------------------------|---------------------------------|----------|--------| |
| | | 7B | Pairs synthetic instructions/code | 74222 | 9 351 532| 2e-4 | |
| |
|
| | **Model Sources** |
| |
|
| | Test Dataset Repository: <https://huggingface.co/datasets/semantixai/LloroV3> |
| |
|
| | Model Dates: Lloro was trained between February 2024 and April 2024. |
| |
|
| | **Performance** |
| | | Modelo | LLM as Judge | Code Bleu Score | Rouge-L | CodeBert- Precision | CodeBert-Recall | CodeBert-F1 | CodeBert-F3 | |
| | |----------------|--------------|------------------|---------|----------------------|-----------------|-------------|-------------| |
| | | GPT 3.5 | 94.29% | 0.3538 | 0.3756 | 0.8099 | 0.8176 | 0.8128 | 0.8164 | |
| | | Instruct -Base | 88.77% | 0.3666 | 0.3351 | 0.8244 | 0.8025 | 0.8121 | 0.8052 | |
| | | Instruct -FT | 97.95% | 0.5967 | 0.6717 | 0.9090 | 0.9182 | 0.9131 | 0.9171 | |
| |
|
| | **Training Infos:** |
| | The following hyperparameters were used during training: |
| |
|
| | | Parameter | Value | |
| | |---------------------------|--------------------------| |
| | | learning_rate | 2e-4 | |
| | | weight_decay | 0.0001 | |
| | | train_batch_size | 7 | |
| | | eval_batch_size | 7 | |
| | | seed | 42 | |
| | | optimizer | Adam - paged_adamw_32bit | |
| | | lr_scheduler_type | cosine | |
| | | lr_scheduler_warmup_ratio | 0.06 | |
| | | num_epochs | 4.0 | |
| |
|
| | **QLoRA hyperparameters** |
| | The following parameters related with the Quantized Low-Rank Adaptation and Quantization were used during training: |
| |
|
| | | Parameter | Value | |
| | |------------------|-----------| |
| | | lora_r | 64 | |
| | | lora_alpha | 256 | |
| | | lora_dropout | 0.1 | |
| | | storage_dtype | "nf4" | |
| | | compute_dtype | "bfloat16"| |
| | |
| | **Experiments** |
| | | Model | Epochs | Overfitting | Final Epochs | Training Hours | CO2 Emission (Kg) | |
| | |-----------------------|--------|-------------|--------------|-----------------|-------------------| |
| | | Code Llama Instruct | 1 | No | 1 | 3.01 | 0.43 | |
| | | Code Llama Instruct | 4 | Yes | 3 | 9.25 | 1.32 | |
| | |
| | **Framework versions** |
| | |
| | | Library | Version | |
| | |---------------|-----------| |
| | | bitsandbytes | 0.40.2 | |
| | | Datasets | 2.14.3 | |
| | | Pytorch | 2.0.1 | |
| | | Tokenizers | 0.14.1 | |
| | | Transformers | 4.34.0 | |
| | |