alfazick's picture
Update README.md
369095e verified
---
license: llama3.1
base_model: meta-llama/Llama-3.1-8B-Instruct
tags:
- function-calling
- tool-use
- fine-tuned
- llama
datasets:
- Salesforce/xlam-function-calling-60k
pipeline_tag: text-generation
---
# Llama 3.1 8B Function Calling
Fine-tuned [Llama 3.1 8B Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) for function/tool calling.
## Training
- **Dataset:** 900 examples from [Salesforce/xlam-function-calling-60k](https://huggingface.co/datasets/Salesforce/xlam-function-calling-60k)
- **Method:** LoRA
- **Trainable params:** 42M / 8B (0.52%)
- **Epochs:** 1
- **Loss:** 0.66 → 0.63
## Evaluation (100 held-out samples)
- **Exact match:** 62%
- **Function name accuracy:** ~90%+
## Usage
```python
from vllm import LLM
llm = LLM(model="alfazick/llama-3.1-8b-function-calling")
```
Or with transformers:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("alfazick/llama-3.1-8b-function-calling")
tokenizer = AutoTokenizer.from_pretrained("alfazick/llama-3.1-8b-function-calling")
```
## Prompt Format
```
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
You are a helpful assistant with access to the following tools or function calls. Your task is to produce a sequence of tools or function calls necessary to generate response to the user utterance. Use the following tools or function calls as required:
[{"name": "func_name", "description": "...", "parameters": {...}}]<|eot_id|><|start_header_id|>user<|end_header_id|>
{query}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
```
## Output Format
```json
[{"name": "function_name", "arguments": {"arg": "value"}}]
```
## Limitations
- Trained on 900 examples (proof of concept)
- May have argument variations vs ground truth
- Best for single/simple tool calls
## Training Details
- **Framework:** Unsloth 2025.11.2 + TRL
- **Hardware:** RTX 5090 (32GB)
- **Method:** LoRA (r=16, alpha=16)