Nexa_0.1_First / README.md
ItsMatti4's picture
Update README.md
a346e76 verified
---
license: apache-2.0
base_model:
- meta-llama/Llama-3.2-1B
---
# Nexa 0.1 First - The Genesis
**Release Date:** March 29, 2026
**Project Phase:** Alpha / Prototype
**Developer:** WeAreNexa
---
## Overview
**Nexa 0.1 First** marks the beginning of the Nexa series. It is a compact, high-performance **Text-to-Text** language model with **1.2 Billion parameters**, designed to demonstrate the efficiency of Low-Rank Adaptation (LoRA) on edge-compatible architectures.
This model was specifically fine-tuned to handle Italian and English instructions with low latency, serving as the foundational proof-of-concept for the subsequent Enhanced and Ultra versions.
## Technical Specifications
| Feature | Specification |
| :--- | :--- |
| **Model Architecture** | Llama 3.2 (Dense) |
| **Parameter Count** | 1.24 Billion |
| **Task Type** | Text-to-Text Generation |
| **Base Model** | unsloth/Llama-3.2-1B-Instruct-bnb-4bit |
| **Quantization** | 4-bit NormalFloat (NF4) via bitsandbytes |
| **Context Length** | 2048 Tokens |
| **Language Support** | Italian, English |
## Fine-Tuning Details (PEFT/LoRA)
* **Method:** LoRA (Low-Rank Adaptation)
* **Rank (r):** 16
* **Alpha:** 16
* **Target Modules:** `q_proj`, `k_proj`, `v_proj`, `o_proj`
* **Dataset:** Alpaca-GPT4 (Filtered subset for logical consistency)
* **Epochs:** 1
* **Optimizer:** AdamW 8-bit
## Lineage & Evolution
1. **Nexa 0.1 First (Current):** Initial alpha release focusing on base instruction following.
2. **Nexa 0.2 Enhanced:** Optimization of response fluidity and broader vocabulary.
3. **Nexa 0.3 Enhanced:** High-density tuning (Rank 128) for advanced reasoning.
4. **Nexa 0.4 Enhanced:** Final optimization targeting all-linear projections for maximum parameter density.
## How to Use
```python
from unsloth import FastLanguageModel
import torch
model, tokenizer = FastLanguageModel.from_pretrained(
model_name = "WeAreNexa/Nexa_0.1_First",
max_seq_length = 2048,
load_in_4bit = True,
)
# Fast inference mode