| --- |
| language: |
| - en |
| tags: |
| - math |
| - education |
| - llama-3 |
| - peft |
| - lora |
| base_model: meta-llama/Llama-3.2-1B-Instruct |
| license: apache-2.0 |
| --- |
| |
| # NexusLLM-Math-1B-v1 |
|
|
| ## Model Details |
| NexusLLM-Math-1B-v1 is a fine-tuned version of Llama 3.2 (1B parameters) optimized specifically for solving advanced high-school mathematics problems, with a focus on JEE Main and Advanced syllabus topics. |
|
|
| - **Developed by:** ZentithLLM |
| - **Model Type:** Causal Language Model (Fine-tuned with LoRA) |
| - **Language:** English |
| - **Base Model:** meta-llama/Llama-3.2-1B-Instruct |
| - **Precision:** FP16 |
|
|
| ## Intended Use |
| This model is designed to act as an educational assistant for 11th-grade mathematics. It is trained to provide step-by-step reasoning and explanations for complex topics, rather than just outputting the final answer. |
|
|
| **Primary Topics Covered:** |
| - Binomial Theorem |
| - Geometry (Circle Theorems, cyclic quadrilaterals, tangents, etc.) |
|
|
| ## Training Data |
| The model was trained on a custom dataset of structured mathematics Q&A pairs. The dataset maps specific mathematical prompts to detailed completions, heavily utilizing an `explanation` field to teach the model the underlying mathematical logic and derivation steps. |
|
|
| ## Training Procedure |
| The model was fine-tuned using the standard Hugging Face `trl` and `peft` libraries on a single NVIDIA T4 GPU, utilizing strictly native FP16 precision to ensure mathematical gradient stability. |
|
|
| - **Training Framework:** Pure Hugging Face (No Unsloth/Quantization) |
| - **Method:** LoRA (Low-Rank Adaptation) |
| - **Rank (r):** 32 |
| - **Alpha:** 32 |
| - **Optimizer:** adamw_torch |
| - **Learning Rate:** 2e-4 |
| - **Max Sequence Length:** 2048 |
| |
| ## How to Use |
| Because this model was trained on a specific dataset structure, you **must** wrap your prompts in the `### Instruction:` and `### Response:` format for it to output the correct mathematical explanations. |
| |
| ```python |
| import torch |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| |
| model_id = "ZentithLLM/NexusLLM-Math-1B-v1" |
|
|
| tokenizer = AutoTokenizer.from_pretrained(model_id) |
| model = AutoModelForCausalLM.from_pretrained( |
| model_id, |
| torch_dtype=torch.float16, |
| device_map="auto" |
| ) |
| |
| question = "What is the general term in the expansion of (x+y)^n?" |
| formatted_prompt = f"### Instruction:\\n{question}\\n\\n### Response:\\n" |
| |
| inputs = tokenizer(formatted_prompt, return_tensors="pt").to(model.device) |
| |
| outputs = model.generate( |
| **inputs, |
| max_new_tokens=250, |
| temperature=0.3, |
| do_sample=True, |
| pad_token_id=tokenizer.eos_token_id |
| ) |
| |
| print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |