|
|
--- |
|
|
license: mit |
|
|
pipeline_tag: text-generation |
|
|
base_model: |
|
|
- meta-llama/Llama-3.2-3B-Instruct |
|
|
--- |
|
|
# π IoraX 3B β Efficient Conversational AI Model |
|
|
 |
|
|
|
|
|
|
|
|
## β¨ Model Overview |
|
|
|
|
|
**IoraX 3B** is a highly efficient 3-billion parameter Transformer, fine-tuned using LoRA adapters on Meta LLaMA 3.2 (3B) β with 4-bit quantization to keep it lightning fast and lightweight! |
|
|
|
|
|
This model specializes in deep conversational understanding, logical reasoning, and coherent long-form generation β your AI companion for research, education, and creative tasks. |
|
|
|
|
|
--- |
|
|
|
|
|
## π― Features & Capabilities |
|
|
|
|
|
- π§ **Size:** 3B parameters |
|
|
- βοΈ **Base:** Meta LLaMA 3.2 (3B) |
|
|
- π§ **Fine-tuning:** LoRA with 4-bit quantization |
|
|
- β³ **Max context length:** 2048 tokens (with RoPE scaling) |
|
|
- π **Training data:** Blend of public conversational datasets + expert-curated Q&A |
|
|
- π **Epochs:** 3 for balanced speed and learning |
|
|
- π **Language:** English |
|
|
|
|
|
--- |
|
|
|
|
|
## π Use Cases |
|
|
|
|
|
| Use Case | Description | |
|
|
|------------------------|-----------------------------------------| |
|
|
| π¬ Conversational AI | Customer support, chatbots, assistants | |
|
|
| π Education | Tutoring, concept explanation, Q&A | |
|
|
| π§ͺ Research Assistant | Drafting, summarizing, brainstorming | |
|
|
| βοΈ Creative Writing | Storytelling, script generation | |
|
|
|
|
|
--- |
|
|
|
|
|
## β οΈ Limitations |
|
|
|
|
|
- π
**Knowledge cutoff:** Data up to 2023 only |
|
|
- βοΈ **Bias:** May reflect biases present in the training corpus |
|
|
- βοΈ **Accuracy:** Verify important outputs, especially in critical domains |
|
|
- π§ββοΈ **Not a replacement for experts:** Use responsibly |
|
|
|
|
|
--- |
|
|
|
|
|
## π‘ Quick Start |
|
|
|
|
|
```python |
|
|
from transformers import AutoTokenizer |
|
|
from unsloth import FastLanguageModel |
|
|
|
|
|
model_name = "XythicK/IoraX-3B" |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
|
model = FastLanguageModel.from_pretrained(model_name, load_in_4bit=True, max_seq_length=2048) |
|
|
|
|
|
messages = [ |
|
|
{"role": "user", "content": "Explain the philosophical significance of the Eiffel Tower. ππ€"} |
|
|
] |
|
|
|
|
|
inputs = tokenizer.apply_chat_template( |
|
|
messages, |
|
|
tokenize=True, |
|
|
add_generation_prompt=True, |
|
|
return_tensors="pt" |
|
|
).to("cuda") |
|
|
|
|
|
outputs = model.generate( |
|
|
input_ids=inputs, |
|
|
max_new_tokens=128, |
|
|
temperature=1.2, |
|
|
use_cache=True |
|
|
) |
|
|
|
|
|
print(tokenizer.batch_decode(outputs, skip_special_tokens=True)) |
|
|
``` |
|
|
|
|
|
## π Contact |
|
|
|
|
|
**Maintainer:** **M Mashhudur Rahim [XythicK]** |
|
|
|
|
|
**Role:** |
|
|
**Independent Machine Learning Researcher & Model Infrastructure Maintainer** |
|
|
|
|
|
(Focused on model quantization, optimization, and efficient deployment) |
|
|
|
|
|
|
|
|
For issues, improvement requests, or additional quantization formats, please use the Hugging Face Discussions or Issues tab. |
|
|
|
|
|
## π Citation |
|
|
|
|
|
If you use IoraX in your work, please cite: |
|
|
```bibtex |
|
|
@misc{ioraX2025, |
|
|
title = {IoraX 3B: Efficient Conversational AI}, |
|
|
author = {M Mashhudur Rahim (XythicK)}, |
|
|
year = {2025}, |
|
|
publisher = {Hugging Face}, |
|
|
howpublished = {\url{https://huggingface.co/XythicK/IoraX-3B}} |
|
|
} |
|
|
``` |
|
|
|
|
|
## β€οΈ Acknowledgements |
|
|
|
|
|
Thanks to Hugging Face and the open-source machine learning community for providing the tools and platforms that make efficient model sharing and deployment possible. |