--- license: mit pipeline_tag: text-generation base_model: - meta-llama/Llama-3.2-3B-Instruct --- # ๐Ÿš€ IoraX 3B โ€” Efficient Conversational AI Model ![IoraX Logo](./IoraX.png) ## โœจ Model Overview **IoraX 3B** is a highly efficient 3-billion parameter Transformer, fine-tuned using LoRA adapters on Meta LLaMA 3.2 (3B) โ€” with 4-bit quantization to keep it lightning fast and lightweight! This model specializes in deep conversational understanding, logical reasoning, and coherent long-form generation โ€” your AI companion for research, education, and creative tasks. --- ## ๐ŸŽฏ Features & Capabilities - ๐Ÿง  **Size:** 3B parameters - โš™๏ธ **Base:** Meta LLaMA 3.2 (3B) - ๐Ÿ”ง **Fine-tuning:** LoRA with 4-bit quantization - โณ **Max context length:** 2048 tokens (with RoPE scaling) - ๐Ÿ“š **Training data:** Blend of public conversational datasets + expert-curated Q&A - ๐Ÿ”„ **Epochs:** 3 for balanced speed and learning - ๐ŸŒ **Language:** English --- ## ๐Ÿš€ Use Cases | Use Case | Description | |------------------------|-----------------------------------------| | ๐Ÿ’ฌ Conversational AI | Customer support, chatbots, assistants | | ๐ŸŽ“ Education | Tutoring, concept explanation, Q&A | | ๐Ÿงช Research Assistant | Drafting, summarizing, brainstorming | | โœ๏ธ Creative Writing | Storytelling, script generation | --- ## โš ๏ธ Limitations - ๐Ÿ“… **Knowledge cutoff:** Data up to 2023 only - โš–๏ธ **Bias:** May reflect biases present in the training corpus - โœ”๏ธ **Accuracy:** Verify important outputs, especially in critical domains - ๐Ÿง‘โ€โš–๏ธ **Not a replacement for experts:** Use responsibly --- ## ๐Ÿ’ก Quick Start ```python from transformers import AutoTokenizer from unsloth import FastLanguageModel model_name = "XythicK/IoraX-3B" tokenizer = AutoTokenizer.from_pretrained(model_name) model = FastLanguageModel.from_pretrained(model_name, load_in_4bit=True, max_seq_length=2048) messages = [ {"role": "user", "content": "Explain the philosophical significance of the Eiffel Tower. ๐ŸŒ‰๐Ÿค”"} ] inputs = tokenizer.apply_chat_template( messages, tokenize=True, add_generation_prompt=True, return_tensors="pt" ).to("cuda") outputs = model.generate( input_ids=inputs, max_new_tokens=128, temperature=1.2, use_cache=True ) print(tokenizer.batch_decode(outputs, skip_special_tokens=True)) ``` ## ๐Ÿ™‹ Contact **Maintainer:** **M Mashhudur Rahim [XythicK]** **Role:** **Independent Machine Learning Researcher & Model Infrastructure Maintainer** (Focused on model quantization, optimization, and efficient deployment) For issues, improvement requests, or additional quantization formats, please use the Hugging Face Discussions or Issues tab. ## ๐Ÿ“„ Citation If you use IoraX in your work, please cite: ```bibtex @misc{ioraX2025, title = {IoraX 3B: Efficient Conversational AI}, author = {M Mashhudur Rahim (XythicK)}, year = {2025}, publisher = {Hugging Face}, howpublished = {\url{https://huggingface.co/XythicK/IoraX-3B}} } ``` ## โค๏ธ Acknowledgements Thanks to Hugging Face and the open-source machine learning community for providing the tools and platforms that make efficient model sharing and deployment possible.