File size: 4,915 Bytes
6eb6514 13494b0 34ca02a 13494b0 5cece9d 13494b0 34ca02a 13494b0 34ca02a 13494b0 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 | ---
license: apache-2.0
datasets:
- axmeeabdhullo/axya-tech-dv100
language:
- dv
metrics:
- accuracy
base_model:
- openai-community/gpt2
new_version: axmeeabdhullo/axya-mini
pipeline_tag: question-answering
library_name: adapter-transformers
---
# Axya-Mini
> A fine-tuned GPT-2 adapter model for Dhivehi (Thaana) language question-answering and text generation tasks.
## Model Description
Axya-Mini is a lightweight, efficient adapter-based language model specifically designed for the Dhivehi language, the national language of the Maldives. Built on the GPT-2 architecture and optimized using adapter layers, this model excels at question-answering tasks while maintaining compact size and fast inference.
**Model Type:** Adapter-based Fine-tuned Model
**Base Model:** GPT-2 (openai-community/gpt2)
**Language:** Dhivehi (ދިވެހި)
**Framework:** Quetzal (Revolutionary CPU-optimized training library)
## Key Features
- 🌟 **Language-Specific:** Optimized for Dhivehi (dv) language processing
- ⚡ **Lightweight:** Efficient adapter architecture for fast inference
- 🎉 **Question Answering:** Trained on question-answering tasks
- 💾 **Safetensors Format:** Secure model serialization
- 🤗 **Adapter-Based:** Uses adapter layers for efficient fine-tuning and storage
## Model Details
### Intended Use
This model is designed for:
- Question answering in Dhivehi
- Text generation tasks in Dhivehi
- Language understanding for Dhivehi content
- Building Dhivehi NLP applications
### Training Data
**Dataset:** [axmeeabdhullo/axya-tech-dv100](https://huggingface.co/datasets/axmeeabdhullo/axya-tech-dv100)
A curated Dhivehi dataset containing 100 high-quality technical and educational content samples.
### Training Methodology
- **Fine-tuning Approach:** Adapter-based fine-tuning
- **Metrics:** Accuracy optimization
- **Library:** adapter-transformers
- **Optimization:** Efficient parameter updating through adapter modules
### Quetzal Library Optimization
**Quetzal** is a revolutionary library that powers the efficient training of this model on CPU, making high-quality AI accessible without expensive GPUs.
**Library Features:**
- 🚀 **3x Faster CPU Training**: Advanced optimizations for CPU-based training
- 📊 **Data Augmentation**: Train accurate models with minimal data (5-10x augmentation)
- 💾 **Memory Efficient**: 4-bit quantization and LoRA for reduced memory usage
- 🎯 **High Accuracy**: Specialized techniques for low-resource scenarios
- 🌍 **Multilingual**: Optimized for languages like Dhivehi, but works for any language
- 🔧 **Easy to Use**: Simple API similar to popular libraries
**Installation:**
```bash
pip install quetzal-ai
```
**Why Quetzal for Dhivehi?**
Quetzal is specifically optimized for low-resource languages like Dhivehi, enabling efficient model training and deployment without requiring expensive GPU infrastructure. This makes it ideal for building NLP models for endangered or underrepresented languages.
## Model Performance
- **Inference Speed:** Fast and efficient due to adapter architecture
- **Model Size:** Compact compared to full model fine-tuning
- **Accuracy:** Optimized for Dhivehi language understanding tasks
## Limitations
- Trained specifically on Dhivehi language content
- Performance may vary with dialects or regional variations
- Requires GPU/TPU for optimal inference speed
- Limited evaluation on diverse downstream tasks
## Recommendations
1. **Fine-tuning:** Can be further fine-tuned on domain-specific Dhivehi data
2. **Deployment:** Use with sufficient computational resources for production
3. **Evaluation:** Test on your specific use case before deployment
4. **Updates:** Check for newer versions of the model for improved performance
## Citation
If you use this model, please cite:
```bibtex
@model{axya_mini,
author = {Abdhullo, Axmee},
title = {Axya-Mini: Dhivehi Language Question-Answering Model},
year = {2025},
publisher = {Hugging Face Model Hub},
howpublished = {https://huggingface.co/axmeeabdhullo/axya-mini}
}
```
## License
This model is licensed under the Apache License 2.0. See the LICENSE file for details.
## Related Resources
- **Dataset:** [axya-tech-dv100](https://huggingface.co/datasets/axmeeabdhullo/axya-tech-dv100)
- **Base Model:** [GPT-2](https://huggingface.co/openai-community/gpt2)
- **Library Documentation:** [Adapter-Transformers](https://adapterhub.ml/)
- **Hugging Face Hub:** [Model Hub](https://huggingface.co/)
## Author
**Axmee Abdhullo**
AI/ML Developer specializing in Dhivehi NLP
[Hugging Face](https://huggingface.co/axmeeabdhullo)
## Contact & Support
For questions, suggestions, or support:
- Open an issue on the model's repository
- Join the Hugging Face community discussions
- Check the model card for updates
---
**Last Updated:** December 2024
**Status:** Active Development
**Version:** 1.0
|