Apex-X commited on
Commit
bcff88a
·
verified ·
1 Parent(s): 7c75776

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +178 -2
README.md CHANGED
@@ -1,4 +1,180 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
 
2
- # TinyLLaMA Alpaca Unsloth Fine-Tuned
3
 
4
- This model is fine-tuned on instruction-following tasks using Unsloth and Alpaca-style prompts.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ library_name: transformers, llama-cpp-python
4
+ tags:
5
+ - dual-model
6
+ - edge-ai
7
+ - instruction-tuned
8
+ - robotics
9
+ - quantized
10
+ - raspberry-pi
11
+ - llama
12
+ - intent-classification
13
+ - text-generation
14
+ - on-device
15
+ language:
16
+ - en
17
+ - ta
18
+ - hi
19
+ datasets:
20
+ - Apex-X/PRODIGY-LAB_SARA
21
+ ---
22
+ ![PRODIGY-Dual](https://img.shields.io/badge/PRODIGY--DOIECHI--SARA-Dual%20Model%20System-blue)
23
+ ![Memory-Efficient](https://img.shields.io/badge/Edge--Optimized-Raspberry%20Pi%20Ready-green)
24
+ ![Multi-Model](https://img.shields.io/badge/Dual--Architecture-Intelligent%20Routing-orange)
25
+ ![Python](https://img.shields.io/badge/Python-3.8%2B-green)
26
+ ![License](https://img.shields.io/badge/License-MIT-yellow)
27
+ ![Multilingual](https://img.shields.io/badge/Multilingual-South%20Indian%20Languages-orange)
28
 
29
+ # Model Card for PRODIGY-DOIECHI & PRODIGY-SARA
30
 
31
+ A revolutionary **dual-model AI system** optimized for **edge devices** (Raspberry Pi, Jetson Nano, etc.), combining an ultra-lightweight intent classifier (**PRODIGY-DOIECHI**) with a powerful reasoning engine (**PRODIGY-SARA**). The system intelligently routes queries to the optimal model based on complexity, enabling both **sub-100ms responses** and **deep reasoning** on low-resource hardware.
32
+
33
+ - **GitHub**: [https://github.com/Apex-X/PRODIGY-DOIECHI-SARA](https://github.com/Apex-X/PRODIGY-DOIECHI-SARA)
34
+ - **Author**: Aadhithya (Apex-X)
35
+ - **Contact**: aadhithyaxll@gmail.com
36
+ - **License**: MIT
37
+
38
+ ---
39
+
40
+ ## Model Details
41
+
42
+ ### 🚀 PRODIGY-DOIECHI
43
+ - **Type**: Neural Network Intent Classifier
44
+ - **Parameters**: 1.2M
45
+ - **Format**: PyTorch (`.pth`)
46
+ - **Size**: < 50 MB
47
+ - **Input**: Natural language query
48
+ - **Output**: Structured intent + confidence score
49
+ - **Use Case**: Commands, calculations, system operations, greetings
50
+
51
+ ### 🧠 PRODIGY-SARA
52
+ - **Base Model**: Llama-7B ,QWEN-238B
53
+ - **Quantization**: GGUF `Q4_K_M`
54
+ - **Format**: GGUF (via `llama-cpp-python`)
55
+ - **Size**: ~3.8 GB
56
+ - **Context Window**: 1024 tokens
57
+ - **Fine-tuned on**: [Apex-X/PRODIGY-LAB_SARA](https://huggingface.co/datasets/Apex-X/PRODIGY-LAB_SARA)
58
+ - **Domains**: Robotics, Tamil Nadu culture, agriculture, medical support, ethics, general reasoning
59
+
60
+ ---
61
+
62
+ ## Intended Use
63
+
64
+ This system is designed for **on-device AI assistants** in:
65
+ - Smart homes
66
+ - Educational robots
67
+ - Industrial IoT monitoring
68
+ - Rural digital kiosks (supporting Tamil, Hindi, English)
69
+
70
+ **Not intended** for high-stakes medical diagnosis, legal advice, or autonomous weapon systems.
71
+
72
+ ---
73
+
74
+ ## Performance (Raspberry Pi 4)
75
+
76
+ | Metric | DOIECHI | SARA |
77
+ |----------------------|-------------|--------------|
78
+ | RAM Usage | 45 MB | 3.8 GB |
79
+ | Avg. Latency | 85 ms | 2.3 sec |
80
+ | Throughput | 11.8 q/s | 0.43 q/s |
81
+ | Intent Accuracy | 89% | — |
82
+ | Generation Speed | — | 2.3 tok/sec |
83
+
84
+ > Combined system averages **0.8s response time** in real-world mixed workloads.
85
+
86
+ ---
87
+
88
+ ## Smart Routing Logic
89
+
90
+ The system uses a 4-stage pipeline:
91
+ 1. **Memory** → short-term context (non-persistent)
92
+ 2. **Function Executor** → direct system commands
93
+ 3. **DOIECHI** → classify intent & complexity
94
+ 4. **SARA** → activated only if confidence < 0.85 or query contains reasoning keywords (`explain`, `how`, `why`, etc.)
95
+
96
+ Routing decision takes **< 5ms** and learns from usage patterns.
97
+
98
+ ---
99
+
100
+ ## How to Use
101
+
102
+ ### Python (Combined System)
103
+ ```python
104
+ from prodigy_system import ProdigyDualSystem
105
+ system = ProdigyDualSystem()
106
+ print(system.process("What's 128 / 4?")) # → DOIECHI
107
+ print(system.process("Explain photosynthesis.")) # → SARA
108
+
109
+ ## 🧠 Hardware Requirements
110
+
111
+ **Minimum**
112
+ - Raspberry Pi Zero 2 W (64-bit OS)
113
+ - OS: Linux (64-bit)
114
+ - Python: 3.8+
115
+ - RAM: 2–4 GB
116
+ - Dependencies:
117
+ `torch`, `llama-cpp-python`, `nltk`, `psutil`
118
+
119
+ **Recommended**
120
+ - Raspberry Pi 4 (4GB+ RAM) or NVIDIA Jetson Nano
121
+ - OS: Ubuntu 20.04+ / Raspberry Pi OS 64-bit
122
+ - Python: 3.10+
123
+ - Dependencies:
124
+ `torch` (CUDA build), `llama-cpp-python`, `nltk`, `psutil`, `huggingface_hub`
125
+
126
+ **Optional**
127
+ - Use Hugging Face Spaces or local FastAPI app for deployment
128
+ - GPU acceleration supported (NVIDIA RTX 20xx or higher)
129
+ - Convert model to GGUF or quantized formats for faster inference
130
+
131
+ ---
132
+
133
+ ## ⚖️ Ethical Considerations
134
+
135
+ - **No Persistent Data Storage**
136
+ The system does not store personal data or user history beyond the current session.
137
+
138
+ - **User Privacy First**
139
+ Every interaction is processed locally or in-memory. No external tracking or telemetry.
140
+
141
+ - **Multilingual Accessibility**
142
+ Built with South Indian language inclusivity in mind, ensuring wider digital access.
143
+
144
+ - **Bias Awareness**
145
+ Model responses are generated from training data that may contain inherent biases.
146
+ Always review critical outputs with human oversight.
147
+
148
+ - **Responsible Usage**
149
+ This model is for research, educational, and robotics-related applications only.
150
+ Avoid use in contexts that generate harmful, discriminatory, or deceptive content.
151
+
152
+ ---
153
+
154
+ *Developed as part of the **PRODIGY 1.2B** open research initiative on Hugging Face.*
155
+ *Optimized for lightweight AI deployment on edge devices like Raspberry Pi and Jetson Nano.*
156
+
157
+ Citation
158
+ @software{prodigy_dual_2025,
159
+ author = {Aadhithya Ravi},
160
+ title = {PRODIGY-DOIECHI \& PRODIGY-SARA: A Dual-Model Edge AI System},
161
+ year = {2025},
162
+ publisher = {GitHub},
163
+ journal = {GitHub repository},
164
+ howpublished = {\url{https://github.com/Apex-X/PRODIGY-DOIECHI-SARA}}
165
+ }
166
+ ---
167
+
168
+ ## 🏁 Acknowledgements
169
+
170
+ - Built on **llama.cpp** and **PyTorch**
171
+ - Inspired by **Alpaca**, **Self-Instruct**, and **TinyLLM** research
172
+ - Special thanks to the **Raspberry Pi** and **open-source AI** communities for enabling lightweight, accessible edge AI innovation
173
+
174
+ © 2025 **Aadhithya (Apex-X)**. Released under the **MIT License**.
175
+
176
+ ---
177
+
178
+ This format follows Hugging Face’s **standard model card structure**, includes all metadata in the YAML frontmatter, and is ready to be used as the `README.md` in a Hugging Face **model repository** (e.g., `Apex-X/PRODIGY-DOIECHI-SARA`).
179
+
180
+ Let me know if you'd like separate cards for each model or a version optimized for the **Hugging Face Spaces** demo!