π± Indro-Veda: The Sovereign Reasoning Model (500M)
Indro-Veda is a state-of-the-art Small Language Model (SLM) developed by Indro-ai. With 500 million parameters, it is specifically engineered to demonstrate high-level reasoning, logical deduction, and structured problem-solvingβcapabilities typically reserved for much larger models.
The name "Indro-Veda" signifies the fusion of supreme intelligence (Indro) and profound knowledge (Veda).
π Model Highlights
- Architecture: Optimized Transformer-based architecture (Llama-style).
- Parameters: 500 Million.
- Training Tokens: 3 Billion curated tokens.
- Specialization: Mathematics, Algorithmic Code, and High-Quality Educational Reasoning.
- Framework: Trained using PyTorch/XLA on TPU infrastructure for maximum efficiency.
π§ Training Philosophy: "Reasoning over Recall"
Unlike traditional small models that focus on memorizing facts, Indro-Veda is trained on a Reasoning-Heavy Dataset Mixture:
- Logical Core (Math): Powered by
UltraData-Mathto ensure the model understands step-by-step derivation. - Structural Core (Code): Trained on
starcoderdatato enhance algorithmic thinking and syntax awareness. - Knowledge Core (Education): Built on
FineWeb-Eduto provide a clean, high-signal educational foundation.
π Dataset Distribution
The model was pre-trained on the Indro-Veda Dataset (3B Tokens) with a fixed-ratio mixture designed to prevent catastrophic forgetting and maintain balanced intelligence across domains.
| Component | Focus | Data Source |
|---|---|---|
| Reasoning | Mathematics & Logic | UltraData-Math |
| Structure | Programming & Algorithms | Starcoderdata |
| Knowledge | High-Quality Educational Web | FineWeb-Edu |
| Identity | Sovereign Alignment | Indro-ai Proprietary |
π οΈ Usage
You can use Indro-Veda with the Hugging Face transformers library:
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "Indro-ai/Indro-Veda-500M"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
prompt = "Explain the concept of logical deduction."
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))