rlm-small-v1 / README.md
prabhatkr's picture
Initial release: ParamTatva RLM Small v1 — Phonetically-Grounded Language Model
b5dd066 verified
---
license: other
license_name: paramtatva-restricted-1.0
license_link: LICENSE
language:
- sa
- en
library_name: transformers
tags:
- paramtatva
- rlm
- resonance
- sanskrit
- maheshwara-sutras
- math
- phonetic-grounding
pipeline_tag: text-generation
---
# ParamTatva RLM-Small-v1
**Resonance Language Model** — A phonetically-grounded transformer trained with insights from the Maheshwara Sutras.
## Model Description
ParamTatva RLM is a novel language model architecture that replaces standard positional encodings with **phonetic graph embeddings** derived from the [Maheshwara Sutras](https://en.wikipedia.org/wiki/Shiva_Sutras), the foundational grammar rules of Sanskrit attributed to Pāṇini.
### Key Innovations
| Feature | Description |
|---------|-------------|
| **Paramtatva Graph Embeddings** | Token embeddings informed by phonetic proximity in the Maheshwara Sutras |
| **Pratyāhāra Attention Bias** | Attention biases derived from Pāṇini's abbreviation system (pratyāhāra) |
| **Mā-Bridge Normalization** | Layer normalization conditioned on phonetic group structure |
### Architecture
```
ParamtatvaTransformer (Small)
├── Embedding: ParamtatvaEmbedding (phonetic graph-aware)
├── Layers: 6 × TransformerBlock
│ ├── Attention: Multi-Head + Pratyāhāra Bias
│ ├── FFN: GELU activation
│ └── Norm: LayerNorm + Mā-Bridge
├── Final LayerNorm
└── LM Head
```
| Parameter | Value |
|-----------|-------|
| Parameters | ~10M |
| Hidden dim | 256 |
| Layers | 6 |
| Attention heads | 8 |
| Intermediate dim | 1024 |
| Max sequence length | 1024 |
| Activation | GELU |
## Intended Use
This model is released for **research and academic purposes**. It demonstrates the viability of phonetically-grounded language modeling using ancient linguistic frameworks.
### Recommended Uses
- Research into phonetic/linguistic priors for language models
- Studies on Sanskrit computational linguistics
- Mathematical reasoning experiments
- Exploration of alternative positional encoding schemes
### Out-of-Scope Uses
- Production/commercial applications (requires separate license)
- Safety-critical systems
- Any use that violates the license terms
## Training
The model was trained using the ParamTatva training pipeline. The training methodology, loss functions, and data curation are proprietary. Only the resulting model weights are released.
**Note**: The full Resonance Learning System (including the proprietary ResonanceEncoder) is NOT included in this release. This release contains only the standard ParamtatvaTransformer weights.
## How to Use
```python
import torch
from safetensors.torch import load_file
# Load weights
state_dict = load_file("model.safetensors")
# The model uses a custom architecture — see paramtatva_transformer.py
# for the full model class definition.
print(f"Parameters: {sum(v.numel() for v in state_dict.values()):,}")
```
## Limitations
- This is a **small** model (~10M parameters) — intended as a proof of concept
- The model was trained on a limited dataset
- Performance on downstream tasks has not been extensively benchmarked
- The proprietary resonance components are not included
## Citation
```bibtex
@misc{paramtatva2026rlm,
title={ParamTatva RLM: A Phonetically-Grounded Language Model
Based on the Maheshwara Sutras},
author={ParamTatva.org},
year={2026},
url={https://huggingface.co/paramtatva/rlm-small-v1}
}
```
## License
This model is released under the **ParamTatva Restricted Use License v1.0**:
- ✅ Research and academic use
- ✅ Non-commercial applications
- ✅ Fine-tuning for research
- ❌ Commercial use (requires written agreement)
- ❌ Reverse engineering of training methodology
See [LICENSE](LICENSE) for full terms.
## Contact
- **Commercial licensing**: licensing@paramtatva.org
- **Research inquiries**: research@paramtatva.org
- **Website**: [paramtatva.org](https://paramtatva.org)