Initial release: ParamTatva RLM Small v1 — Phonetically-Grounded Language Model
Browse files- LICENSE +98 -0
- README.md +128 -0
- config.json +25 -0
- pytorch_model.bin +3 -0
LICENSE
ADDED
|
@@ -0,0 +1,98 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
ParamTatva Restricted Use License
|
| 2 |
+
Version 1.0, February 2026
|
| 3 |
+
|
| 4 |
+
Copyright (c) 2025-2026 ParamTatva.org
|
| 5 |
+
All rights reserved.
|
| 6 |
+
|
| 7 |
+
TERMS AND CONDITIONS
|
| 8 |
+
|
| 9 |
+
1. DEFINITIONS
|
| 10 |
+
|
| 11 |
+
"Software" means the model weights, configuration files, and
|
| 12 |
+
associated documentation distributed under this License.
|
| 13 |
+
|
| 14 |
+
"Licensor" means ParamTatva.org.
|
| 15 |
+
|
| 16 |
+
"You" means the individual or entity exercising permissions
|
| 17 |
+
granted by this License.
|
| 18 |
+
|
| 19 |
+
"Commercial Use" means any use intended for or directed toward
|
| 20 |
+
commercial advantage or monetary compensation.
|
| 21 |
+
|
| 22 |
+
"Derivative Work" means any work that is based on or derived
|
| 23 |
+
from the Software, including but not limited to fine-tuned models,
|
| 24 |
+
distilled models, merged models, or quantized versions.
|
| 25 |
+
|
| 26 |
+
2. GRANT OF RIGHTS
|
| 27 |
+
|
| 28 |
+
Subject to the terms of this License, the Licensor grants You a
|
| 29 |
+
worldwide, non-exclusive, non-transferable, revocable license to:
|
| 30 |
+
|
| 31 |
+
(a) Use the Software for research and academic purposes;
|
| 32 |
+
(b) Use the Software for personal, non-commercial applications;
|
| 33 |
+
(c) Create Derivative Works for research purposes only;
|
| 34 |
+
(d) Publish research results obtained using the Software,
|
| 35 |
+
provided proper attribution is given.
|
| 36 |
+
|
| 37 |
+
3. RESTRICTIONS
|
| 38 |
+
|
| 39 |
+
You may NOT:
|
| 40 |
+
|
| 41 |
+
(a) Use the Software or any Derivative Work for Commercial Use
|
| 42 |
+
without a separate written agreement with the Licensor;
|
| 43 |
+
(b) Distribute Derivative Works without:
|
| 44 |
+
(i) including this License;
|
| 45 |
+
(ii) providing clear attribution to ParamTatva.org;
|
| 46 |
+
(c) Attempt to reverse engineer, decompile, or derive the
|
| 47 |
+
training methodology, training data, or proprietary
|
| 48 |
+
components of the Licensor's systems from the Software;
|
| 49 |
+
(d) Remove or alter any copyright, trademark, or attribution
|
| 50 |
+
notices contained in the Software;
|
| 51 |
+
(e) Use the Software to train competing commercial products
|
| 52 |
+
without written permission;
|
| 53 |
+
(f) Sublicense the Software.
|
| 54 |
+
|
| 55 |
+
4. ATTRIBUTION
|
| 56 |
+
|
| 57 |
+
Any use of the Software must include the following attribution:
|
| 58 |
+
|
| 59 |
+
"This work uses the ParamTatva Resonance Language Model (RLM),
|
| 60 |
+
developed by ParamTatva.org."
|
| 61 |
+
|
| 62 |
+
Academic citations should reference:
|
| 63 |
+
@misc{paramtatva2026rlm,
|
| 64 |
+
title={ParamTatva RLM: A Phonetically-Grounded Language Model},
|
| 65 |
+
author={ParamTatva.org},
|
| 66 |
+
year={2026},
|
| 67 |
+
url={https://huggingface.co/paramtatva/rlm-small-v1}
|
| 68 |
+
}
|
| 69 |
+
|
| 70 |
+
5. COMMERCIAL LICENSING
|
| 71 |
+
|
| 72 |
+
For commercial use, contact: licensing@paramtatva.org
|
| 73 |
+
|
| 74 |
+
Commercial licenses are available for:
|
| 75 |
+
- Integration into commercial products
|
| 76 |
+
- Commercial API services
|
| 77 |
+
- Enterprise deployments
|
| 78 |
+
|
| 79 |
+
6. NO WARRANTY
|
| 80 |
+
|
| 81 |
+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
| 82 |
+
EXPRESS OR IMPLIED. IN NO EVENT SHALL THE LICENSOR BE LIABLE FOR
|
| 83 |
+
ANY CLAIM, DAMAGES, OR OTHER LIABILITY ARISING FROM THE SOFTWARE.
|
| 84 |
+
|
| 85 |
+
7. TERMINATION
|
| 86 |
+
|
| 87 |
+
This License automatically terminates if You violate any of its
|
| 88 |
+
terms. Upon termination, You must destroy all copies of the
|
| 89 |
+
Software in Your possession.
|
| 90 |
+
|
| 91 |
+
8. GOVERNING LAW
|
| 92 |
+
|
| 93 |
+
This License shall be governed by the laws of India.
|
| 94 |
+
|
| 95 |
+
---
|
| 96 |
+
|
| 97 |
+
For questions: legal@paramtatva.org
|
| 98 |
+
Website: https://paramtatva.org
|
README.md
ADDED
|
@@ -0,0 +1,128 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: other
|
| 3 |
+
license_name: paramtatva-restricted-1.0
|
| 4 |
+
license_link: LICENSE
|
| 5 |
+
language:
|
| 6 |
+
- sa
|
| 7 |
+
- en
|
| 8 |
+
library_name: transformers
|
| 9 |
+
tags:
|
| 10 |
+
- paramtatva
|
| 11 |
+
- rlm
|
| 12 |
+
- resonance
|
| 13 |
+
- sanskrit
|
| 14 |
+
- maheshwara-sutras
|
| 15 |
+
- math
|
| 16 |
+
- phonetic-grounding
|
| 17 |
+
pipeline_tag: text-generation
|
| 18 |
+
---
|
| 19 |
+
|
| 20 |
+
# ParamTatva RLM-Small-v1
|
| 21 |
+
|
| 22 |
+
**Resonance Language Model** — A phonetically-grounded transformer trained with insights from the Maheshwara Sutras.
|
| 23 |
+
|
| 24 |
+
## Model Description
|
| 25 |
+
|
| 26 |
+
ParamTatva RLM is a novel language model architecture that replaces standard positional encodings with **phonetic graph embeddings** derived from the [Maheshwara Sutras](https://en.wikipedia.org/wiki/Shiva_Sutras), the foundational grammar rules of Sanskrit attributed to Pāṇini.
|
| 27 |
+
|
| 28 |
+
### Key Innovations
|
| 29 |
+
|
| 30 |
+
| Feature | Description |
|
| 31 |
+
|---------|-------------|
|
| 32 |
+
| **Paramtatva Graph Embeddings** | Token embeddings informed by phonetic proximity in the Maheshwara Sutras |
|
| 33 |
+
| **Pratyāhāra Attention Bias** | Attention biases derived from Pāṇini's abbreviation system (pratyāhāra) |
|
| 34 |
+
| **Mā-Bridge Normalization** | Layer normalization conditioned on phonetic group structure |
|
| 35 |
+
|
| 36 |
+
### Architecture
|
| 37 |
+
|
| 38 |
+
```
|
| 39 |
+
ParamtatvaTransformer (Small)
|
| 40 |
+
├── Embedding: ParamtatvaEmbedding (phonetic graph-aware)
|
| 41 |
+
├── Layers: 6 × TransformerBlock
|
| 42 |
+
│ ├── Attention: Multi-Head + Pratyāhāra Bias
|
| 43 |
+
│ ├── FFN: GELU activation
|
| 44 |
+
│ └── Norm: LayerNorm + Mā-Bridge
|
| 45 |
+
├── Final LayerNorm
|
| 46 |
+
└── LM Head
|
| 47 |
+
```
|
| 48 |
+
|
| 49 |
+
| Parameter | Value |
|
| 50 |
+
|-----------|-------|
|
| 51 |
+
| Parameters | ~10M |
|
| 52 |
+
| Hidden dim | 256 |
|
| 53 |
+
| Layers | 6 |
|
| 54 |
+
| Attention heads | 8 |
|
| 55 |
+
| Intermediate dim | 1024 |
|
| 56 |
+
| Max sequence length | 1024 |
|
| 57 |
+
| Activation | GELU |
|
| 58 |
+
|
| 59 |
+
## Intended Use
|
| 60 |
+
|
| 61 |
+
This model is released for **research and academic purposes**. It demonstrates the viability of phonetically-grounded language modeling using ancient linguistic frameworks.
|
| 62 |
+
|
| 63 |
+
### Recommended Uses
|
| 64 |
+
- Research into phonetic/linguistic priors for language models
|
| 65 |
+
- Studies on Sanskrit computational linguistics
|
| 66 |
+
- Mathematical reasoning experiments
|
| 67 |
+
- Exploration of alternative positional encoding schemes
|
| 68 |
+
|
| 69 |
+
### Out-of-Scope Uses
|
| 70 |
+
- Production/commercial applications (requires separate license)
|
| 71 |
+
- Safety-critical systems
|
| 72 |
+
- Any use that violates the license terms
|
| 73 |
+
|
| 74 |
+
## Training
|
| 75 |
+
|
| 76 |
+
The model was trained using the ParamTatva training pipeline. The training methodology, loss functions, and data curation are proprietary. Only the resulting model weights are released.
|
| 77 |
+
|
| 78 |
+
**Note**: The full Resonance Learning System (including the proprietary ResonanceEncoder) is NOT included in this release. This release contains only the standard ParamtatvaTransformer weights.
|
| 79 |
+
|
| 80 |
+
## How to Use
|
| 81 |
+
|
| 82 |
+
```python
|
| 83 |
+
import torch
|
| 84 |
+
from safetensors.torch import load_file
|
| 85 |
+
|
| 86 |
+
# Load weights
|
| 87 |
+
state_dict = load_file("model.safetensors")
|
| 88 |
+
|
| 89 |
+
# The model uses a custom architecture — see paramtatva_transformer.py
|
| 90 |
+
# for the full model class definition.
|
| 91 |
+
print(f"Parameters: {sum(v.numel() for v in state_dict.values()):,}")
|
| 92 |
+
```
|
| 93 |
+
|
| 94 |
+
## Limitations
|
| 95 |
+
|
| 96 |
+
- This is a **small** model (~10M parameters) — intended as a proof of concept
|
| 97 |
+
- The model was trained on a limited dataset
|
| 98 |
+
- Performance on downstream tasks has not been extensively benchmarked
|
| 99 |
+
- The proprietary resonance components are not included
|
| 100 |
+
|
| 101 |
+
## Citation
|
| 102 |
+
|
| 103 |
+
```bibtex
|
| 104 |
+
@misc{paramtatva2026rlm,
|
| 105 |
+
title={ParamTatva RLM: A Phonetically-Grounded Language Model
|
| 106 |
+
Based on the Maheshwara Sutras},
|
| 107 |
+
author={ParamTatva.org},
|
| 108 |
+
year={2026},
|
| 109 |
+
url={https://huggingface.co/paramtatva/rlm-small-v1}
|
| 110 |
+
}
|
| 111 |
+
```
|
| 112 |
+
|
| 113 |
+
## License
|
| 114 |
+
|
| 115 |
+
This model is released under the **ParamTatva Restricted Use License v1.0**:
|
| 116 |
+
- ✅ Research and academic use
|
| 117 |
+
- ✅ Non-commercial applications
|
| 118 |
+
- ✅ Fine-tuning for research
|
| 119 |
+
- ❌ Commercial use (requires written agreement)
|
| 120 |
+
- ❌ Reverse engineering of training methodology
|
| 121 |
+
|
| 122 |
+
See [LICENSE](LICENSE) for full terms.
|
| 123 |
+
|
| 124 |
+
## Contact
|
| 125 |
+
|
| 126 |
+
- **Commercial licensing**: licensing@paramtatva.org
|
| 127 |
+
- **Research inquiries**: research@paramtatva.org
|
| 128 |
+
- **Website**: [paramtatva.org](https://paramtatva.org)
|
config.json
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"ParamtatvaTransformer"
|
| 4 |
+
],
|
| 5 |
+
"model_type": "paramtatva-rlm",
|
| 6 |
+
"vocab_size": 1224,
|
| 7 |
+
"hidden_size": 128,
|
| 8 |
+
"num_hidden_layers": 4,
|
| 9 |
+
"num_attention_heads": 2,
|
| 10 |
+
"intermediate_size": 512,
|
| 11 |
+
"max_position_embeddings": 1024,
|
| 12 |
+
"hidden_act": "gelu",
|
| 13 |
+
"hidden_dropout_prob": 0.1,
|
| 14 |
+
"attention_probs_dropout_prob": 0.1,
|
| 15 |
+
"layer_norm_eps": 1e-06,
|
| 16 |
+
"initializer_range": 0.02,
|
| 17 |
+
"torch_dtype": "float32",
|
| 18 |
+
"transformers_version": "4.40.0",
|
| 19 |
+
"paramtatva_config": {
|
| 20 |
+
"use_graph_embeddings": true,
|
| 21 |
+
"use_pratyahara_bias": true,
|
| 22 |
+
"use_ma_bridge": true,
|
| 23 |
+
"phonetic_basis": "maheshwara_sutras"
|
| 24 |
+
}
|
| 25 |
+
}
|
pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7e345e1eb32297a516fe4ce74bb5c93ffea5fc76338b19b72758511a07c35bce
|
| 3 |
+
size 4121338
|