prabhatkr commited on
Commit
b5dd066
·
verified ·
1 Parent(s): 00a1bb6

Initial release: ParamTatva RLM Small v1 — Phonetically-Grounded Language Model

Browse files
Files changed (4) hide show
  1. LICENSE +98 -0
  2. README.md +128 -0
  3. config.json +25 -0
  4. pytorch_model.bin +3 -0
LICENSE ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ParamTatva Restricted Use License
2
+ Version 1.0, February 2026
3
+
4
+ Copyright (c) 2025-2026 ParamTatva.org
5
+ All rights reserved.
6
+
7
+ TERMS AND CONDITIONS
8
+
9
+ 1. DEFINITIONS
10
+
11
+ "Software" means the model weights, configuration files, and
12
+ associated documentation distributed under this License.
13
+
14
+ "Licensor" means ParamTatva.org.
15
+
16
+ "You" means the individual or entity exercising permissions
17
+ granted by this License.
18
+
19
+ "Commercial Use" means any use intended for or directed toward
20
+ commercial advantage or monetary compensation.
21
+
22
+ "Derivative Work" means any work that is based on or derived
23
+ from the Software, including but not limited to fine-tuned models,
24
+ distilled models, merged models, or quantized versions.
25
+
26
+ 2. GRANT OF RIGHTS
27
+
28
+ Subject to the terms of this License, the Licensor grants You a
29
+ worldwide, non-exclusive, non-transferable, revocable license to:
30
+
31
+ (a) Use the Software for research and academic purposes;
32
+ (b) Use the Software for personal, non-commercial applications;
33
+ (c) Create Derivative Works for research purposes only;
34
+ (d) Publish research results obtained using the Software,
35
+ provided proper attribution is given.
36
+
37
+ 3. RESTRICTIONS
38
+
39
+ You may NOT:
40
+
41
+ (a) Use the Software or any Derivative Work for Commercial Use
42
+ without a separate written agreement with the Licensor;
43
+ (b) Distribute Derivative Works without:
44
+ (i) including this License;
45
+ (ii) providing clear attribution to ParamTatva.org;
46
+ (c) Attempt to reverse engineer, decompile, or derive the
47
+ training methodology, training data, or proprietary
48
+ components of the Licensor's systems from the Software;
49
+ (d) Remove or alter any copyright, trademark, or attribution
50
+ notices contained in the Software;
51
+ (e) Use the Software to train competing commercial products
52
+ without written permission;
53
+ (f) Sublicense the Software.
54
+
55
+ 4. ATTRIBUTION
56
+
57
+ Any use of the Software must include the following attribution:
58
+
59
+ "This work uses the ParamTatva Resonance Language Model (RLM),
60
+ developed by ParamTatva.org."
61
+
62
+ Academic citations should reference:
63
+ @misc{paramtatva2026rlm,
64
+ title={ParamTatva RLM: A Phonetically-Grounded Language Model},
65
+ author={ParamTatva.org},
66
+ year={2026},
67
+ url={https://huggingface.co/paramtatva/rlm-small-v1}
68
+ }
69
+
70
+ 5. COMMERCIAL LICENSING
71
+
72
+ For commercial use, contact: licensing@paramtatva.org
73
+
74
+ Commercial licenses are available for:
75
+ - Integration into commercial products
76
+ - Commercial API services
77
+ - Enterprise deployments
78
+
79
+ 6. NO WARRANTY
80
+
81
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
82
+ EXPRESS OR IMPLIED. IN NO EVENT SHALL THE LICENSOR BE LIABLE FOR
83
+ ANY CLAIM, DAMAGES, OR OTHER LIABILITY ARISING FROM THE SOFTWARE.
84
+
85
+ 7. TERMINATION
86
+
87
+ This License automatically terminates if You violate any of its
88
+ terms. Upon termination, You must destroy all copies of the
89
+ Software in Your possession.
90
+
91
+ 8. GOVERNING LAW
92
+
93
+ This License shall be governed by the laws of India.
94
+
95
+ ---
96
+
97
+ For questions: legal@paramtatva.org
98
+ Website: https://paramtatva.org
README.md ADDED
@@ -0,0 +1,128 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: paramtatva-restricted-1.0
4
+ license_link: LICENSE
5
+ language:
6
+ - sa
7
+ - en
8
+ library_name: transformers
9
+ tags:
10
+ - paramtatva
11
+ - rlm
12
+ - resonance
13
+ - sanskrit
14
+ - maheshwara-sutras
15
+ - math
16
+ - phonetic-grounding
17
+ pipeline_tag: text-generation
18
+ ---
19
+
20
+ # ParamTatva RLM-Small-v1
21
+
22
+ **Resonance Language Model** — A phonetically-grounded transformer trained with insights from the Maheshwara Sutras.
23
+
24
+ ## Model Description
25
+
26
+ ParamTatva RLM is a novel language model architecture that replaces standard positional encodings with **phonetic graph embeddings** derived from the [Maheshwara Sutras](https://en.wikipedia.org/wiki/Shiva_Sutras), the foundational grammar rules of Sanskrit attributed to Pāṇini.
27
+
28
+ ### Key Innovations
29
+
30
+ | Feature | Description |
31
+ |---------|-------------|
32
+ | **Paramtatva Graph Embeddings** | Token embeddings informed by phonetic proximity in the Maheshwara Sutras |
33
+ | **Pratyāhāra Attention Bias** | Attention biases derived from Pāṇini's abbreviation system (pratyāhāra) |
34
+ | **Mā-Bridge Normalization** | Layer normalization conditioned on phonetic group structure |
35
+
36
+ ### Architecture
37
+
38
+ ```
39
+ ParamtatvaTransformer (Small)
40
+ ├── Embedding: ParamtatvaEmbedding (phonetic graph-aware)
41
+ ├── Layers: 6 × TransformerBlock
42
+ │ ├── Attention: Multi-Head + Pratyāhāra Bias
43
+ │ ├── FFN: GELU activation
44
+ │ └── Norm: LayerNorm + Mā-Bridge
45
+ ├── Final LayerNorm
46
+ └── LM Head
47
+ ```
48
+
49
+ | Parameter | Value |
50
+ |-----------|-------|
51
+ | Parameters | ~10M |
52
+ | Hidden dim | 256 |
53
+ | Layers | 6 |
54
+ | Attention heads | 8 |
55
+ | Intermediate dim | 1024 |
56
+ | Max sequence length | 1024 |
57
+ | Activation | GELU |
58
+
59
+ ## Intended Use
60
+
61
+ This model is released for **research and academic purposes**. It demonstrates the viability of phonetically-grounded language modeling using ancient linguistic frameworks.
62
+
63
+ ### Recommended Uses
64
+ - Research into phonetic/linguistic priors for language models
65
+ - Studies on Sanskrit computational linguistics
66
+ - Mathematical reasoning experiments
67
+ - Exploration of alternative positional encoding schemes
68
+
69
+ ### Out-of-Scope Uses
70
+ - Production/commercial applications (requires separate license)
71
+ - Safety-critical systems
72
+ - Any use that violates the license terms
73
+
74
+ ## Training
75
+
76
+ The model was trained using the ParamTatva training pipeline. The training methodology, loss functions, and data curation are proprietary. Only the resulting model weights are released.
77
+
78
+ **Note**: The full Resonance Learning System (including the proprietary ResonanceEncoder) is NOT included in this release. This release contains only the standard ParamtatvaTransformer weights.
79
+
80
+ ## How to Use
81
+
82
+ ```python
83
+ import torch
84
+ from safetensors.torch import load_file
85
+
86
+ # Load weights
87
+ state_dict = load_file("model.safetensors")
88
+
89
+ # The model uses a custom architecture — see paramtatva_transformer.py
90
+ # for the full model class definition.
91
+ print(f"Parameters: {sum(v.numel() for v in state_dict.values()):,}")
92
+ ```
93
+
94
+ ## Limitations
95
+
96
+ - This is a **small** model (~10M parameters) — intended as a proof of concept
97
+ - The model was trained on a limited dataset
98
+ - Performance on downstream tasks has not been extensively benchmarked
99
+ - The proprietary resonance components are not included
100
+
101
+ ## Citation
102
+
103
+ ```bibtex
104
+ @misc{paramtatva2026rlm,
105
+ title={ParamTatva RLM: A Phonetically-Grounded Language Model
106
+ Based on the Maheshwara Sutras},
107
+ author={ParamTatva.org},
108
+ year={2026},
109
+ url={https://huggingface.co/paramtatva/rlm-small-v1}
110
+ }
111
+ ```
112
+
113
+ ## License
114
+
115
+ This model is released under the **ParamTatva Restricted Use License v1.0**:
116
+ - ✅ Research and academic use
117
+ - ✅ Non-commercial applications
118
+ - ✅ Fine-tuning for research
119
+ - ❌ Commercial use (requires written agreement)
120
+ - ❌ Reverse engineering of training methodology
121
+
122
+ See [LICENSE](LICENSE) for full terms.
123
+
124
+ ## Contact
125
+
126
+ - **Commercial licensing**: licensing@paramtatva.org
127
+ - **Research inquiries**: research@paramtatva.org
128
+ - **Website**: [paramtatva.org](https://paramtatva.org)
config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ParamtatvaTransformer"
4
+ ],
5
+ "model_type": "paramtatva-rlm",
6
+ "vocab_size": 1224,
7
+ "hidden_size": 128,
8
+ "num_hidden_layers": 4,
9
+ "num_attention_heads": 2,
10
+ "intermediate_size": 512,
11
+ "max_position_embeddings": 1024,
12
+ "hidden_act": "gelu",
13
+ "hidden_dropout_prob": 0.1,
14
+ "attention_probs_dropout_prob": 0.1,
15
+ "layer_norm_eps": 1e-06,
16
+ "initializer_range": 0.02,
17
+ "torch_dtype": "float32",
18
+ "transformers_version": "4.40.0",
19
+ "paramtatva_config": {
20
+ "use_graph_embeddings": true,
21
+ "use_pratyahara_bias": true,
22
+ "use_ma_bridge": true,
23
+ "phonetic_basis": "maheshwara_sutras"
24
+ }
25
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e345e1eb32297a516fe4ce74bb5c93ffea5fc76338b19b72758511a07c35bce
3
+ size 4121338