Update README.md
Browse files
README.md
CHANGED
|
@@ -1,78 +1,65 @@
|
|
| 1 |
-
#
|
| 2 |
|
| 3 |
-
|
| 4 |
|
| 5 |
-
|
| 6 |
-
- **Training:** Train the model on tokenized data.
|
| 7 |
-
- **Sampling:** Generate text from trained models.
|
| 8 |
-
- **Exporting:** Save models and minimal tokenizer configurations in formats compatible with Hugging Face.
|
| 9 |
-
- **Knowledge Distillation:** Train a smaller student model using a larger teacher model.
|
| 10 |
-
- **Fine-Tuning:** Adapt a distilled model on conversational data (from local sources or directly from Hugging Face).
|
| 11 |
|
| 12 |
-
|
| 13 |
|
| 14 |
-
|
| 15 |
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
- [Model Overview](#model-overview)
|
| 19 |
-
- [Architecture](#architecture)
|
| 20 |
-
- [Installation](#installation)
|
| 21 |
-
- [Usage](#usage)
|
| 22 |
-
- [Tokenization](#tokenization)
|
| 23 |
-
- [Training](#training)
|
| 24 |
-
- [Sampling](#sampling)
|
| 25 |
-
- [Exporting](#exporting)
|
| 26 |
-
- [Knowledge Distillation](#knowledge-distillation)
|
| 27 |
-
- [Fine-Tuning](#fine-tuning)
|
| 28 |
-
- [Local Fine-Tuning](#local-fine-tuning)
|
| 29 |
-
- [Hugging Face Fine-Tuning](#hugging-face-fine-tuning)
|
| 30 |
-
- [Hyperparameters](#hyperparameters)
|
| 31 |
-
- [Contributing](#contributing)
|
| 32 |
-
- [License](#license)
|
| 33 |
-
- [Acknowledgements](#acknowledgements)
|
| 34 |
|
| 35 |
---
|
| 36 |
|
| 37 |
-
##
|
| 38 |
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
|
|
|
|
|
|
|
| 47 |
|
| 48 |
---
|
| 49 |
|
| 50 |
-
##
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 51 |
|
| 52 |
-
|
| 53 |
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
| 57 |
-
- **MLP (Feed-Forward Network):** Utilizes GELU activation for non-linearity.
|
| 58 |
-
- **Residual Connections and Layer Normalization:** Stabilize training and improve convergence.
|
| 59 |
-
- **Final Projection Layer:** Maps embeddings to logits over the vocabulary.
|
| 60 |
|
| 61 |
-
|
|
|
|
|
|
|
|
|
|
| 62 |
|
| 63 |
---
|
| 64 |
|
| 65 |
-
##
|
| 66 |
|
| 67 |
-
|
| 68 |
|
| 69 |
-
|
| 70 |
-
- transformers
|
| 71 |
-
- datasets
|
| 72 |
-
- tqdm
|
| 73 |
-
- safetensors (for export functionality)
|
| 74 |
|
| 75 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 76 |
|
| 77 |
-
|
| 78 |
-
|
|
|
|
| 1 |
+
# 🧠 TNSA AI
|
| 2 |
|
| 3 |
+
**Welcome to TNSA AI on Hugging Face 🤖**
|
| 4 |
|
| 5 |
+
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
|
| 7 |
+
## 🌟 About Us
|
| 8 |
|
| 9 |
+
**TNSA AI** is a frontier research group focused on building open, powerful, and human-aligned **Artificial General Intelligence (AGI)**. From cutting-edge language models to multimodal systems, our mission is to create intelligent systems that empower humanity — not just compete with it.
|
| 10 |
|
| 11 |
+
We build from scratch, at scale, and with purpose.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
|
| 13 |
---
|
| 14 |
|
| 15 |
+
## 🏗️ Core Projects
|
| 16 |
|
| 17 |
+
| 🔬 Project | 🚀 Description |
|
| 18 |
+
|------------------|---------------------------------------------------------------------------------|
|
| 19 |
+
| **NGen Series** | High-performance transformer models (NGen1, NGen2, NGen3Jax_N, and more) |
|
| 20 |
+
| **Tokenize2** | Our in-house tokenizer built on optimized Byte-Pair Encoding (BPE) |
|
| 21 |
+
| **DiffuseX2** | Industry-level Stable Diffusion architecture for powerful image generation |
|
| 22 |
+
| **Adhyaapak-2** | Personalized Teaching AI powered by LLM + vision/audio understanding |
|
| 23 |
+
| **Neura** | An AGI-grade modular architecture with memory, deliberation & vision modules |
|
| 24 |
+
| **ARCH-X 9** | ML framework architecture for scalable multimodal development |
|
| 25 |
+
| **StellarTTS** | Text-to-speech engine based on Wi LLM capable of human-like speech synthesis |
|
| 26 |
+
| **SongGEN** | AI music generation engine inspired by OpenAI’s Jukebox |
|
| 27 |
|
| 28 |
---
|
| 29 |
|
| 30 |
+
## 📦 Featured Models & Tools
|
| 31 |
+
|
| 32 |
+
- `tnsa-ai/NGen2.1Base`
|
| 33 |
+
- `tnsa-ai/Tokenize2`
|
| 34 |
+
- `tnsa-ai/DiffuseX2-mini`
|
| 35 |
+
- `tnsa-ai/Adhyaapak-2-Beta`
|
| 36 |
+
- `tnsa-ai/StellarTTS-Nano`
|
| 37 |
+
- `tnsa-ai/SongGEN-Core`
|
| 38 |
|
| 39 |
+
> All models are trained or fine-tuned using our custom pipelines and follow the [TNSA Standard].
|
| 40 |
|
| 41 |
+
---
|
| 42 |
+
|
| 43 |
+
## 🔧 Dev Philosophy
|
|
|
|
|
|
|
|
|
|
| 44 |
|
| 45 |
+
- ✅ Fully open-sourced codebases (no black boxes)
|
| 46 |
+
- 🔁 Reinvent core components when needed
|
| 47 |
+
- 🔬 Prioritize interpretability, capability, and safety
|
| 48 |
+
- 🧪 Models designed for education, research, and real-world use
|
| 49 |
|
| 50 |
---
|
| 51 |
|
| 52 |
+
## 🧑💻 Join Our Movement
|
| 53 |
|
| 54 |
+
We’re not a company. We’re a **mission**.
|
| 55 |
|
| 56 |
+
🚀 If you believe in building AGI for collective good — not just profit — you’re one of us.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 57 |
|
| 58 |
+
📫 Reach out: tnsaresearch@proton.me
|
| 59 |
+
🔗 Website (Coming Soon): [TNSA.ai](https://tnsa.ai)
|
| 60 |
+
🌐 Follow our open work here and contribute!
|
| 61 |
+
|
| 62 |
+
---
|
| 63 |
|
| 64 |
+
> “A mind once stretched by a new idea never returns to its original dimensions.”
|
| 65 |
+
> — **TNSA AI**
|