JiRackBaseDataset / README.md
kgrabko's picture
Update README.md
b3b3f43 verified
---
viewer: false
license: other
license_name: cms-manhattan-jirack-v1.2
license_link: LICENSE
language:
- en
- ru
- fr
- de
- cn
- jp
tags:
- llama
pipeline_tag: text-generation
---
# 💎 JiRack Base dataset for 1.5B model
**Dataset:** the dataset formated for JiRack tokenizer . I recommend initializing the model with a 4K context window for initial stability, followed by scaling to 8K context using specialized JiRack 8K datasets. This two-stage approach ensures robust positional encoding before extending the model's long-range dependency.
**Time:** JiRack 1.5B: High-Efficiency Financial Modeling
- We are training a compact 1.5B parameter model on an extensive 11 billion token corpus. By training on a token-to-parameter ratio of nearly 7:1, we achieve exceptional knowledge density and reasoning capabilities in a lightweight architecture.
- Performance: JiRack Ternary Pro 1.5b about 28–36 hours per epoch on NVIDIA BlackWell 96 Gb VRAM
- Performance: JiRack Ternary Pro 10b about 7-9 days per epoch on NVIDIA BlackWell 96 Gb VRAM
- Optimization: Optimized for secure, low-latency banking applications.
**Inventor:** Konstantin Vladimirovich Grabko
**Organization:** CMS Manhattan JiRack Technology
**Official Site:** [www.cmsmanhattan.com](http://www.cmsmanhattan.com)
Designed for Banking and Fintech Institutions
**Banks and Fintech** Build secure, internal models tailored for the banking sector. We provide end-to-end solutions to pre-train models for fraud prevention, spam filtering, risk assessment, and Anti-Money Laundering (AML) detectio
- This is the base checkpoint, evaluated prior to fine-tuning on domain-specific datasets. The primary objective is to validate RoPE (Rotary Positional Embeddings) stability and coherence following the initial pre-training phase.
⚠️ **IMPORTANT NOTICE — PROPRIETARY TECHNOLOGY**
**Allowed:**
- Personal and non-commercial research use only
**Strictly Prohibited without a written commercial license:**
- Any commercial use (SaaS, mobile apps, edge devices, paid services, etc.)
- Creating and distributing derivative models for profit
- Removing or modifying any copyright or legal notices
- Patenting any part of this technology
Commercial users **must** obtain a signed license and pay **5% royalty** on net revenue.
Any unauthorized commercial use will be pursued legally under New York law.
Contact for commercial license: grabko@cmsmanhattan.com
There is fix price for FinTech
## ⚠️ Finch tech AL solution
Custom AI Solutions with JiRack
- Deploy your own secure, high-performance model from scratch. I specialize in delivering the JiRack modern architecture on NVIDIA Clusters, fully optimized for your private datasets.
- Let's build your sovereign AI today. DM for inquiries.
- Please contact to CMS Manhttan for the solution
- # Tesr Tokenizer size !
(venv_ji) root@jirack2:# python -c '
from transformers import AutoTokenizer
tok = AutoTokenizer.from_pretrained("./jirack_code_tokenizer_fixed")
print("Vocab size:", len(tok))
print("pad_token_id:", tok.pad_token_id)
print("eos_token_id:", tok.eos_token_id)
'
- Vocab size: 128259
- pad_token_id: 128001
- eos_token_id: 128001