JiRack_empty / README.md
kgrabko's picture
Update README.md
5f2979e verified
---
license: gpl-3.0
---
# JiRack GPT-2 Initial Weights
This file is strictly intended for saving the **initial weights (checkpoint)** of the JiRack GPT model.
The model is **"clean"**: it contains no data and has never undergone any pre-training.
- Powered by CMS Manhattan’s cutting-edge Vision-BERT architecture.
It is engineered to be a maximally safe and robust base for **training from scratch** for specialized, smaller models, such as:
- **SPAM Detection Systems**
- **FRAUD Detection Models**
- **Background Check (BG Check) Models**
_A product of CMS Manhattan._
---
## Tokenizer Choices
- For English: **GPT-2 Hugging Face tokenizer**
- For multilingual use: **BERT tokenizer** from the Hugging Face library
---
## Model Architecture Details
### GPT-2 Architecture (Classic, Transformer-like)
```
CustomEmbedding
FrozenSignatureLayer
LearnedPositionalEmbedding
[TransformerBlock]
β”œβ”€β”€ MultiHeadAttention
β”œβ”€β”€ LayerNorm
β”œβ”€β”€ LayerNorm
β”œβ”€β”€ FFN
β”œβ”€β”€ Linear
β”œβ”€β”€ Activation: GELU
└── Linear
LayerNorm
Linear
```
---
## Model Checkpoint File Explanations
### **12-head Attention Model**
**Parameters:**
- `VOCAB_SIZE = 50257`
- `MODEL_DIM = 768`
- `NUM_HEADS = 12`
- `NUM_LAYERS = 6`
- `MAX_SEQ_LEN = 8192`
- `FFN_HIDDEN_DIM = 4 * MODEL_DIM`
- `HEAD_DIM = MODEL_DIM // NUM_HEADS`
**File:**
`JiRack_H12_L6_V50257_D768_MSL8192_FF768x4.pt`
---
### **6-head Attention Model**
**Parameters:**
- `VOCAB_SIZE = 50257`
- `MODEL_DIM = 768`
- `NUM_HEADS = 6`
- `NUM_LAYERS = 6`
- `MAX_SEQ_LEN = 8192`
- `FFN_HIDDEN_DIM = 4 * MODEL_DIM`
- `HEAD_DIM = MODEL_DIM // NUM_HEADS`
**File:**
`JiRack_H6_L6_V50257_D768_MSL8192_FF768x4.pt`
- So About PyTorch script . You can use Pytorch script for AI classification task .
- Do not Jit for Chatbot task . Use just state dict PyTorch for GPT (Chatbot) tasks
---
See other models with same patterns for read parameters
# install tokenizer before run
---
- mkdir -p tokenizer
- wget -O tokenizer/tokenizer.json https://huggingface.co/gpt2/resolve/main/tokenizer.json
- wget -O tokenizer/vocab.json https://huggingface.co/gpt2/resolve/main/vocab.json
- wget -O tokenizer/merges.txt https://huggingface.co/gpt2/resolve/main/merges.txt
- wget -O tokenizer/tokenizer_config.json https://huggingface.co/gpt2/resolve/main/tokenizer_config.json
---
### JiRack RAG System
- It is microservice architecture with API Gateway and Service Discovery
- Framework Spring boot and Google embeddings model for JiRack RAG System with Chatbot and JiRach model deployment with docker scipt
- video https://www.youtube.com/watch?v=vHClQu76kMc
- RAG System https://bitbucket.org/cmsmanhattan/rag/src/main/
# Copyright Office
- From:
- cop-rc@loc.gov
- To:
- konstantin.grabko@yahoo.com
- Mon, Dec 15 at 7:31 AM
- THIS IS AN AUTOMATED EMAIL. PLEASE DO NOT REPLY.
- Thank you for submitting your registration claim using the Electronic Copyright Office (ECO) System.
- The following files were successfully uploaded for service request 1-15058193231
- File Name :jirack_gpt2_class_pytorch.zip
- File Size :2993 KB
- Date/Time :12/15/2025 7:27:48 AM
- [THREAD ID: 1-6X1C895]
- United States Copyright Office
---
Welcome to ask to design your corp model over 33B or 70B or more parameters
##
CMS Manhattan
Copyright Β© 2002–2026