File size: 2,379 Bytes
1a835ec
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
351a5c3
1a835ec
 
 
 
c5a3be6
1a835ec
 
 
 
 
 
 
c5a3be6
1a835ec
 
 
c5a3be6
1a835ec
 
 
 
4f3747d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8d14b1a
4f3747d
 
 
 
1a835ec
 
 
789b1a6
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
---
license: apache-2.0
datasets:
- TopAI-1/WebText-5
- TopAI-1/Reddit-WebText
- TopAI-1/Syntetic-Data-1
- TopAI-1/Minecraft-WebText-2
language:
- en
- he
pipeline_tag: text-generation
tags:
- art
- code
- agent
- text-generation-inference
- merge
- moe
library_name: transformers
---

# MCGPT-1: Mixture of Experts (MoE) Language Model

**MCGPT-1** is a custom-built MoE model developed by **TopAI-IL**. It is designed to demonstrate specialized knowledge in Minecraft, Reddit-style conversations, and model self-identity.

## Model Details
- **Architecture:** Mixture of Experts (MoE)
- **Total Experts:** 4
- **Layers:** 4
- **Attention Heads:** 8
- **Hidden Size:** 256
- **Training Domains:** 1. Identity (TopAI-IL)
  2. Minecraft Technical Data & Guides
  3. Reddit/Web Slang & Conversations
  4. General Hebrew/English Knowledge
  5. Instructions Syntetic-Data

## How to use
This model uses a custom architecture (`mcgpt`). To run inference, ensure you include the architecture class in your code or use the `trust_remote_code=True` flag if the modeling script is provided.

## Use example

  ```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model_id = "TopAI-1/MCGPT-1" 

# 2. load the model and tokienizer
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id, 
    trust_remote_code=True, 
    torch_dtype=torch.float32
)

# 3. GPU If have
device = "cuda" if torch.cuda.is_available() else "cpu"
model.to(device)
model.eval()

# 4. fast text generation
def generate(prompt, max_new_tokens=50):
    inputs = tokenizer(prompt, return_tensors="pt").to(device)
    with torch.no_grad():
        outputs = model.generate(
            **inputs, 
            max_new_tokens=max_new_tokens, 
            do_sample=True, 
            top_k=50, 
            temperature=0.8,
            pad_token_id=tokenizer.eos_token_id
        )
    return tokenizer.decode(outputs[0], skip_special_tokens=True)

# Infrence Check
print("Testing MCGPT-1 from Hub:")
prompt = "use the following search parameters to narrow your results: e.g."
print(generate(prompt))
```

## Capabilities
The model successfully identifies itself as **MCGPT-1** and can switch between experts based on the prompt (e.g., providing Minecraft-related advice when prompted with "help").

**Developed by @ TopAI-IL (2026)**