File size: 2,834 Bytes
9685ec3
6f48ea9
9685ec3
 
6f48ea9
 
 
 
 
 
 
 
316e9f4
6f48ea9
316e9f4
9685ec3
 
316e9f4
9685ec3
316e9f4
9685ec3
316e9f4
9685ec3
316e9f4
 
9685ec3
316e9f4
 
 
 
9685ec3
316e9f4
 
 
 
9685ec3
316e9f4
9685ec3
316e9f4
 
 
 
 
9685ec3
316e9f4
9685ec3
316e9f4
9685ec3
316e9f4
 
 
 
 
9685ec3
316e9f4
 
9685ec3
316e9f4
 
9685ec3
316e9f4
 
 
 
9685ec3
316e9f4
9685ec3
316e9f4
 
9685ec3
316e9f4
 
9685ec3
316e9f4
 
 
 
9685ec3
 
 
316e9f4
 
 
 
 
9685ec3
316e9f4
9685ec3
316e9f4
 
 
9685ec3
316e9f4
9685ec3
316e9f4
9685ec3
316e9f4
9685ec3
6f48ea9
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
---
license: agpl-3.0
base_model: unsloth/DeepSeek-R1-0528-Qwen3-8B
tags:
- marxism-leninism
- grpo
- llama-cpp
- ollama
- political-education
- marxism
- communism
- political-extremism
language:
- en
pipeline_tag: text-generation
---

# MLMLML - Machine Learning Marxist-Leninist Models of Language

A GRPO fine-tuned language model for Marxist-Leninist political education and analysis.

## Model Description

This model is fine-tuned from `unsloth/DeepSeek-R1-0528-Qwen3-8B` using Group Relative Policy Optimization (GRPO) 
on a curated dataset of Marxist-Leninist Q&A pairs from [ProleWiki](https://en.prolewiki.org/).

The training rewards:
- **Ideological firmness**: Clear positions grounded in material analysis
- **Coherence**: Self-consistent, well-structured responses
- **Accuracy**: Faithful to Marxist-Leninist theory and historical evidence

The training penalizes:
- "Both-sidesing" and false balance
- Hedging and evasive language
- Bourgeois framing and ahistorical claims

## Writing Style

Following Chairman Mao's guidance in "Oppose Stereotyped Party Writing":
- **Vigorous, lively, fresh and forceful** - never drab or stereotyped
- **Audience-aware** - "When shooting an arrow, one must aim at the target"
- **Investigation-based** - "No investigation, no right to speak"
- **Clear positions** - FOR and AGAINST, using scientific argument

## Usage

### Download and Convert to GGUF

```bash
# Clone the repo
git lfs install
git clone https://huggingface.co/percyraskova/MLMLML
cd MLMLML

# Convert to GGUF (requires llama.cpp)
python ~/llama.cpp/convert_hf_to_gguf.py . --outfile MLMLML-F16.gguf --outtype f16

# Quantize to Q4_K_M
~/llama.cpp/build/bin/llama-quantize MLMLML-F16.gguf MLMLML-Q4_K_M.gguf Q4_K_M

# Create Ollama model
ollama create mlmlml -f Modelfile
ollama run mlmlml
```

### Direct with Transformers

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("percyraskova/MLMLML")
tokenizer = AutoTokenizer.from_pretrained("percyraskova/MLMLML")

inputs = tokenizer("What is imperialism?", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0]))
```

## Training Details

- **Base model**: unsloth/DeepSeek-R1-0528-Qwen3-8B
- **Method**: GRPO (Group Relative Policy Optimization)
- **Dataset**: ProleWiki Q&A pairs (~4500 samples)
- **Epochs**: 2
- **Hardware**: NVIDIA A100 80GB

## Limitations

This model is designed for educational purposes about Marxist-Leninist theory and analysis.
It takes clear ideological positions and is not intended to be "neutral" on class struggle,
imperialism, or other questions where Marxism-Leninism has definite answers.

## License

Apache 2.0

## Citation

If you use this model, please cite ProleWiki as the source of training data.