File size: 2,834 Bytes
9685ec3 6f48ea9 9685ec3 6f48ea9 316e9f4 6f48ea9 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 316e9f4 9685ec3 6f48ea9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 |
---
license: agpl-3.0
base_model: unsloth/DeepSeek-R1-0528-Qwen3-8B
tags:
- marxism-leninism
- grpo
- llama-cpp
- ollama
- political-education
- marxism
- communism
- political-extremism
language:
- en
pipeline_tag: text-generation
---
# MLMLML - Machine Learning Marxist-Leninist Models of Language
A GRPO fine-tuned language model for Marxist-Leninist political education and analysis.
## Model Description
This model is fine-tuned from `unsloth/DeepSeek-R1-0528-Qwen3-8B` using Group Relative Policy Optimization (GRPO)
on a curated dataset of Marxist-Leninist Q&A pairs from [ProleWiki](https://en.prolewiki.org/).
The training rewards:
- **Ideological firmness**: Clear positions grounded in material analysis
- **Coherence**: Self-consistent, well-structured responses
- **Accuracy**: Faithful to Marxist-Leninist theory and historical evidence
The training penalizes:
- "Both-sidesing" and false balance
- Hedging and evasive language
- Bourgeois framing and ahistorical claims
## Writing Style
Following Chairman Mao's guidance in "Oppose Stereotyped Party Writing":
- **Vigorous, lively, fresh and forceful** - never drab or stereotyped
- **Audience-aware** - "When shooting an arrow, one must aim at the target"
- **Investigation-based** - "No investigation, no right to speak"
- **Clear positions** - FOR and AGAINST, using scientific argument
## Usage
### Download and Convert to GGUF
```bash
# Clone the repo
git lfs install
git clone https://huggingface.co/percyraskova/MLMLML
cd MLMLML
# Convert to GGUF (requires llama.cpp)
python ~/llama.cpp/convert_hf_to_gguf.py . --outfile MLMLML-F16.gguf --outtype f16
# Quantize to Q4_K_M
~/llama.cpp/build/bin/llama-quantize MLMLML-F16.gguf MLMLML-Q4_K_M.gguf Q4_K_M
# Create Ollama model
ollama create mlmlml -f Modelfile
ollama run mlmlml
```
### Direct with Transformers
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("percyraskova/MLMLML")
tokenizer = AutoTokenizer.from_pretrained("percyraskova/MLMLML")
inputs = tokenizer("What is imperialism?", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0]))
```
## Training Details
- **Base model**: unsloth/DeepSeek-R1-0528-Qwen3-8B
- **Method**: GRPO (Group Relative Policy Optimization)
- **Dataset**: ProleWiki Q&A pairs (~4500 samples)
- **Epochs**: 2
- **Hardware**: NVIDIA A100 80GB
## Limitations
This model is designed for educational purposes about Marxist-Leninist theory and analysis.
It takes clear ideological positions and is not intended to be "neutral" on class struggle,
imperialism, or other questions where Marxism-Leninism has definite answers.
## License
Apache 2.0
## Citation
If you use this model, please cite ProleWiki as the source of training data. |