putin-bot-lora / README.md
kennethpayne01's picture
Upload README.md with huggingface_hub
160be7c verified
---
license: llama3.3
base_model: meta-llama/Llama-3.3-70B-Instruct
tags:
- lora
- llama
- character-simulation
- political-persona
language:
- en
- ru
---
# Putin Bot - LoRA Adapter
LoRA fine-tuning weights for Llama 3.3 70B Instruct, trained to simulate Vladimir Putin's communication style for strategy simulation games.
## Model Details
- **Base Model**: meta-llama/Llama-3.3-70B-Instruct
- **Training Method**: LoRA (Low-Rank Adaptation)
- **Adapter Size**: 32
- **Training Data**: 3,696 samples from press conferences and interviews (2000-2024)
- **Training Platform**: Google Cloud Vertex AI
- **Training Cost**: ~$40-120
- **Adapter Size**: 1.5GB
## Training Data Sources
- Kremlin press conference transcripts (575 documents)
- Tucker Carlson interview (Feb 2024)
- Oliver Stone interviews (2017)
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
# Load base model
base_model = "meta-llama/Llama-3.3-70B-Instruct"
model = AutoModelForCausalLM.from_pretrained(
base_model,
torch_dtype="auto",
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(base_model)
# Load LoRA adapter
model = PeftModel.from_pretrained(model, "kennethpayne01/putin-bot-lora")
# Generate
messages = [
{"role": "system", "content": "You are Vladimir Putin, President of Russia."},
{"role": "user", "content": "What is your view on NATO expansion?"}
]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True)
outputs = model.generate(inputs, max_new_tokens=512, temperature=0.7)
response = tokenizer.decode(outputs[0][inputs.shape[1]:], skip_special_tokens=True)
print(response)
```
## Hardware Requirements
- **Full precision**: ~140GB VRAM (2-4x A100 80GB)
- **4-bit quantization**: ~40GB VRAM (1x A100 80GB)
- **8-bit quantization**: ~70GB VRAM (1x A100 80GB)
## Training Configuration
- Learning rate: 0.0001
- Epochs: 3
- LoRA rank (r): 32
- LoRA alpha: 64
- Target modules: All attention layers
- Training time: ~2-4 hours
- Platform: Vertex AI with A100 GPUs
## Limitations
- Model trained on public statements; may not reflect private views
- Data range 2000-2024; current events after Dec 2024 not included
- English translations may lose nuance from original Russian
- Designed for simulation/entertainment, not policy analysis
## Ethical Considerations
This model is created for strategy simulation games and educational purposes. It should not be used to:
- Spread misinformation or propaganda
- Impersonate real individuals for deception
- Generate harmful or misleading content
## License
This adapter is released under the Llama 3.3 license. See base model license for details.
## Citation
```bibtex
@misc{putin-bot-lora,
author = {Your Name},
title = {Putin Bot LoRA Adapter},
year = {2025},
publisher = {Hugging Face},
howpublished = {\url{https://huggingface.co/kennethpayne01/putin-bot-lora}}
}
```
## Repository
Full code and training pipeline: [GitHub Repository](#)