|
|
--- |
|
|
license: llama3.3 |
|
|
base_model: meta-llama/Llama-3.3-70B-Instruct |
|
|
tags: |
|
|
- lora |
|
|
- llama |
|
|
- character-simulation |
|
|
- political-persona |
|
|
language: |
|
|
- en |
|
|
- ru |
|
|
--- |
|
|
|
|
|
# Putin Bot - LoRA Adapter |
|
|
|
|
|
LoRA fine-tuning weights for Llama 3.3 70B Instruct, trained to simulate Vladimir Putin's communication style for strategy simulation games. |
|
|
|
|
|
## Model Details |
|
|
|
|
|
- **Base Model**: meta-llama/Llama-3.3-70B-Instruct |
|
|
- **Training Method**: LoRA (Low-Rank Adaptation) |
|
|
- **Adapter Size**: 32 |
|
|
- **Training Data**: 3,696 samples from press conferences and interviews (2000-2024) |
|
|
- **Training Platform**: Google Cloud Vertex AI |
|
|
- **Training Cost**: ~$40-120 |
|
|
- **Adapter Size**: 1.5GB |
|
|
|
|
|
## Training Data Sources |
|
|
|
|
|
- Kremlin press conference transcripts (575 documents) |
|
|
- Tucker Carlson interview (Feb 2024) |
|
|
- Oliver Stone interviews (2017) |
|
|
|
|
|
## Usage |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
from peft import PeftModel |
|
|
|
|
|
# Load base model |
|
|
base_model = "meta-llama/Llama-3.3-70B-Instruct" |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
base_model, |
|
|
torch_dtype="auto", |
|
|
device_map="auto" |
|
|
) |
|
|
tokenizer = AutoTokenizer.from_pretrained(base_model) |
|
|
|
|
|
# Load LoRA adapter |
|
|
model = PeftModel.from_pretrained(model, "kennethpayne01/putin-bot-lora") |
|
|
|
|
|
# Generate |
|
|
messages = [ |
|
|
{"role": "system", "content": "You are Vladimir Putin, President of Russia."}, |
|
|
{"role": "user", "content": "What is your view on NATO expansion?"} |
|
|
] |
|
|
|
|
|
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True) |
|
|
outputs = model.generate(inputs, max_new_tokens=512, temperature=0.7) |
|
|
response = tokenizer.decode(outputs[0][inputs.shape[1]:], skip_special_tokens=True) |
|
|
print(response) |
|
|
``` |
|
|
|
|
|
## Hardware Requirements |
|
|
|
|
|
- **Full precision**: ~140GB VRAM (2-4x A100 80GB) |
|
|
- **4-bit quantization**: ~40GB VRAM (1x A100 80GB) |
|
|
- **8-bit quantization**: ~70GB VRAM (1x A100 80GB) |
|
|
|
|
|
## Training Configuration |
|
|
|
|
|
- Learning rate: 0.0001 |
|
|
- Epochs: 3 |
|
|
- LoRA rank (r): 32 |
|
|
- LoRA alpha: 64 |
|
|
- Target modules: All attention layers |
|
|
- Training time: ~2-4 hours |
|
|
- Platform: Vertex AI with A100 GPUs |
|
|
|
|
|
## Limitations |
|
|
|
|
|
- Model trained on public statements; may not reflect private views |
|
|
- Data range 2000-2024; current events after Dec 2024 not included |
|
|
- English translations may lose nuance from original Russian |
|
|
- Designed for simulation/entertainment, not policy analysis |
|
|
|
|
|
## Ethical Considerations |
|
|
|
|
|
This model is created for strategy simulation games and educational purposes. It should not be used to: |
|
|
- Spread misinformation or propaganda |
|
|
- Impersonate real individuals for deception |
|
|
- Generate harmful or misleading content |
|
|
|
|
|
## License |
|
|
|
|
|
This adapter is released under the Llama 3.3 license. See base model license for details. |
|
|
|
|
|
## Citation |
|
|
|
|
|
```bibtex |
|
|
@misc{putin-bot-lora, |
|
|
author = {Your Name}, |
|
|
title = {Putin Bot LoRA Adapter}, |
|
|
year = {2025}, |
|
|
publisher = {Hugging Face}, |
|
|
howpublished = {\url{https://huggingface.co/kennethpayne01/putin-bot-lora}} |
|
|
} |
|
|
``` |
|
|
|
|
|
## Repository |
|
|
|
|
|
Full code and training pipeline: [GitHub Repository](#) |
|
|
|