File size: 4,988 Bytes
2293606
3dca452
2293606
 
 
3dca452
 
2293606
 
 
 
 
 
 
 
 
 
 
 
3dca452
 
2293606
3dca452
2293606
 
 
 
3dca452
2293606
3dca452
2293606
3dca452
2293606
3dca452
2293606
3dca452
2293606
3dca452
2293606
 
 
 
3dca452
 
 
2293606
 
 
 
 
 
3dca452
2293606
 
 
 
 
 
 
 
 
 
3dca452
2293606
3dca452
2293606
 
 
3dca452
2293606
 
 
 
3dca452
2293606
 
3dca452
2293606
 
 
 
 
 
 
3dca452
2293606
 
 
3dca452
2293606
3dca452
2293606
 
3dca452
2293606
3dca452
2293606
 
3dca452
2293606
3dca452
2293606
 
 
 
 
 
 
 
 
3dca452
2293606
 
 
3dca452
2293606
3dca452
2293606
 
3dca452
2293606
 
3dca452
2293606
3dca452
2293606
 
 
 
 
3dca452
2293606
3dca452
2293606
3dca452
2293606
3dca452
2293606
3dca452
2293606
3dca452
2293606
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143

---
license: apache-2.0
language:
  - en
library_name: peft
tags:
  - text-generation
  - transformers
  - peft
  - lora
  - qwen
  - qwen2
  - reddit
  - llama-factory
datasets:
  - olmo-data/dolma-v1_6-reddit
base_model: Qwen/Qwen2-0.5B
pipeline_tag: text-generation
---

# Qwen2-0.5B Reddit LoRA Adapter

**Repo:** [iko-01/LLaMA-1](https://huggingface.co/iko-01/LLaMA-1)  
**Base model:** [Qwen/Qwen2-0.5B](https://huggingface.co/Qwen/Qwen2-0.5B)  
**Adapter type:** LoRA (via LLaMA-Factory + QLoRA)  
**Intended use:** Simulating casual, Reddit-style comments, discussions, and thread replies

## Model Description

This is a **LoRA adapter** fine-tuned on top of **Qwen2-0.5B** using a filtered subset of Reddit posts & comments from the Dolma dataset (v1.6 Reddit portion).

The model is trained to generate informal, conversational text typical of Reddit threads β€” including sarcasm, memes references, casual opinions, upvotes/downvotes vibe, and natural thread continuations.

Despite the repository name (`LLaMA-1`), this is **not** a LLaMA model β€” it is purely **Qwen2** architecture.

### Key Characteristics

- Extremely lightweight (only ~0.5B base + small LoRA adapter)
- Runs comfortably on consumer GPUs, laptops, or even decent CPUs
- Fast inference (very suitable for local prototyping, chatbots, Reddit simulators, etc.)
- Casual / internet / meme-friendly tone

## Training Details

- **Framework:** LLaMA-Factory
- **Training method:** QLoRA (4-bit base quantization + LoRA)
- **Dataset size:** ~6,000 high-quality, deduplicated Reddit samples
- **Hardware:** Google Colab T4 (single GPU)
- **Training duration:** β‰ˆ 30 minutes
- **Hyperparameters:**

  | Parameter              | Value     |
  |------------------------|-----------|
  | LoRA rank (r)          | 32        |
  | LoRA alpha             | 64        |
  | Learning rate          | 2e-4      |
  | Batch size             | 2         |
  | Gradient accumulation  | 16        |
  | Epochs                 | 3         |
  | Optimizer              | AdamW     |
  | Warmup ratio           | 0.03      |

## Usage

```bash
pip install -U transformers peft torch accelerate bitsandbytes  # bitsandbytes optional but recommended
```

```python
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

base_model_id = "Qwen/Qwen2-0.5B"
adapter_id    = "iko-01/LLaMA-1"

# Load base model
model = AutoModelForCausalLM.from_pretrained(
    base_model_id,
    torch_dtype=torch.float16,
    device_map="auto",
    trust_remote_code=True
)

# Apply LoRA adapter
model = PeftModel.from_pretrained(model, adapter_id)
model.eval()

tokenizer = AutoTokenizer.from_pretrained(base_model_id, trust_remote_code=True)

# Example prompt
prompt = """Continue this r/AskReddit thread:

After playing for 50 hours I finally"""

messages = [{"role": "user", "content": prompt}]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)

inputs = tokenizer(text, return_tensors="pt").to(model.device)

with torch.no_grad():
    out = model.generate(
        **inputs,
        max_new_tokens=120,
        temperature=0.75,
        top_p=0.92,
        repetition_penalty=1.08,
        do_sample=True
    )

response = tokenizer.decode(out[0][len(inputs.input_ids[0]):], skip_special_tokens=True)
print(response)
```

### Example Outputs

**Prompt:**  
`Complete this r/gaming discussion: After playing for 50 hours I finally`

**Typical model output:**  
`...realized the main story is mid but the side content is actually peak. The open world exploration in the frozen north hits different. Spent like 6 hours just fishing and upgrading my house and I don't even feel bad about it lmao. Anyone else 100% the fishing minigame before the final boss?`

## Limitations & Responsible Use

- **Model size** β€” Being a 0.5B model, it has limited world knowledge, reasoning depth, and coherence over very long contexts compared to 7B+ models.
- **Reddit bias** β€” The training data comes from Reddit β†’ expect informal language, slang, sarcasm, exaggeration, memes, controversial/hot-take opinions, and sometimes toxic phrasing.
- **Hallucinations** β€” Can confidently generate plausible but incorrect facts, especially outside popular Reddit topics.
- **Not for production / sensitive use** β€” Not suitable for factual Q&A, customer support, education, legal/medical advice, or any high-stakes application.
- **English only** β€” The fine-tune was done exclusively on English Reddit content.

Use this model mainly for **creative**, **entertainment**, or **research** purposes (e.g. generating synthetic discussion data, building Reddit-style bots, style transfer experiments).

## Citation / Thanks

If you use this adapter in your work, feel free to mention:

> Fine-tuned with LLaMA-Factory on Qwen2-0.5B using Reddit data from Dolma.

Big thanks to the Qwen team, LLaMA-Factory contributors, and AllenAI (Dolma dataset).

Happy hacking! πŸš€
```