File size: 2,543 Bytes
d20ab1b
 
16e6009
 
a90d1ee
 
 
 
 
 
 
 
d20ab1b
 
a90d1ee
d20ab1b
 
 
 
a90d1ee
 
d20ab1b
a90d1ee
 
 
 
d20ab1b
a90d1ee
 
d20ab1b
a90d1ee
d20ab1b
 
 
 
a90d1ee
 
d20ab1b
a90d1ee
 
 
d20ab1b
 
a90d1ee
 
d20ab1b
a90d1ee
d20ab1b
 
a90d1ee
 
 
d20ab1b
 
a90d1ee
d20ab1b
a90d1ee
d20ab1b
 
 
a90d1ee
 
 
d20ab1b
a90d1ee
723bf6e
d20ab1b
a90d1ee
 
 
d20ab1b
a90d1ee
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
---
library_name: transformers
tags:
- unsloth
- lora
- gemma
- whatsapp
license: apache-2.0
language:
- es
base_model:
- unsloth/gemma-2-2b-it-bnb-4bit
---

# Model Card for Whatsapp-Finetuned LoRA (Gemma-2-2B-IT-4bit)

## Model Details

### Model Description
This is a LoRA adapter trained on personal WhatsApp conversations, applied on top of **`unsloth/gemma-2-2b-it-bnb-4bit`**, an instruction-tuned Gemma 2B model in 4-bit quantization.  
The adapter specializes the base model toward informal Spanish conversational style, slang, and context typical of WhatsApp chats.  

- **Developed by:** Private (Boni)  
- **Model type:** LoRA adapter for causal language modeling  
- **Language(s):** Spanish (es)  
- **Finetuned from model:** `unsloth/gemma-2-2b-it-bnb-4bit`  

### Model Sources
- **Base model:** [unsloth/gemma-2-2b-it-bnb-4bit](https://huggingface.co/unsloth/gemma-2-2b-it-bnb-4bit)  

---

## Uses

### Direct Use
- Chatbots and assistants that mimic WhatsApp-style Spanish conversations.  
- Experimentation with low-rank adapters on personal datasets.  

### Downstream Use
- Can be merged with the base model for full fine-tuned inference.  
- Can be combined with other adapters for multi-domain behavior.  

### Out-of-Scope Use
- Production deployment without careful filtering (the dataset is personal, informal, and may not generalize).  
- Sensitive domains like healthcare, law, or safety-critical applications.  

---

## Bias, Risks, and Limitations
- The dataset consists of personal WhatsApp conversations, which may include biases, informal expressions, and idiosyncratic slang.  
- The model may reflect private communication style and does not guarantee factual correctness.  
- Limited training size means performance outside the conversational domain is reduced.  

### Recommendations
Users should treat outputs as experimental. Avoid relying on this model for factual or professional contexts.  

---

## How to Get Started with the Model

```python
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

base_model = "unsloth/gemma-2-2b-it-bnb-4bit"
adapter = "Jabr7/Mini-Boni"

tokenizer = AutoTokenizer.from_pretrained(base_model)
model = AutoModelForCausalLM.from_pretrained(base_model, device_map="auto")
model = PeftModel.from_pretrained(model, adapter)

prompt = "Hola, ¿cómo estás?"
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))