File size: 1,716 Bytes
5e36f43
4ff4745
 
 
37ce44b
 
 
 
 
 
 
4ff4745
37ce44b
5e36f43
4ff4745
5e36f43
 
4ff4745
5e36f43
4ff4745
5e36f43
4ff4745
5e36f43
4ff4745
 
 
 
 
 
5e36f43
 
 
4ff4745
 
 
 
 
 
5e36f43
4ff4745
5e36f43
4ff4745
 
 
5e36f43
4ff4745
 
5e36f43
4ff4745
 
 
 
 
5e36f43
4ff4745
5e36f43
4ff4745
 
 
 
 
5e36f43
4ff4745
5e36f43
4ff4745
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
---
license: gemma
base_model: google/gemma-2-2b-it
tags:
  - gemma2
  - fine-tuned
  - qlora
  - technical-assistant
  - aws
  - security
  - finance
language:
  - en
library_name: transformers
pipeline_tag: text-generation
---

# gemma2-2b-technical-assistant

Fine-tuned Gemma 2 2B IT model for personalized technical assistance.

## Model Description

This model is a QLoRA fine-tuned version of `google/gemma-2-2b-it`, specialized for:
- AWS cloud security guidance
- FastAPI/Python backend development
- Finance application development
- Kubernetes workload management
- ISO 27001:2022 compliance

## Training Details

- **Base Model:** google/gemma-2-2b-it
- **Fine-tuning Method:** QLoRA (4-bit quantization)
- **LoRA Rank:** 16
- **LoRA Alpha:** 32
- **Training Epochs:** 5
- **Hardware:** Google Colab T4 GPU

## Usage

### Transformers
```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("satejh/gemma2-2b-technical-assistant")
tokenizer = AutoTokenizer.from_pretrained("satejh/gemma2-2b-technical-assistant")

prompt = "<start_of_turn>user\nWhat database should I use?<end_of_turn>\n<start_of_turn>model\n"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0]))
```

### Ollama

Download the GGUF file and Modelfile from this repo, then:
```bash
ollama create gemma2-2b-technical-assistant -f Modelfile
ollama run gemma2-2b-technical-assistant
```

## Intended Use

This model is designed as a personalized technical assistant with:
- Security-first approach
- Read-only database interactions
- Direct, actionable responses
- AWS and Kubernetes expertise