DSDUDEd commited on
Commit
4eb6f1f
Β·
verified Β·
1 Parent(s): ae427a0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +29 -14
README.md CHANGED
@@ -1,23 +1,38 @@
1
- # Dave – Fully Custom AI Model
 
 
2
 
3
- **Model type:** Fully Custom Transformer from Scratch
4
- **Purpose:** Conversational AI with a unique personality: fun, smart, coding-savvy, sometimes sad or depressed, talks about drugs and alcohol, lost his wife, happy about his kid, sometimes cringe, close friends, swears often.
5
 
6
- **Training:**
7
- - 3000 generated prompts and responses
8
- - Custom tokenizer and transformer trained from scratch
9
- - Trained entirely in Google Colab
10
- - Saved as `pytorch_model.bin`, `config.json`, `tokenizer.json`
11
 
12
- **Example usage:**
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
  ```python
15
  from transformers import AutoTokenizer, AutoModelForCausalLM
16
 
17
  tokenizer = AutoTokenizer.from_pretrained("DSDUDEd/Dave")
18
  model = AutoModelForCausalLM.from_pretrained("DSDUDEd/Dave")
19
-
20
- prompt = "Write a Python function to reverse a string."
21
- inputs = tokenizer(prompt, return_tensors="pt")
22
- outputs = model.generate(**inputs, max_new_tokens=64)
23
- print(tokenizer.decode(outputs[0], skip_special_tokens=True))
 
1
+ # ──────────────────────────────
2
+ # Prepare "Dave" Hugging Face-ready repo
3
+ # ──────────────────────────────
4
 
5
+ !pip install -q torch transformers tokenizers
 
6
 
7
+ import os
8
+ import json
9
+ import torch
10
+ import torch.nn as nn
11
+ from tokenizers import Tokenizer, models, pre_tokenizers, trainers, processors
12
 
13
+ # -----------------------------
14
+ # Settings
15
+ # -----------------------------
16
+ model_name = "dave"
17
+ output_dir = f"./{model_name}"
18
+ os.makedirs(output_dir, exist_ok=True)
19
+ print(f"Using folder: {output_dir}")
20
+
21
+ # -----------------------------
22
+ # README.md for Hugging Face
23
+ # -----------------------------
24
+ readme_content = """# Dave – Fully Custom AI Model
25
+
26
+ Dave is a fully custom-built transformer trained from scratch.
27
+ - Personality: fun, smart, coding-savvy, depressed, talks about drugs/alcohol, lost wife, has a kid, sometimes cringe, has close friends, swears frequently.
28
+ - 3000 prompts/responses were used for training.
29
+ - Architecture: custom Transformer implementation.
30
+ - Trained from scratch.
31
+
32
+ ## How to use
33
 
34
  ```python
35
  from transformers import AutoTokenizer, AutoModelForCausalLM
36
 
37
  tokenizer = AutoTokenizer.from_pretrained("DSDUDEd/Dave")
38
  model = AutoModelForCausalLM.from_pretrained("DSDUDEd/Dave")