0xAbhi commited on
Commit
669b264
·
verified ·
1 Parent(s): eead72f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +113 -17
README.md CHANGED
@@ -1,28 +1,124 @@
 
 
 
 
 
 
 
 
1
  # Kavi
2
 
3
- ## Version
4
- **v0.5**
5
 
6
- ## What’s New
7
- - Improved reasoning and clarity
8
- - Fine-tuned on curated life-advice conversations
9
- - Better consistency in responses
10
 
11
- ## Base Model
12
- - TinyLlama/TinyLlama-1.1B-Chat-v1.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
  ## Training
15
- - QLoRA fine-tuning
16
- - Tesla T4 GPU
17
- - English life-advice dataset
 
 
 
 
 
 
18
 
19
  ## Intended Use
20
- - Life guidance
21
- - Educational conversations
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
 
23
  ## Limitations
24
- - Not professional advice
25
- - Tamil support planned in future versions
26
 
27
- ## License
28
- Apache 2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ - ta
6
+ base_model:
7
+ - TinyLlama/TinyLlama-1.1B-Chat-v1.0
8
+ ---
9
  # Kavi
10
 
11
+ **Kavi** is a compact, instruction-tuned language model focused on delivering **clear, simple, and practical life advice**. It is designed to be approachable, consistent, and easy to deploy, making it suitable for educational and personal guidance use cases.
 
12
 
13
+ ---
 
 
 
14
 
15
+ ## Model Details
16
+
17
+ - **Model name:** Kavi
18
+ - **Version:** v0.5
19
+ - **Author:** Abhinivesh (0xAbhi)
20
+ - **Base model:** TinyLlama/TinyLlama-1.1B-Chat-v1.0
21
+ - **Model type:** Causal Language Model
22
+ - **Language:** English (Tamil support planned)
23
+ - **License:** Apache 2.0
24
+
25
+ ---
26
+
27
+ ## What’s New in v0.5
28
+
29
+ - Improved reasoning and response clarity
30
+ - Fine-tuned on curated life-advice conversations
31
+ - More consistent and structured answers
32
+ - Better alignment for supportive, guidance-oriented dialogue
33
+
34
+ ---
35
 
36
  ## Training
37
+
38
+ - **Fine-tuning method:** QLoRA (Low-Rank Adaptation)
39
+ - **Hardware:** NVIDIA Tesla T4 (Google Colab)
40
+ - **Precision:** 4-bit base model with LoRA adapters (merged)
41
+ - **Dataset:** Curated English life-advice and guidance conversations
42
+
43
+ The model was fine-tuned to improve conversational quality, tone, and practical reasoning without significantly increasing model size or inference cost.
44
+
45
+ ---
46
 
47
  ## Intended Use
48
+
49
+ Kavi is intended for:
50
+
51
+ - General life guidance and self-reflection prompts
52
+ - Educational and learning-oriented conversations
53
+ - Supportive, non-clinical advice interactions
54
+ - Chatbots focused on clarity, simplicity, and encouragement
55
+
56
+ ---
57
+
58
+ ## Out-of-Scope Use
59
+
60
+ Kavi is **not** intended for:
61
+
62
+ - Medical, legal, or financial advice
63
+ - Crisis counseling or mental health diagnosis
64
+ - Professional or authoritative decision-making systems
65
+
66
+ ---
67
 
68
  ## Limitations
 
 
69
 
70
+ - Not a substitute for professional advice
71
+ - English-first; Tamil reasoning and responses are planned for future releases
72
+ - As a small (1.1B) model, complex multi-step reasoning may be limited
73
+
74
+ ---
75
+
76
+ ## Ethical Considerations
77
+
78
+ - Responses are generated based on patterns learned from training data and may not always be accurate or complete
79
+ - Users should apply human judgment when interpreting outputs
80
+ - The model does not possess awareness, intent, or personal understanding
81
+
82
+ ---
83
+
84
+ ## Usage Example
85
+
86
+ ```python
87
+ from transformers import pipeline
88
+
89
+ pipe = pipeline(
90
+ "text-generation",
91
+ model="0xAbhi/kavi",
92
+ device_map="auto",
93
+ )
94
+
95
+ pipe("I feel stuck in life and don’t know what to do next.")
96
+ ```
97
+ ---
98
+
99
+ ## Future Work
100
+
101
+ Incremental Tamil language fine-tuning
102
+
103
+ Multilingual reasoning improvements
104
+
105
+ Additional alignment for emotional nuance and long-form guidance
106
+
107
+ Citation
108
+
109
+ ```python
110
+ @misc{tinyllama,
111
+ title={TinyLlama: An Open-Source Small Language Model},
112
+ author={TinyLlama Team},
113
+ year={2023},
114
+ url={https://huggingface.co/TinyLlama}
115
+ }
116
+ ```
117
+
118
+ ## Acknowledgements
119
+
120
+ TinyLlama team for the base model
121
+
122
+ Hugging Face ecosystem (Transformers, PEFT, TRL)
123
+
124
+ Unsloth for efficient fine-tuning on low-resource hardware