Jay1121 commited on
Commit
b838204
·
verified ·
1 Parent(s): 384dcd8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +108 -173
README.md CHANGED
@@ -1,220 +1,155 @@
1
  ---
2
  license: apache-2.0
3
- library_name: transformers
 
4
  pipeline_tag: text-generation
5
  tags:
6
  - chat
7
  - instruction-tuned
8
  - korean
9
- ---
10
-
11
-
12
- ---
13
- base_model: Upstage/SOLAR-10.7B-Instruct-v1.0
14
- library_name: peft
15
- pipeline_tag: text-generation
16
- tags:
17
- - base_model:adapter:Upstage/SOLAR-10.7B-Instruct-v1.0
18
  - lora
19
  - sft
20
  - transformers
21
  - trl
 
 
 
22
  ---
23
 
24
- # Model Card for Model ID
25
-
26
- <!-- Provide a quick summary of what the model is/does. -->
27
 
 
 
28
 
 
29
 
30
  ## Model Details
31
 
32
- ### Model Description
33
-
34
- <!-- Provide a longer summary of what this model is. -->
35
-
36
-
37
-
38
  - **Developed by:** Jihee Cho
39
- - **Funded by [optional]:** [More Information Needed]
40
- - **Shared by [optional]:** [More Information Needed]
41
- - **Model type:** [More Information Needed]
42
- - **Language(s) (NLP):** [More Information Needed]
43
- - **License:** [More Information Needed]
44
- - **Finetuned from model [optional]:** [More Information Needed]
45
 
46
- ### Model Sources [optional]
47
 
48
- <!-- Provide the basic links for the model. -->
 
49
 
50
- - **Repository:** [More Information Needed]
51
- - **Paper [optional]:** [More Information Needed]
52
- - **Demo [optional]:** [More Information Needed]
53
-
54
- ## Uses
55
-
56
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
57
-
58
- ### Direct Use
59
-
60
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
61
 
62
- [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
63
 
64
- ### Downstream Use [optional]
65
 
66
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
67
 
68
- [More Information Needed]
 
 
 
 
69
 
70
  ### Out-of-Scope Use
 
 
 
 
71
 
72
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
73
-
74
- [More Information Needed]
75
-
76
- ## Bias, Risks, and Limitations
77
-
78
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
79
-
80
- [More Information Needed]
81
-
82
- ### Recommendations
83
-
84
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
85
-
86
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
87
-
88
- ## How to Get Started with the Model
89
-
90
- Use the code below to get started with the model.
91
-
92
- [More Information Needed]
93
-
94
- ## Training Details
95
-
96
- ### Training Data
97
-
98
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
99
-
100
- [More Information Needed]
101
-
102
- ### Training Procedure
103
-
104
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
105
-
106
- #### Preprocessing [optional]
107
-
108
- [More Information Needed]
109
-
110
-
111
- #### Training Hyperparameters
112
-
113
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
114
-
115
- #### Speeds, Sizes, Times [optional]
116
-
117
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
118
-
119
- [More Information Needed]
120
-
121
- ## Evaluation
122
-
123
- <!-- This section describes the evaluation protocols and provides the results. -->
124
-
125
- ### Testing Data, Factors & Metrics
126
-
127
- #### Testing Data
128
-
129
- <!-- This should link to a Dataset Card if possible. -->
130
-
131
- [More Information Needed]
132
-
133
- #### Factors
134
-
135
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
136
-
137
- [More Information Needed]
138
-
139
- #### Metrics
140
-
141
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
142
-
143
- [More Information Needed]
144
-
145
- ### Results
146
-
147
- [More Information Needed]
148
-
149
- #### Summary
150
-
151
-
152
-
153
- ## Model Examination [optional]
154
-
155
- <!-- Relevant interpretability work for the model goes here -->
156
-
157
- [More Information Needed]
158
-
159
- ## Environmental Impact
160
-
161
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
162
-
163
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
164
-
165
- - **Hardware Type:** [More Information Needed]
166
- - **Hours used:** [More Information Needed]
167
- - **Cloud Provider:** [More Information Needed]
168
- - **Compute Region:** [More Information Needed]
169
- - **Carbon Emitted:** [More Information Needed]
170
-
171
- ## Technical Specifications [optional]
172
-
173
- ### Model Architecture and Objective
174
-
175
- [More Information Needed]
176
-
177
- ### Compute Infrastructure
178
-
179
- [More Information Needed]
180
-
181
- #### Hardware
182
-
183
- [More Information Needed]
184
 
185
- #### Software
186
 
187
- [More Information Needed]
 
 
 
 
188
 
189
- ## Citation [optional]
 
 
 
190
 
191
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
192
 
193
- **BibTeX:**
194
 
195
- [More Information Needed]
 
 
 
 
 
 
196
 
197
- **APA:**
 
198
 
199
- [More Information Needed]
200
 
201
- ## Glossary [optional]
202
 
203
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
 
204
 
205
- [More Information Needed]
 
206
 
207
- ## More Information [optional]
208
 
209
- [More Information Needed]
210
 
211
- ## Model Card Authors [optional]
212
 
213
- [More Information Needed]
 
 
 
 
 
 
 
 
214
 
215
- ## Model Card Contact
216
 
217
- [More Information Needed]
218
- ### Framework versions
219
 
220
- - PEFT 0.17.0
 
1
  ---
2
  license: apache-2.0
3
+ base_model: Upstage/SOLAR-10.7B-Instruct-v1.0
4
+ library_name: peft
5
  pipeline_tag: text-generation
6
  tags:
7
  - chat
8
  - instruction-tuned
9
  - korean
 
 
 
 
 
 
 
 
 
10
  - lora
11
  - sft
12
  - transformers
13
  - trl
14
+ language:
15
+ - ko
16
+ - en
17
  ---
18
 
19
+ # solar-chatbot-final (LoRA Adapter)
 
 
20
 
21
+ **Summary:** A Korean conversational **LoRA adapter** fine-tuned on Upstage's `SOLAR-10.7B-Instruct-v1.0`
22
+ Designed for natural, responsive dialogue with capabilities in general Q&A, summarization, and text generation.
23
 
24
+ ---
25
 
26
  ## Model Details
27
 
 
 
 
 
 
 
28
  - **Developed by:** Jihee Cho
29
+ - **Model type:** Causal Language Model (chat/instruction-tuned), **LoRA adapter** for SOLAR-10.7B
30
+ - **Languages:** Korean (primary), English (basic support)
31
+ - **License:** Apache-2.0
32
+ - **Finetuned from:** [Upstage/SOLAR-10.7B-Instruct-v1.0](https://huggingface.co/Upstage/SOLAR-10.7B-Instruct-v1.0)
 
 
33
 
34
+ ### Model Sources
35
 
36
+ - **Base Model:** https://huggingface.co/Upstage/SOLAR-10.7B-Instruct-v1.0
37
+ - **Repository:** https://huggingface.co/Jay1121/solar-chatbot-final
38
 
39
+ ---
 
 
 
 
 
 
 
 
 
 
40
 
41
+ ## Quick Start
42
+
43
+ ### Requirements
44
+ ```bash
45
+ pip install transformers peft torch
46
+ ```
47
+
48
+ ### Usage Example
49
+ ```python
50
+ from transformers import AutoTokenizer, AutoModelForCausalLM
51
+ from peft import PeftModel
52
+
53
+ # Load base model and adapter
54
+ base_model_name = "Upstage/SOLAR-10.7B-Instruct-v1.0"
55
+ adapter_name = "Jay1121/solar-chatbot-final"
56
+
57
+ tokenizer = AutoTokenizer.from_pretrained(base_model_name, use_fast=True)
58
+ model = AutoModelForCausalLM.from_pretrained(
59
+ base_model_name,
60
+ torch_dtype="auto",
61
+ device_map="auto"
62
+ )
63
+ model = PeftModel.from_pretrained(model, adapter_name)
64
+
65
+ # Generate response
66
+ prompt = "안녕! 오늘 뭐 하지?" # "Hi! What should I do today?"
67
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
68
+ outputs = model.generate(
69
+ **inputs,
70
+ max_new_tokens=256,
71
+ do_sample=True,
72
+ temperature=0.8,
73
+ pad_token_id=tokenizer.eos_token_id
74
+ )
75
+ response = tokenizer.decode(outputs[0], skip_special_tokens=True)
76
+ print(response)
77
+ ```
78
 
79
+ ---
80
 
81
+ ## Intended Use
82
 
83
+ ### Primary Use Cases
84
+ - Casual Korean conversation and chatbot demos
85
+ - General question-answering systems
86
+ - Text summarization and draft writing
87
+ - Prototyping and educational purposes
88
 
89
  ### Out-of-Scope Use
90
+ - Professional advice in medical, legal, or financial domains
91
+ - Fact-checking or authoritative information provision
92
+ - Production-level commercial services
93
+ - Processing personal or sensitive data
94
 
95
+ ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
96
 
97
+ ## Limitations and Risks
98
 
99
+ ### Known Limitations
100
+ - **No factual accuracy guarantee:** The model cannot ensure the correctness of generated information
101
+ - **Hallucination:** May generate false or non-existent content
102
+ - **Bias:** Training data biases may be reflected in outputs
103
+ - **Consistency issues:** May struggle to maintain consistency in long conversations
104
 
105
+ ### Ethical Considerations
106
+ - Users should critically review all generated content
107
+ - Always consult reliable sources for important decisions
108
+ - Do not blindly trust model outputs; verification is essential
109
 
110
+ ---
111
 
112
+ ## Technical Details
113
 
114
+ ### Training Details
115
+ - **Fine-tuning Method:** LoRA (Low-Rank Adaptation)
116
+ - **Target Modules:** Query, Key, Value projection layers
117
+ - **Rank:** 16
118
+ - **Alpha:** 32
119
+ - **Training Data:** Korean conversation datasets
120
+ - **Training Framework:** Transformers + PEFT + TRL
121
 
122
+ ### Performance
123
+ This model is optimized for natural conversational experiences rather than benchmark performance, focusing on engaging and contextually appropriate Korean dialogue.
124
 
125
+ ---
126
 
127
+ ## Disclaimer and Usage Terms
128
 
129
+ ### No Impersonation Policy
130
+ This project is not affiliated with or endorsed by any person or organization. Do not use this model to impersonate specific individuals or create systems that could mislead others about official endorsements.
131
 
132
+ ### Branding and Attribution
133
+ When using this model, please provide appropriate attribution and avoid using it in ways that could be mistaken for official statements from the original author or related organizations.
134
 
135
+ ---
136
 
137
+ ## Citation
138
 
139
+ If you use this model in your research or projects, please cite:
140
 
141
+ ```bibtex
142
+ @misc{solar-chatbot-final,
143
+ author = {Jihee Cho},
144
+ title = {solar-chatbot-final: Korean Conversational LoRA Adapter for SOLAR-10.7B},
145
+ year = {2024},
146
+ publisher = {Hugging Face},
147
+ howpublished = {\url{https://huggingface.co/Jay1121/solar-chatbot-final}}
148
+ }
149
+ ```
150
 
151
+ ---
152
 
153
+ ## Contact
 
154
 
155
+ For questions or feedback, please use the Discussion tab on the Hugging Face model page.