LimYeri commited on
Commit
2dd373b
ยท
verified ยท
1 Parent(s): 7bc9622

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -34
README.md CHANGED
@@ -3,12 +3,7 @@ library_name: transformers
3
  tags: []
4
  ---
5
 
6
- # Model Card for Model ID
7
-
8
- <!-- Provide a quick summary of what the model is/does. -->
9
-
10
-
11
-
12
  ## Model Details
13
 
14
  ### Model Description
@@ -16,10 +11,10 @@ tags: []
16
  ์ด ๋ชจ๋ธ์€ ํ•œ๊ตญ์–ด ์ผ์ƒ ํ…์ŠคํŠธ(ํŠนํžˆ ์ผ๊ธฐ/์‹ฌ๋ฆฌ ๊ธฐ๋ก)์— ๋‚˜ํƒ€๋‚˜๋Š” ๊ฐ์ •์„ ์ž๋™์œผ๋กœ ๋ถ„๋ฅ˜ํ•˜๊ธฐ ์œ„ํ•ด ๊ตฌ์ถ•๋œ ๊ฐ์ • ๋ถ„๋ฅ˜ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.
17
  ๊ธฐ๋ฐ˜ ๋ชจ๋ธ์€ monologg/koelectra-base-v3-discriminator์ด๋ฉฐ, ๊ฐ์ • ๋ถ„์„ Task์— ์ตœ์ ํ™”๋˜๋„๋ก 8๊ฐœ ๊ฐ์ • ๋ผ๋ฒจ(๊ธฐ์จ, ์„ค๋ ˜, ํ‰๋ฒ”ํ•จ, ๋†€๋ผ์›€, ๋ถˆ์พŒํ•จ, ๋‘๋ ค์›€, ์Šฌํ””, ๋ถ„๋…ธ) ๊ตฌ์กฐ๋กœ ํŒŒ์ธํŠœ๋‹๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
18
 
19
- - **Model type:** [Text Classification (Emotion Recognition)]
20
- - **Language (NLP):** [Korean (ํ•œ๊ตญ์–ด, ko)]
21
- - **License:** [MIT]
22
- - **Finetuned from model:** [monologg/koelectra-base-v3-discriminator]
23
 
24
  ### Emotion Classes
25
 
@@ -69,26 +64,26 @@ print(predict("์˜ค๋Š˜ ํ•˜๋ฃจ๊ฐ€ ์œ ๋‚œํžˆ ํž˜๋“ค๊ณ  ์ง€์ณค๋‹ค."))
69
  1. [LimYeri/kor-diary-emotion_v2]("https://huggingface.co/datasets/LimYeri/kor-diary-emotion_v2")
70
  2. [qowlsdud/CounselGPT]("https://huggingface.co/datasets/qowlsdud/CounselGPT")
71
 
72
- - **Total(8:2๋กœ ๋ถ„ํ• ):** [50,000ํ–‰]
73
- - **Train:** [40,000ํ–‰]
74
- - **Validation:** [10,000ํ–‰]
75
 
76
  ### Training Procedure
77
 
78
  - **Base Model**: [monologg/koelectra-base-v3-discriminator]("https://huggingface.co/monologg/koelectra-base-v3-discriminator")
79
- - **Objective**: [Single-label classification]
80
- - **Precision**: [fp16 mixed precision]
81
- - **Max Length**: [512]
82
 
83
  #### Training Hyperparameters
84
 
85
- - **num_train_epochs**: [3]
86
- - **learning_rate**: [3e-5]
87
- - **weight_decay**: [0.02]
88
- - **warmup_ratio**: [0.15]
89
- - **per_device_train_batch_size**: [32]
90
- - **per_device_eval_batch_size**: [64]
91
- - **max_grad_norm**: [1.0]
92
 
93
  ---
94
 
@@ -101,23 +96,35 @@ print(predict("์˜ค๋Š˜ ํ•˜๋ฃจ๊ฐ€ ์œ ๋‚œํžˆ ํž˜๋“ค๊ณ  ์ง€์ณค๋‹ค."))
101
  | **Eval Loss** | 0.16 |
102
 
103
  ---
104
-
105
  ## Model Architecture
106
 
107
  ### 1) ELECTRA Encoder (Base-size)
108
 
109
- - **Hidden size:** [768]
110
- - **Layers:** [12 Transformer blocks]
111
- - **Attention heads:** [12]
112
- - **MLP intermediate size:** [3072]
113
- - **Activation:** [GELU]
114
- - **Dropout:** [0.1]
115
 
116
  ### 2) Classification Head
117
 
118
  ๊ฐ์ • 8๊ฐœ ํด๋ž˜์Šค๋ฅผ ์˜ˆ์ธกํ•˜๊ธฐ ์œ„ํ•œ ์ถ”๊ฐ€ ๋ถ„๋ฅ˜ ํ—ค๋“œ:
119
 
120
- - **Dense Layer**: [768 โ†’ 768]
121
- - **Activation**: [GELU]
122
- - **Dropout**: [0.1]
123
- - **Output Projection**: [768 โ†’ 8]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  tags: []
4
  ---
5
 
6
+ # HowRU-KoELECTRA-Emotion-Classifier
 
 
 
 
 
7
  ## Model Details
8
 
9
  ### Model Description
 
11
  ์ด ๋ชจ๋ธ์€ ํ•œ๊ตญ์–ด ์ผ์ƒ ํ…์ŠคํŠธ(ํŠนํžˆ ์ผ๊ธฐ/์‹ฌ๋ฆฌ ๊ธฐ๋ก)์— ๋‚˜ํƒ€๋‚˜๋Š” ๊ฐ์ •์„ ์ž๋™์œผ๋กœ ๋ถ„๋ฅ˜ํ•˜๊ธฐ ์œ„ํ•ด ๊ตฌ์ถ•๋œ ๊ฐ์ • ๋ถ„๋ฅ˜ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.
12
  ๊ธฐ๋ฐ˜ ๋ชจ๋ธ์€ monologg/koelectra-base-v3-discriminator์ด๋ฉฐ, ๊ฐ์ • ๋ถ„์„ Task์— ์ตœ์ ํ™”๋˜๋„๋ก 8๊ฐœ ๊ฐ์ • ๋ผ๋ฒจ(๊ธฐ์จ, ์„ค๋ ˜, ํ‰๋ฒ”ํ•จ, ๋†€๋ผ์›€, ๋ถˆ์พŒํ•จ, ๋‘๋ ค์›€, ์Šฌํ””, ๋ถ„๋…ธ) ๊ตฌ์กฐ๋กœ ํŒŒ์ธํŠœ๋‹๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
13
 
14
+ - **Model type:** Text Classification (Emotion Recognition)
15
+ - **Language (NLP):** Korean (ํ•œ๊ตญ์–ด, ko)
16
+ - **License:** MIT
17
+ - **Finetuned from model:** monologg/koelectra-base-v3-discriminator
18
 
19
  ### Emotion Classes
20
 
 
64
  1. [LimYeri/kor-diary-emotion_v2]("https://huggingface.co/datasets/LimYeri/kor-diary-emotion_v2")
65
  2. [qowlsdud/CounselGPT]("https://huggingface.co/datasets/qowlsdud/CounselGPT")
66
 
67
+ - **Total(8:2๋กœ ๋ถ„ํ• ):** 50,000ํ–‰
68
+ - **Train:** 40,000ํ–‰
69
+ - **Validation:** 10,000ํ–‰
70
 
71
  ### Training Procedure
72
 
73
  - **Base Model**: [monologg/koelectra-base-v3-discriminator]("https://huggingface.co/monologg/koelectra-base-v3-discriminator")
74
+ - **Objective**: Single-label classification
75
+ - **Precision**: fp16 mixed precision
76
+ - **Max Length**: 512
77
 
78
  #### Training Hyperparameters
79
 
80
+ - **num_train_epochs**: 3
81
+ - **learning_rate**: 3e-5
82
+ - **weight_decay**: 0.02
83
+ - **warmup_ratio**: 0.15
84
+ - **per_device_train_batch_size**: 32
85
+ - **per_device_eval_batch_size**: 64
86
+ - **max_grad_norm**: 1.0
87
 
88
  ---
89
 
 
96
  | **Eval Loss** | 0.16 |
97
 
98
  ---
 
99
  ## Model Architecture
100
 
101
  ### 1) ELECTRA Encoder (Base-size)
102
 
103
+ - **Hidden size:** 768
104
+ - **Layers:** 12 Transformer blocks
105
+ - **Attention heads:** 12
106
+ - **MLP intermediate size:** 3072
107
+ - **Activation:** GELU
108
+ - **Dropout:** 0.1
109
 
110
  ### 2) Classification Head
111
 
112
  ๊ฐ์ • 8๊ฐœ ํด๋ž˜์Šค๋ฅผ ์˜ˆ์ธกํ•˜๊ธฐ ์œ„ํ•œ ์ถ”๊ฐ€ ๋ถ„๋ฅ˜ ํ—ค๋“œ:
113
 
114
+ - **Dense Layer**: 768 โ†’ 768
115
+ - **Activation**: GELU
116
+ - **Dropout**: 0.1
117
+ - **Output Projection**: 768 โ†’ 8
118
+
119
+ ---
120
+
121
+ ## Citation
122
+ ```bibtex
123
+ @misc{HowRUEmotion2025,
124
+ title={HowRU KoELECTRA Emotion Classifier},
125
+ author={Lim, Yeri},
126
+ year={2025},
127
+ publisher={Hugging Face},
128
+ howpublished={\url{https://huggingface.co/LimYeri/HowRU-KoELECTRA-Emotion-Classifier}}
129
+ }
130
+ ```