ZhiyangQi97 commited on
Commit
debc210
Β·
verified Β·
1 Parent(s): 65330f0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -10
README.md CHANGED
@@ -24,9 +24,9 @@ Unlike response-generation models, this version is trained to **predict client f
24
 
25
  ## πŸ’‘ Overview
26
 
27
- - βœ… Fine-tuned on **2,601 dialogues** with client feedback scores between **70 and 98**
28
- - βœ… Data collected through **text-based role-play** by trained counselors
29
- - βœ… Covers a wide range of topics: depression, family, school, career, relationships, and more
30
  - βœ… Base Model: [`tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3`](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3)
31
 
32
  ---
@@ -95,17 +95,15 @@ Fine-tuning was performed using **QLoRA** with the following configuration:
95
 
96
  ### Dataset Split
97
 
98
- - **Training Data**: 2,601 dialogues with feedback scores between 70 and 98
99
- - **Train/Validation Split**: 90% train, 10% validation
100
 
101
  ### Hyperparameter Settings
102
 
103
- - **Optimizer**: `adamw_8bit`
104
  - **Warm-up Steps**: `100`
105
- - **Learning Rate**: `1e-3`
106
- - **Epochs**: `5`
107
- - **Batch Size**: `8`
108
- - **Validation Frequency**: every 400 steps
109
 
110
  ---
111
 
@@ -131,5 +129,6 @@ If you use this model or dataset, please cite the following paper:
131
  - [KokoroChat on GitHub (UEC-InabaLab)](https://github.com/UEC-InabaLab/KokoroChat)
132
  - πŸ€– **Model Variants**:
133
  - [Llama-3.1-KokoroChat-Low](https://huggingface.co/UEC-InabaLab/Llama-3.1-KokoroChat-Low): fine-tuned on **3,870 dialogues** with client feedback scores **< 70**
 
134
  - [Llama-3.1-KokoroChat-Full](https://huggingface.co/UEC-InabaLab/Llama-3.1-KokoroChat-Full): fine-tuned on **6,471 dialogues** with client feedback scores **≀ 98**
135
  - πŸ“„ **Paper**: [ACL 2025 Paper](https://aclanthology.org/2025.acl-long.608/)
 
24
 
25
  ## πŸ’‘ Overview
26
 
27
+ - βœ… Task: Predict the **overall counseling quality score** as rated by the client
28
+ - βœ… Dataset: 6,589 dialogues with feedback scores between 0 and 100
29
+ - βœ… Data source: Text-based role-play by trained counselors
30
  - βœ… Base Model: [`tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3`](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3)
31
 
32
  ---
 
95
 
96
  ### Dataset Split
97
 
98
+ - **Training/Validation/Test ratio**: 8:1:1
 
99
 
100
  ### Hyperparameter Settings
101
 
102
+ - **Optimizer**: `adamw_torch_fused`
103
  - **Warm-up Steps**: `100`
104
+ - **Learning Rate**: `2e-4`
105
+ - **Epochs**: `4`
106
+ - **Batch Size**: `4`
 
107
 
108
  ---
109
 
 
129
  - [KokoroChat on GitHub (UEC-InabaLab)](https://github.com/UEC-InabaLab/KokoroChat)
130
  - πŸ€– **Model Variants**:
131
  - [Llama-3.1-KokoroChat-Low](https://huggingface.co/UEC-InabaLab/Llama-3.1-KokoroChat-Low): fine-tuned on **3,870 dialogues** with client feedback scores **< 70**
132
+ - [Llama-3.1-KokoroChat-High](https://huggingface.co/UEC-InabaLab/Llama-3.1-KokoroChat-High): fine-tuned on **2,601 dialogues** with client feedback scores between **70 and 98**
133
  - [Llama-3.1-KokoroChat-Full](https://huggingface.co/UEC-InabaLab/Llama-3.1-KokoroChat-Full): fine-tuned on **6,471 dialogues** with client feedback scores **≀ 98**
134
  - πŸ“„ **Paper**: [ACL 2025 Paper](https://aclanthology.org/2025.acl-long.608/)