Update README.md
Browse files
README.md
CHANGED
|
@@ -22,8 +22,7 @@ datasets:
|
|
| 22 |
|
| 23 |
## π‘ Overview
|
| 24 |
|
| 25 |
-
- β
Fine-tuned on **
|
| 26 |
-
(from the full KokoroChat dataset of 6,589 dialogues; 118 high-score dialogues reserved for testing)
|
| 27 |
- β
Data collected through **text-based role-play** by trained counselors
|
| 28 |
- β
Covers a wide range of topics: depression, family, school, career, relationships, and more
|
| 29 |
- β
Base Model: [`tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3`](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3)
|
|
@@ -94,8 +93,7 @@ Fine-tuning was performed using **QLoRA** with the following configuration:
|
|
| 94 |
|
| 95 |
### Dataset Split
|
| 96 |
|
| 97 |
-
- **Training Data**:
|
| 98 |
-
*(from the full KokoroChat dataset of 6,589 dialogues; 118 dialogues with scores of 99 or 100 were reserved for testing)*
|
| 99 |
- **Train/Validation Split**: 90% train, 10% validation
|
| 100 |
|
| 101 |
### Hyperparameter Settings
|
|
@@ -131,5 +129,5 @@ If you use this model or dataset, please cite the following paper:
|
|
| 131 |
- [KokoroChat on GitHub (UEC-InabaLab)](https://github.com/UEC-InabaLab/KokoroChat)
|
| 132 |
- π€ **Model Variants**:
|
| 133 |
- [KokoroChat-Low](https://huggingface.co/UEC-InabaLab/KokoroChat-Low): fine-tuned on **3,870 dialogues** with client feedback scores **< 70**
|
| 134 |
-
- [KokoroChat-
|
| 135 |
- π **Paper**: [ACL 2025 Paper (PDF)](https://drive.google.com/file/d/1T6XgvZii8rZ1kKLgOUGqm3BMvqQAvxEM/view?usp=sharing)
|
|
|
|
| 22 |
|
| 23 |
## π‘ Overview
|
| 24 |
|
| 25 |
+
- β
Fine-tuned on **2,601 dialogues** with client feedback scores between **70 and 98**
|
|
|
|
| 26 |
- β
Data collected through **text-based role-play** by trained counselors
|
| 27 |
- β
Covers a wide range of topics: depression, family, school, career, relationships, and more
|
| 28 |
- β
Base Model: [`tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3`](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3)
|
|
|
|
| 93 |
|
| 94 |
### Dataset Split
|
| 95 |
|
| 96 |
+
- **Training Data**: 2,601 dialogues with feedback scores between 70 and 98
|
|
|
|
| 97 |
- **Train/Validation Split**: 90% train, 10% validation
|
| 98 |
|
| 99 |
### Hyperparameter Settings
|
|
|
|
| 129 |
- [KokoroChat on GitHub (UEC-InabaLab)](https://github.com/UEC-InabaLab/KokoroChat)
|
| 130 |
- π€ **Model Variants**:
|
| 131 |
- [KokoroChat-Low](https://huggingface.co/UEC-InabaLab/KokoroChat-Low): fine-tuned on **3,870 dialogues** with client feedback scores **< 70**
|
| 132 |
+
- [KokoroChat-Full](https://huggingface.co/UEC-InabaLab/KokoroChat-Full): fine-tuned on **6,471 dialogues** with client feedback scores **β€ 98**
|
| 133 |
- π **Paper**: [ACL 2025 Paper (PDF)](https://drive.google.com/file/d/1T6XgvZii8rZ1kKLgOUGqm3BMvqQAvxEM/view?usp=sharing)
|