Update README.md
Browse files
README.md
CHANGED
|
@@ -16,7 +16,7 @@ datasets:
|
|
| 16 |
|
| 17 |
# 🧠 KokoroChat-High (LoRA Adapter for Japanese Counseling Dialogue)
|
| 18 |
|
| 19 |
-
This repository contains the **LoRA adapter weights** for KokoroChat-High, a version of the KokoroChat model fine-tuned on **high-feedback counseling dialogues** (client score ≥ 70 and ≤ 98) from the [KokoroChat dataset](https://huggingface.co/datasets/
|
| 20 |
|
| 21 |
The base model is [tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3), and this adapter enhances it for generating **high-quality, empathetic Japanese counseling responses**.
|
| 22 |
|
|
@@ -51,7 +51,7 @@ from peft import PeftModel
|
|
| 51 |
|
| 52 |
# === Base + Adapter Paths ===
|
| 53 |
base_model_id = "tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3"
|
| 54 |
-
adapter_id = "
|
| 55 |
|
| 56 |
# === Load Tokenizer ===
|
| 57 |
tokenizer = AutoTokenizer.from_pretrained(base_model_id)
|
|
@@ -94,3 +94,18 @@ print(tokenizer.decode(output[0][input_ids.shape[-1]:], skip_special_tokens=True
|
|
| 94 |
- 📁 **Dataset**: [KokoroChat Dataset on Hugging Face](https://huggingface.co/datasets/UEC-InabaLab/KokoroChat)
|
| 95 |
- 🧠 **Base Model**: [tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3)
|
| 96 |
- 📄 **Paper**: [KokoroChat: A Japanese Psychological Counseling Dialogue Dataset (ACL 2025)](https://drive.google.com/file/d/1T6XgvZii8rZ1kKLgOUGqm3BMvqQAvxEM/view?usp=sharing)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 16 |
|
| 17 |
# 🧠 KokoroChat-High (LoRA Adapter for Japanese Counseling Dialogue)
|
| 18 |
|
| 19 |
+
This repository contains the **LoRA adapter weights** for KokoroChat-High, a version of the KokoroChat model fine-tuned on **high-feedback counseling dialogues** (client score ≥ 70 and ≤ 98) from the [KokoroChat dataset](https://huggingface.co/datasets/UEC-InabaLab/KokoroChat).
|
| 20 |
|
| 21 |
The base model is [tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3), and this adapter enhances it for generating **high-quality, empathetic Japanese counseling responses**.
|
| 22 |
|
|
|
|
| 51 |
|
| 52 |
# === Base + Adapter Paths ===
|
| 53 |
base_model_id = "tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3"
|
| 54 |
+
adapter_id = "UEC-InabaLab/KokoroChat-High"
|
| 55 |
|
| 56 |
# === Load Tokenizer ===
|
| 57 |
tokenizer = AutoTokenizer.from_pretrained(base_model_id)
|
|
|
|
| 94 |
- 📁 **Dataset**: [KokoroChat Dataset on Hugging Face](https://huggingface.co/datasets/UEC-InabaLab/KokoroChat)
|
| 95 |
- 🧠 **Base Model**: [tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3)
|
| 96 |
- 📄 **Paper**: [KokoroChat: A Japanese Psychological Counseling Dialogue Dataset (ACL 2025)](https://drive.google.com/file/d/1T6XgvZii8rZ1kKLgOUGqm3BMvqQAvxEM/view?usp=sharing)
|
| 97 |
+
|
| 98 |
+
|
| 99 |
+
## 📄 Citation
|
| 100 |
+
|
| 101 |
+
If you use this dataset, please cite the following paper:
|
| 102 |
+
|
| 103 |
+
```bibtex
|
| 104 |
+
@inproceedings{qi2025kokorochat,
|
| 105 |
+
title = {KokoroChat: A Japanese Psychological Counseling Dialogue Dataset Collected via Role-Playing by Trained Counselors},
|
| 106 |
+
author = {Zhiyang Qi and Takumasa Kaneko and Keiko Takamizo and Mariko Ukiyo and Michimasa Inaba},
|
| 107 |
+
booktitle = {Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics},
|
| 108 |
+
year = {2025},
|
| 109 |
+
url = {https://github.com/UEC-InabaLab/KokoroChat}
|
| 110 |
+
}
|
| 111 |
+
```
|