ZhiyangQi97 commited on
Commit
74e198e
·
verified ·
1 Parent(s): 53dfd76

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -2
README.md CHANGED
@@ -16,7 +16,7 @@ datasets:
16
 
17
  # 🧠 KokoroChat-High (LoRA Adapter for Japanese Counseling Dialogue)
18
 
19
- This repository contains the **LoRA adapter weights** for KokoroChat-High, a version of the KokoroChat model fine-tuned on **high-feedback counseling dialogues** (client score ≥ 70 and ≤ 98) from the [KokoroChat dataset](https://huggingface.co/datasets/your-username/kokorochat).
20
 
21
  The base model is [tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3), and this adapter enhances it for generating **high-quality, empathetic Japanese counseling responses**.
22
 
@@ -51,7 +51,7 @@ from peft import PeftModel
51
 
52
  # === Base + Adapter Paths ===
53
  base_model_id = "tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3"
54
- adapter_id = "your-username/kokorochat-lora"
55
 
56
  # === Load Tokenizer ===
57
  tokenizer = AutoTokenizer.from_pretrained(base_model_id)
@@ -94,3 +94,18 @@ print(tokenizer.decode(output[0][input_ids.shape[-1]:], skip_special_tokens=True
94
  - 📁 **Dataset**: [KokoroChat Dataset on Hugging Face](https://huggingface.co/datasets/UEC-InabaLab/KokoroChat)
95
  - 🧠 **Base Model**: [tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3)
96
  - 📄 **Paper**: [KokoroChat: A Japanese Psychological Counseling Dialogue Dataset (ACL 2025)](https://drive.google.com/file/d/1T6XgvZii8rZ1kKLgOUGqm3BMvqQAvxEM/view?usp=sharing)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
 
17
  # 🧠 KokoroChat-High (LoRA Adapter for Japanese Counseling Dialogue)
18
 
19
+ This repository contains the **LoRA adapter weights** for KokoroChat-High, a version of the KokoroChat model fine-tuned on **high-feedback counseling dialogues** (client score ≥ 70 and ≤ 98) from the [KokoroChat dataset](https://huggingface.co/datasets/UEC-InabaLab/KokoroChat).
20
 
21
  The base model is [tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3), and this adapter enhances it for generating **high-quality, empathetic Japanese counseling responses**.
22
 
 
51
 
52
  # === Base + Adapter Paths ===
53
  base_model_id = "tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3"
54
+ adapter_id = "UEC-InabaLab/KokoroChat-High"
55
 
56
  # === Load Tokenizer ===
57
  tokenizer = AutoTokenizer.from_pretrained(base_model_id)
 
94
  - 📁 **Dataset**: [KokoroChat Dataset on Hugging Face](https://huggingface.co/datasets/UEC-InabaLab/KokoroChat)
95
  - 🧠 **Base Model**: [tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3)
96
  - 📄 **Paper**: [KokoroChat: A Japanese Psychological Counseling Dialogue Dataset (ACL 2025)](https://drive.google.com/file/d/1T6XgvZii8rZ1kKLgOUGqm3BMvqQAvxEM/view?usp=sharing)
97
+
98
+
99
+ ## 📄 Citation
100
+
101
+ If you use this dataset, please cite the following paper:
102
+
103
+ ```bibtex
104
+ @inproceedings{qi2025kokorochat,
105
+ title = {KokoroChat: A Japanese Psychological Counseling Dialogue Dataset Collected via Role-Playing by Trained Counselors},
106
+ author = {Zhiyang Qi and Takumasa Kaneko and Keiko Takamizo and Mariko Ukiyo and Michimasa Inaba},
107
+ booktitle = {Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics},
108
+ year = {2025},
109
+ url = {https://github.com/UEC-InabaLab/KokoroChat}
110
+ }
111
+ ```