nielsr HF Staff commited on
Commit
1085800
·
verified ·
1 Parent(s): 0ec1422

Improve dataset card: add task category, paper/code links, and citation

Browse files

Hi, I'm Niels from the Hugging Face community science team!

This PR improves the dataset card for JF-ICR by:
- Adding the `text-classification` task category and `finance` tag to the YAML metadata.
- Linking the paper to the Hugging Face paper page (https://huggingface.co/papers/2602.01479).
- Adding the official GitHub repository link (https://github.com/The-FinAI/Ebisu).
- Including the evaluation command provided in the GitHub repository to help users run the benchmark.
- Updating the BibTeX citation with the actual authors and title from the paper.

Let me know if you have any questions!

Files changed (1) hide show
  1. README.md +30 -11
README.md CHANGED
@@ -1,11 +1,17 @@
1
  ---
2
- license: apache-2.0
3
  language:
4
  - ja
 
 
 
 
 
5
  ---
6
 
7
  # Dataset Card for JF-ICR
8
 
 
 
9
  ### Dataset Summary
10
 
11
  **JF-ICR (Japanese Financial Implicit Commitment Recognition)** is a benchmark dataset for evaluating implicit commitment and refusal recognition in Japanese investor-facing Q&A. The dataset consists of 94 single-turn question-answer pairs from 4 companies spanning 3 years, annotated into 5 implicit stance categories that distinguish agreement, hedging, and refusal under high-context discourse. This dataset addresses the challenge of understanding indirect expression and implicit commitment in Japanese financial communication, where refusals are frequently conveyed through pragmatic cues rather than explicit statements.
@@ -17,7 +23,22 @@ language:
17
 
18
  ### Languages
19
 
20
- * Japanese (jp)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
 
22
  ## Dataset Structure
23
 
@@ -159,15 +180,13 @@ JF-ICR contributes to research in Japanese financial NLP by enabling evaluation
159
  If you use this dataset, please cite:
160
 
161
  ```bibtex
162
- @misc{ebisu2025,
163
- title={EBISU: Benchmarking Large Language Models in Japanese Finance},
164
- author={[Authors]},
165
- year={2025},
166
- eprint={[arXiv number]},
167
  archivePrefix={arXiv},
168
  primaryClass={cs.CL},
169
- url={https://arxiv.org/abs/[number]},
170
  }
171
- ```
172
-
173
- **Note:** Please check the official paper for the complete citation once published.
 
1
  ---
 
2
  language:
3
  - ja
4
+ license: apache-2.0
5
+ task_categories:
6
+ - text-classification
7
+ tags:
8
+ - finance
9
  ---
10
 
11
  # Dataset Card for JF-ICR
12
 
13
+ [**Paper**](https://huggingface.co/papers/2602.01479) | [**Code**](https://github.com/The-FinAI/Ebisu)
14
+
15
  ### Dataset Summary
16
 
17
  **JF-ICR (Japanese Financial Implicit Commitment Recognition)** is a benchmark dataset for evaluating implicit commitment and refusal recognition in Japanese investor-facing Q&A. The dataset consists of 94 single-turn question-answer pairs from 4 companies spanning 3 years, annotated into 5 implicit stance categories that distinguish agreement, hedging, and refusal under high-context discourse. This dataset addresses the challenge of understanding indirect expression and implicit commitment in Japanese financial communication, where refusals are frequently conveyed through pragmatic cues rather than explicit statements.
 
23
 
24
  ### Languages
25
 
26
+ * Japanese (ja)
27
+
28
+ ## Usage
29
+
30
+ This dataset is part of the **Ebisu** benchmark. To evaluate a model using the official evaluation harness, you can use a command like the following:
31
+
32
+ ```bash
33
+ lm_eval --model vllm \
34
+ --model_args "pretrained=$MODEL,tensor_parallel_size=4,gpu_memory_utilization=0.95,max_model_len=1024" \
35
+ --tasks jp \
36
+ --batch_size auto \
37
+ --output_path ../results/jp \
38
+ --log_samples \
39
+ --apply_chat_template \
40
+ --include_path ../tasks/jp
41
+ ```
42
 
43
  ## Dataset Structure
44
 
 
180
  If you use this dataset, please cite:
181
 
182
  ```bibtex
183
+ @misc{peng2026ebisu,
184
+ title={Ebisu: Benchmarking Large Language Models in Japanese Finance},
185
+ author={Xueqing Peng and Ruoyu Xiang and Fan Zhang and Mingzi Song and Mingyang Jiang and Yan Wang and Lingfei Qian and Taiki Hara and Yuqing Guo and Jimin Huang and Junichi Tsujii and Sophia Ananiadou},
186
+ year={2026},
187
+ eprint={2602.01479},
188
  archivePrefix={arXiv},
189
  primaryClass={cs.CL},
190
+ url={https://huggingface.co/papers/2602.01479},
191
  }
192
+ ```