Improve dataset card: add paper link, GitHub link, and task category
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,4 +1,9 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
dataset_info:
|
| 3 |
features:
|
| 4 |
- name: query
|
|
@@ -19,17 +24,18 @@ configs:
|
|
| 19 |
data_files:
|
| 20 |
- split: test
|
| 21 |
path: data/test-*
|
| 22 |
-
license: apache-2.0
|
| 23 |
-
language:
|
| 24 |
-
- ja
|
| 25 |
---
|
| 26 |
|
| 27 |
# Dataset Card for JF-TE
|
| 28 |
|
|
|
|
|
|
|
| 29 |
### Dataset Summary
|
| 30 |
|
| 31 |
**JF-TE (Japanese Financial Term Extraction)** is a benchmark dataset for evaluating hierarchical extraction and ranking of nested financial terminology from Japanese professional disclosures. The dataset consists of 202 note-level instances extracted from 10 professional disclosures, containing 2,412 expert-curated term mentions covering 777 unique finance terms after normalization. The task addresses the challenge of boundary-sensitive grounding of finance terminology in mixed-script text, where nested compounds and script-variant loanwords (kanji, hiragana, and katakana) make term boundaries and semantic scope ambiguous. This dataset provides a linguistically grounded evaluation of financial term extraction that reflects the complexity of real Japanese financial disclosures.
|
| 32 |
|
|
|
|
|
|
|
| 33 |
### Supported Tasks and Leaderboards
|
| 34 |
|
| 35 |
* **Evaluation Metrics:**
|
|
@@ -40,7 +46,23 @@ language:
|
|
| 40 |
|
| 41 |
### Languages
|
| 42 |
|
| 43 |
-
* Japanese (
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
|
| 45 |
## Dataset Structure
|
| 46 |
|
|
@@ -101,11 +123,7 @@ Annotations were conducted following a rigorous protocol:
|
|
| 101 |
|
| 102 |
1. **Guideline Development**: Task-specific guidelines were developed tailored to hierarchical financial terminology, with explicit rules governing term boundary decisions and the distinction between financial terminology and general expressions.
|
| 103 |
|
| 104 |
-
2. **Iterative Refinement**: Annotation guidelines were iteratively refined through pre-annotation rounds, with particular attention to
|
| 105 |
-
- Term boundary decisions
|
| 106 |
-
- Distinction between financial terminology and general expressions
|
| 107 |
-
- Handling of nested and overlapping terms
|
| 108 |
-
- Mixed-script variant identification
|
| 109 |
|
| 110 |
3. **Span-Level Annotation**: Annotation was conducted at the span level, allowing multi-word expressions, compound nouns, and nested structures to be marked explicitly.
|
| 111 |
|
|
@@ -125,50 +143,28 @@ Annotations were conducted following a rigorous protocol:
|
|
| 125 |
|
| 126 |
All annotations were performed by the same native-level Japanese financial experts as in JF-ICR:
|
| 127 |
|
| 128 |
-
- **Annotator 1**: A doctoral student in Japan with a solid academic background in financial mathematics and several years of study and residence in Japan. Their research focuses on AI applications in finance
|
| 129 |
-
|
| 130 |
-
- **Annotator 2**: A researcher at a Japanese fintech company with nearly two decades of experience in the Japanese financial market. They hold a Ph.D. in Economics and possess deep expertise in Japanese corporate finance, financial reporting, and investor communications. As a native-level Japanese speaker, they are familiar with the linguistic, cultural, and institutional conventions of Japanese financial disclosures.
|
| 131 |
-
|
| 132 |
-
The annotation process was conducted using the Label Studio platform, ensuring consistency and reproducibility.
|
| 133 |
|
| 134 |
#### Inter-Annotator Agreement
|
| 135 |
|
| 136 |
-
Inter-annotator agreement results demonstrate strong consistency in identifying financial term spans
|
| 137 |
-
- **Span-level F1**: [Reported in paper]
|
| 138 |
-
- **Cohen's κ**: [Reported in paper]
|
| 139 |
-
- **Krippendorff's α**: [Reported in paper]
|
| 140 |
-
|
| 141 |
-
These results indicate that the annotation rules governing term boundaries and financial termhood are operational and reproducible, supporting the reliability of the resulting high-quality JF-TE dataset.
|
| 142 |
|
| 143 |
### Personal and Sensitive Information
|
| 144 |
|
| 145 |
-
The dataset contains no personal or sensitive information. All data is derived from publicly available professional disclosures (有価証券報告書)
|
| 146 |
|
| 147 |
## Considerations for Using the Data
|
| 148 |
|
| 149 |
### Social Impact of Dataset
|
| 150 |
|
| 151 |
-
JF-TE contributes to research in Japanese financial NLP by enabling evaluation of financial term extraction capabilities in linguistically complex settings. Accurate extraction of financial terminology is essential for downstream tasks such as financial analysis, regulatory compliance, and automated financial document processing.
|
| 152 |
|
| 153 |
### Discussion of Biases
|
| 154 |
|
| 155 |
-
* **Document Type**: The dataset focuses on Annual Securities Reports (有価証券報告書)
|
| 156 |
-
|
| 157 |
-
* **
|
| 158 |
-
|
| 159 |
-
* **Term Coverage**: The dataset emphasizes financial and accounting terminology, which may underrepresent other types of domain-specific language that appear in financial disclosures.
|
| 160 |
-
|
| 161 |
-
* **Script Representation**: While the dataset includes mixed-script terms, the distribution may reflect patterns specific to the selected disclosures.
|
| 162 |
-
|
| 163 |
-
### Other Known Limitations
|
| 164 |
-
|
| 165 |
-
* **Limited Size**: The dataset contains 202 note-level instances, which may limit coverage of all possible financial term patterns.
|
| 166 |
-
|
| 167 |
-
* **Note-Level Focus**: The dataset focuses on note-level excerpts rather than full documents, which may miss some contextual information relevant for term extraction.
|
| 168 |
-
|
| 169 |
-
* **Nested Term Complexity**: The hierarchical nature of nested terms may make evaluation challenging, as different models may prioritize different levels of the hierarchy.
|
| 170 |
-
|
| 171 |
-
* **Mixed-Script Handling**: The complexity of mixed-script Japanese text (kanji, hiragana, katakana) may pose challenges for models not specifically trained on Japanese text.
|
| 172 |
|
| 173 |
## Additional Information
|
| 174 |
|
|
@@ -185,15 +181,13 @@ JF-TE contributes to research in Japanese financial NLP by enabling evaluation o
|
|
| 185 |
If you use this dataset, please cite:
|
| 186 |
|
| 187 |
```bibtex
|
| 188 |
-
@misc{
|
| 189 |
-
title={
|
| 190 |
-
author={
|
| 191 |
-
year={
|
| 192 |
-
eprint={
|
| 193 |
archivePrefix={arXiv},
|
| 194 |
primaryClass={cs.CL},
|
| 195 |
-
url={https://
|
| 196 |
}
|
| 197 |
-
```
|
| 198 |
-
|
| 199 |
-
**Note:** Please check the official paper for the complete citation once published.
|
|
|
|
| 1 |
---
|
| 2 |
+
language:
|
| 3 |
+
- ja
|
| 4 |
+
license: apache-2.0
|
| 5 |
+
task_categories:
|
| 6 |
+
- text-retrieval
|
| 7 |
dataset_info:
|
| 8 |
features:
|
| 9 |
- name: query
|
|
|
|
| 24 |
data_files:
|
| 25 |
- split: test
|
| 26 |
path: data/test-*
|
|
|
|
|
|
|
|
|
|
| 27 |
---
|
| 28 |
|
| 29 |
# Dataset Card for JF-TE
|
| 30 |
|
| 31 |
+
[**Paper**](https://huggingface.co/papers/2602.01479) | [**GitHub**](https://github.com/The-FinAI/Ebisu)
|
| 32 |
+
|
| 33 |
### Dataset Summary
|
| 34 |
|
| 35 |
**JF-TE (Japanese Financial Term Extraction)** is a benchmark dataset for evaluating hierarchical extraction and ranking of nested financial terminology from Japanese professional disclosures. The dataset consists of 202 note-level instances extracted from 10 professional disclosures, containing 2,412 expert-curated term mentions covering 777 unique finance terms after normalization. The task addresses the challenge of boundary-sensitive grounding of finance terminology in mixed-script text, where nested compounds and script-variant loanwords (kanji, hiragana, and katakana) make term boundaries and semantic scope ambiguous. This dataset provides a linguistically grounded evaluation of financial term extraction that reflects the complexity of real Japanese financial disclosures.
|
| 36 |
|
| 37 |
+
It is part of the **Ebisu** benchmark, presented in the paper [Ebisu: Benchmarking Large Language Models in Japanese Finance](https://huggingface.co/papers/2602.01479).
|
| 38 |
+
|
| 39 |
### Supported Tasks and Leaderboards
|
| 40 |
|
| 41 |
* **Evaluation Metrics:**
|
|
|
|
| 46 |
|
| 47 |
### Languages
|
| 48 |
|
| 49 |
+
* Japanese (ja)
|
| 50 |
+
|
| 51 |
+
## Usage
|
| 52 |
+
|
| 53 |
+
You can evaluate models on this dataset using the [lm-evaluation-harness](https://github.com/ASCRX/lm-evaluation-harness) provided by the authors:
|
| 54 |
+
|
| 55 |
+
```bash
|
| 56 |
+
lm_eval --model vllm \
|
| 57 |
+
--model_args "pretrained=$MODEL,tensor_parallel_size=4,gpu_memory_utilization=0.95,max_model_len=1024" \
|
| 58 |
+
--tasks jp \
|
| 59 |
+
--batch_size auto \
|
| 60 |
+
--output_path ../results/jp \
|
| 61 |
+
--hf_hub_log_args "hub_results_org=TheFinAI,details_repo_name=lm-eval-results-jp,push_results_to_hub=True,push_samples_to_hub=True,public_repo=False" \
|
| 62 |
+
--log_samples \
|
| 63 |
+
--apply_chat_template \
|
| 64 |
+
--include_path ../tasks/jp
|
| 65 |
+
```
|
| 66 |
|
| 67 |
## Dataset Structure
|
| 68 |
|
|
|
|
| 123 |
|
| 124 |
1. **Guideline Development**: Task-specific guidelines were developed tailored to hierarchical financial terminology, with explicit rules governing term boundary decisions and the distinction between financial terminology and general expressions.
|
| 125 |
|
| 126 |
+
2. **Iterative Refinement**: Annotation guidelines were iteratively refined through pre-annotation rounds, with particular attention to term boundary decisions, handling of nested terms, and mixed-script variant identification.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 127 |
|
| 128 |
3. **Span-Level Annotation**: Annotation was conducted at the span level, allowing multi-word expressions, compound nouns, and nested structures to be marked explicitly.
|
| 129 |
|
|
|
|
| 143 |
|
| 144 |
All annotations were performed by the same native-level Japanese financial experts as in JF-ICR:
|
| 145 |
|
| 146 |
+
- **Annotator 1**: A doctoral student in Japan with a solid academic background in financial mathematics and several years of study and residence in Japan. Their research focuses on AI applications in finance.
|
| 147 |
+
- **Annotator 2**: A researcher at a Japanese fintech company with nearly two decades of experience in the Japanese financial market. They hold a Ph.D. in Economics and possess deep expertise in Japanese corporate finance.
|
|
|
|
|
|
|
|
|
|
| 148 |
|
| 149 |
#### Inter-Annotator Agreement
|
| 150 |
|
| 151 |
+
Inter-annotator agreement results demonstrate strong consistency in identifying financial term spans. Detailed scores for Span-level F1, Cohen's κ, and Krippendorff's α are reported in the official paper.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 152 |
|
| 153 |
### Personal and Sensitive Information
|
| 154 |
|
| 155 |
+
The dataset contains no personal or sensitive information. All data is derived from publicly available professional disclosures (有価証券報告書).
|
| 156 |
|
| 157 |
## Considerations for Using the Data
|
| 158 |
|
| 159 |
### Social Impact of Dataset
|
| 160 |
|
| 161 |
+
JF-TE contributes to research in Japanese financial NLP by enabling evaluation of financial term extraction capabilities in linguistically complex settings. Accurate extraction of financial terminology is essential for downstream tasks such as financial analysis, regulatory compliance, and automated financial document processing.
|
| 162 |
|
| 163 |
### Discussion of Biases
|
| 164 |
|
| 165 |
+
* **Document Type**: The dataset focuses on Annual Securities Reports (有価証券報告書).
|
| 166 |
+
* **Company Representation**: The dataset covers 10 professional disclosures, which may not represent the full diversity of Japanese corporate communication styles.
|
| 167 |
+
* **Term Coverage**: The dataset emphasizes financial and accounting terminology.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 168 |
|
| 169 |
## Additional Information
|
| 170 |
|
|
|
|
| 181 |
If you use this dataset, please cite:
|
| 182 |
|
| 183 |
```bibtex
|
| 184 |
+
@misc{peng2026ebisu,
|
| 185 |
+
title={Ebisu: Benchmarking Large Language Models in Japanese Finance},
|
| 186 |
+
author={Xueqing Peng and Ruoyu Xiang and Fan Zhang and Mingzi Song and Mingyang Jiang and Yan Wang and Lingfei Qian and Taiki Hara and Yuqing Guo and Jimin Huang and Junichi Tsujii and Sophia Ananiadou},
|
| 187 |
+
year={2026},
|
| 188 |
+
eprint={2602.01479},
|
| 189 |
archivePrefix={arXiv},
|
| 190 |
primaryClass={cs.CL},
|
| 191 |
+
url={https://huggingface.co/papers/2602.01479},
|
| 192 |
}
|
| 193 |
+
```
|
|
|
|
|
|