Dataset Viewer
Auto-converted to Parquet Duplicate
query-id
stringlengths
36
36
corpus-id
stringlengths
24
24
score
int64
1
1
668f910a-08a9-410e-84c4-7a7c5fd3f9bb
67bc6d3381586e5aec8512c9
1
82c99906-f40f-43de-a191-bdb38dc2ccf2
67bc6d3381586e5aec8512c9
1
746393e9-77c7-40d2-b248-c3be80b63bb0
67bc6d3381586e5aec8512c9
1
c7e23c69-32ad-41a9-81fe-135dc652d3b3
67bc6d3381586e5aec8512c9
1
a53af6f3-c44b-469c-a39a-271b7aa19e84
67bc6d3381586e5aec8512c9
1
47c92141-5b11-426e-b486-cd5d036f086b
67bc6fc581586e5aec8512ce
1
95b58439-67ae-4af3-9eaa-e830dc079108
67bc6fc581586e5aec8512ce
1
1d1a10a0-9ffa-4504-9843-f49343af67ea
67bc6fc581586e5aec8512ce
1
424225ce-7131-4bb6-969f-04f3e08c0b3c
67bc6fc581586e5aec8512ce
1
ac12f5ce-9093-45d0-95fb-4b0c41b552f9
67bc6fc581586e5aec8512ce
1
05102bf2-7a28-4e00-a85f-bce34ba5efbe
67bc6fc581586e5aec8512ce
1
256971ab-0f03-4e46-91aa-bb784ab4b3b0
67bc708c81586e5aec8512d1
1
3cb079f1-247f-4e65-876d-5f3e34facd82
67bc708c81586e5aec8512d1
1
fa905b3b-d8a2-4879-baee-7d96964ac2a4
67bc708c81586e5aec8512d1
1
1dae2292-b01d-406a-9006-2fd2cf2853ce
67bc708c81586e5aec8512d1
1
66e2ec1a-e1d3-43af-a056-976e8b71875d
67bc708c81586e5aec8512d1
1
0b128bfc-b97a-4a16-b84f-a4c03f6f2c5f
67bc719481586e5aec8512d4
1
a85a63d2-3222-4071-8c6f-0047e464f881
67bc719481586e5aec8512d4
1
5acd8b94-bf96-4e1f-b49d-6e1e80ed22b2
67bc719481586e5aec8512d4
1
f57e114b-973e-4278-8408-9a61b333a013
67bc719481586e5aec8512d4
1
fca056cb-4db6-491f-9770-a3dfdaa4d5a8
67bc719481586e5aec8512d4
1
baba4298-08fc-4414-bac5-da73a64ea255
67bc760e31fb03c020176349
1
7e04f636-1cda-421a-8a65-f9d1f0748332
67bc760e31fb03c020176349
1
39fd811b-bcec-4474-a6f1-99ab75cc87ff
67bc760e31fb03c020176349
1
24f5d8f5-eda5-47bb-80c8-5d8c35b05e94
67bc760e31fb03c020176349
1
ac2a2897-55af-4e3f-8e19-4dd1388984c0
67bc760e31fb03c020176349
1
737829c0-c7ad-47ee-8560-c93d2d64d2e0
67bc760e31fb03c020176349
1
5e79ff10-0a86-4787-9edb-d2669c7cc813
67bc7819e3c0a2d5a0a10785
1
f56d893c-a965-4314-99c6-4a0a51518a2c
67bc7819e3c0a2d5a0a10785
1
31ac2a7b-49b6-4f44-a8ae-b8edc2b10ab5
67bc7819e3c0a2d5a0a10785
1
2840c607-a9f7-4cc7-bfa8-a67fd1b779d1
67bc7819e3c0a2d5a0a10785
1
953a814e-8837-40ff-8f7c-8408c42e176d
67bc7819e3c0a2d5a0a10785
1
07bc5ec0-1ae1-4f0f-af2d-471fde1faddc
67bc7819e3c0a2d5a0a10785
1
7ffd468e-867a-4a92-95bd-2341dffdcb5e
67bc784be3c0a2d5a0a10786
1
89533c0f-2bd5-40ec-8290-39c2529bd9cb
67bc784be3c0a2d5a0a10786
1
090c8a5b-ff10-4ce8-8582-fdecf819ded7
67bc784be3c0a2d5a0a10786
1
735c3395-3498-4623-99c4-47e5e146db39
67bc7882e3c0a2d5a0a10787
1
dfd794d1-14d9-4a78-b1d8-2277670dfbfc
67bc7882e3c0a2d5a0a10787
1
8433e7cb-bad3-47ad-9d41-b73e1ce0799a
67bc7882e3c0a2d5a0a10787
1
c8c42716-7f2f-4c02-a0a0-dc13131cf2ad
67bc7882e3c0a2d5a0a10787
1
b4be8a23-76e2-494e-a131-4c99ae15f47b
67bc7882e3c0a2d5a0a10787
1
889aaec9-8800-45a8-a0da-0d059f4f76f8
67bc78e5e3c0a2d5a0a10788
1
a90d4d75-8f68-43c7-b3c9-c950982c1f09
67bc78e5e3c0a2d5a0a10788
1
77eabf7c-ced6-4751-af96-6c4fceddde2d
67bc78e5e3c0a2d5a0a10788
1
cc332cb3-93a6-4b36-98bd-e848b2a69895
67bc78e5e3c0a2d5a0a10788
1
63f8076a-e6cd-476e-8ee5-df0c83992a78
67bc78e5e3c0a2d5a0a10788
1
c187c342-1588-4c93-b3f0-4b6becaddd9d
67bc78e5e3c0a2d5a0a10788
1
267c7676-cb9b-49ad-880f-1020cbe67265
67bc7913e3c0a2d5a0a10789
1
a494ecb6-dd2a-4883-8052-85ed60fc41fd
67bc7913e3c0a2d5a0a10789
1
77a7a7fd-a887-4aa6-ad39-4a8996b9f5ad
67bc7913e3c0a2d5a0a10789
1
67ccc143-d31b-4ac9-8088-c30cbd036495
67bc7913e3c0a2d5a0a10789
1
8138a08a-c7ef-4d6b-b511-dc2054b6b4f4
67bc7913e3c0a2d5a0a10789
1
99029536-5014-4471-800f-de7693720346
67bc7913e3c0a2d5a0a10789
1
be9b6c1e-c104-4df5-bfb5-0431e5fe8fcb
67bc793de3c0a2d5a0a1078a
1
c1590f6b-7a6a-4afd-93fc-66b950f1a216
67bc793de3c0a2d5a0a1078a
1
34cadcd0-0752-43dc-98ab-17ad70cec1b1
67bc793de3c0a2d5a0a1078a
1
3f0af472-218a-4c0b-b5f1-c16bc641e6bd
67bc793de3c0a2d5a0a1078a
1
3db868b4-7358-4583-acfc-650837c77a38
67bc793de3c0a2d5a0a1078a
1
2fc64499-5f8b-4eee-a803-b0b4ebae30b6
67bc7a43e3c0a2d5a0a1078e
1
8b9e09b4-a0b5-49e1-87b3-de9c0a3b21f3
67bc7a43e3c0a2d5a0a1078e
1
61c91af4-e324-4a8a-b391-44166de85e37
67bc7a43e3c0a2d5a0a1078e
1
0c0f11ef-a299-409b-9726-5792ec93fa94
67bc7a43e3c0a2d5a0a1078e
1
8f152617-15bd-4b6f-b358-7e37a9f63747
67bc7a43e3c0a2d5a0a1078e
1
36a61b6f-1987-4623-a5fe-1e0b4e99312d
67bc7a43e3c0a2d5a0a1078e
1
b6ae15ca-68ac-4b93-92b1-e8b99402ae4c
67bc7a43e3c0a2d5a0a1078e
1
f136c9fe-d307-40b7-8844-ada0e0e0be38
67bc7a77e3c0a2d5a0a1078f
1
beffa870-d8de-4c22-9291-505d9d7996bb
67bc7a77e3c0a2d5a0a1078f
1
d749e979-0a29-4bfd-a3df-e00dd511e127
67bc7a77e3c0a2d5a0a1078f
1
39871530-3814-47ab-8be9-b42bae85846e
67bc7a77e3c0a2d5a0a1078f
1
d5b90937-308c-4643-b050-066d2469f57f
67bc7a77e3c0a2d5a0a1078f
1
2ddeecb0-0bd2-418c-ae22-7d9f5aa0fd6d
67bc7a77e3c0a2d5a0a1078f
1
63f01814-5a9c-4663-9908-b4514d8b09ac
67bc7a77e3c0a2d5a0a1078f
1
588ba229-e122-4c23-9eec-8f5f1cfe4496
67bc7b41e3c0a2d5a0a10791
1
ad5daa0b-80a8-47e0-81c9-59591bad84bc
67bc7b41e3c0a2d5a0a10791
1
dfe46970-8c78-4a3f-83ae-d28ab2861b33
67bc7b41e3c0a2d5a0a10791
1
be1aa321-c3f6-4807-b631-91fcf0779463
67bc7b41e3c0a2d5a0a10791
1
1311480d-2119-450b-8191-fd356ff8bf3c
67bc7b41e3c0a2d5a0a10791
1
daad90e2-a6f9-4d40-8a28-b4fb3ea6b7d7
67bc7b41e3c0a2d5a0a10791
1
b0a824f4-5587-4741-8e6d-974bb4b271fa
67bc7b41e3c0a2d5a0a10791
1
e9212c50-a818-4492-a556-588bab4d31d3
67bc7bc4e3c0a2d5a0a10792
1
061860a8-8030-47ff-a88d-2a4278a9c53b
67bc7bc4e3c0a2d5a0a10792
1
37e92f5b-46e4-438c-964d-d5ff4f23e339
67bc7bc4e3c0a2d5a0a10792
1
8a3ba477-a530-4913-bde2-f1f2c8269f8d
67bc7bc4e3c0a2d5a0a10792
1
d2a85f36-761c-4d18-a568-2f40fc54ce6b
67bc7bc4e3c0a2d5a0a10792
1
e5890897-5109-4d09-921b-e92ded6283e2
67bc7c1de3c0a2d5a0a10793
1
a9a4bc1e-6882-4bd8-adbd-bc198a7d2ed8
67bc7c51e3c0a2d5a0a10794
1
308d2d46-4039-4d90-a71f-b8f37f4228f8
67bc7c51e3c0a2d5a0a10794
1
2a1e4ae5-5efb-43fb-adc2-93fdfcbd6a6f
67bc7c51e3c0a2d5a0a10794
1
adfc32ca-877d-4913-b07c-4058eb391a76
67bc7c51e3c0a2d5a0a10794
1
daebf6b4-2988-440e-a49d-0bf187d5152a
67bc7c83e3c0a2d5a0a10795
1
95990a5f-c2b5-44fd-b85e-71630f8006f5
67bc7c83e3c0a2d5a0a10795
1
6df7c153-a995-4fb8-a162-fb2da0983f91
67bc7c83e3c0a2d5a0a10795
1
5de525cd-ae59-4034-b9ed-c3503c9e5aa5
67bc7c83e3c0a2d5a0a10795
1
4b62590e-ff9a-418d-903a-6b723906449d
67bc7c83e3c0a2d5a0a10795
1
4eb62fa9-0180-4564-a10f-696a6ea99dff
67bc7c83e3c0a2d5a0a10795
1
e547482e-e225-4d32-8f65-5b7c2f257907
67bc7cbae3c0a2d5a0a10796
1
b411f7ac-f474-4c62-ac8f-6897fb1b8028
67bc7cbae3c0a2d5a0a10796
1
da607a78-63d8-44d7-9d29-0b80f3ffea83
67bc7cbae3c0a2d5a0a10796
1
affc32c3-9f04-428a-8ecf-1b0572ade10b
67bc7cbae3c0a2d5a0a10796
1
b0ac9d57-7e20-4821-9e32-18708bf92236
67bc7cbae3c0a2d5a0a10796
1
End of preview. Expand in Data Studio

Turkish Legal Özelge Corpus Dataset

📊 Dataset Summary

Turkish Legal Özelge Corpus is a comprehensive Information Retrieval dataset consisting of özelge (tax ruling) decisions published by the Turkish Revenue Administration (Gelir İdaresi Başkanlığı - GİB).

Key Features

  • Format: BEIR (Benchmarking IR) format with corpus-queries-qrels structure
  • Language: Turkish 🇹🇷
  • Domain: Tax Law, Administrative Law, Turkish Law
  • Source: GİB Özelge Decisions
  • Use Cases: Information retrieval, question-answering systems, RAG applications

Dataset Structure

The dataset follows the BEIR format and consists of three main components:

2. Queries (Query Collection)

Legal information pieces extracted from 7 different perspectives for each document.

7 Query Types:

  1. Subject: Main topic of the özelge
  2. Article Text: Text of relevant law articles
  3. Communique Text: Content of relevant communiques and circulars
  4. Regulation Text: Regulation and legislation texts
  5. Justification Text: Legal justifications
  6. Decision Text: Administrative opinions and final decisions
  7. Condition Text: Application conditions and requirements

Tokenizer Benchmark & Data Filtering Summary

This process is not a training error and does not involve any training failure. It is a data analysis and preprocessing step performed before model training.

We benchmarked seven tokenizers (MPNet, Qwen2, Gemma, XLM-R, BERT, Pretrained, T5) on all datasets to measure token lengths and identify extreme long-sequence outliers. Among these, MPNetTokenizerFast generated the highest total token count, making it the most sensitive tokenizer for detecting unusually long samples.

Using MPNet as the reference tokenizer, we removed samples that exceeded the dataset-specific average by ~7000 tokens. This filtering was applied independently to each dataset to ensure balanced sequence distributions and cleaner input data.

The number of removed and remaining samples is summarized in the table below.

Tokenizer vocab_size total_tokens avg_tokens min_tokens max_tokens median_tokens
MPNetTokenizerFast 30,527 276,476,811 2,281 263 12,383 1,998
Qwen2TokenizerFast 151,669 219,326,828 1,810 190 9,201 1,594
GemmaTokenizerFast 262,144 183,710,411 1,516 158 7,578 1,341
XLMRobertaTokenizerFast 250,002 151,008,441 1,246 132 6,397 1,099
BertTokenizerFast 32,000 127,503,718 1,052 103 5,386 931
PretrainedTokenizerFast 32,000 122,387,578 1,010 102 5,227 893
T5TokenizerFast 32,128 121,315,289 1,001 100 5,238 885

Tokenizer / Total Token

Corr of Vocab Size – Total Token
Dataset max_tokens avg_tokens deleted_samples total_samples
newmindai/regulation-retrieval 276,476,811 2281.19945 1,300 121,300
newmindai/caselaw-retrieval 1,386 2,281 0 1,386
newmindai/court-of-cassation-caselaw 30,527 186.4827586 11 272

3. Default (Relevance Matrix)

Relationship table showing which query belongs to which document.

Field Description
query-id Query identifier
corpus-id Related document identifier
score Relevance score (all 1)

Dataset Statistics

Total Statistics:
├─  Corpus Records: 23,701 documents
├─  Query Records: 121,198 queries
└─  Relevance Records: 121,198 relations

Per Document:
├─ 1 corpus entry (full ruling text)
├─ 2–7 queries (legal perspectives)
└─ Average ~5.1 queries per document

Field Coverage (Queries per Document)

On average, each özelge is represented by around 5.1 distinct queries, corresponding to different legal fields. The distribution of populated query types per document is as follows:

  • 2 query types: ~0.1% of documents (e.g., Subject + Article Text)
  • 3 query types: ~12.3% of documents (e.g., Subject + Article Text + Decision Text)
  • 4 query types: ~26.2% of documents (e.g., Subject + Article Text + Communique Text + Decision Text)
  • 5 query types: ~23.9% of documents (e.g., Subject + Article Text + Communique Text + Regulation Text + Decision Text)
  • 6 query types: ~12.6% of documents (e.g., Subject + Article Text + Communique Text + Regulation Text + Justification Text + Decision Text)
  • 7 query types: ~24.9% of documents (All fields: Subject + Article Text + Communique Text + Regulation Text + Justification Text + Decision Text + Condition Text)

Query Types Available:

  1. Subject: Main topic/issue of the ruling
  2. Article Text: Relevant law article content
  3. Communique Text: Official communique/circular content
  4. Regulation Text: Regulation and legislation texts
  5. Justification Text: Legal reasoning and justifications
  6. Decision Text: Administrative opinion and final decision
  7. Condition Text: Application conditions and requirements

In other words, roughly 61% of the corpus has 5 or more query types populated, making them rich multi-perspective legal cases rather than shallow single-label examples.

Queries per document distribution

Text Length Distribution

For corpus texts (original full özelge rulings with non-empty ozelge_content, currently 100 documents):

  • Mean length: ~1,736 words
  • Median (p50): ~1,658 words
  • 90th percentile (p90): ~2,393 words

These are long, dense legal rulings, comparable to typical tax/administrative decisions with full reasoning and references.

For query texts (legal snippets extracted from seven perspectives across all 23k+ records):

  • Mean length: ~41.6 words
  • Median (p50): ~24 words
  • 90th percentile (p90): ~97 words

This makes queries similar to short legal questions, issue statements, justifications or excerpts from statutes/communiques, while the associated corpus entries provide the full ruling context for the subset of records where the full original özelge text is available.

Corpus vs. query text length histograms

Use Cases

1. Information Retrieval Systems

  • Training for semantic search models
  • Dense retrieval systems (DPR, ANCE, ColBERT)
  • Sparse retrieval systems (BM25, TF-IDF) benchmark

2. RAG (Retrieval-Augmented Generation) Applications

  • Legal chatbots
  • Tax consultation assistants
  • Automatic özelge analysis systems

3. Question-Answering Systems

  • Legal QA models
  • Extractive and abstractive QA
  • Multi-hop reasoning

4. 📊 Model Evaluation

  • Benchmarking Turkish IR models
  • Retrieval performance analysis
  • Domain adaptation studies

Data Collection and Processing

Data Source

The data is sourced from official özelge decisions of the Turkish Revenue Administration. Each özelge:

  • Responds to specific questions asked by taxpayers
  • References relevant legislation, communiques, and regulations
  • Contains the Administration's opinion for concrete applications

Ethics and Legal Notices

License

This dataset is published under CC-BY 4.0 license. Please cite the source when using.


Citation

@article{mecellem2026,
  title={Mecellem Models: Turkish Models Trained from Scratch and Continually Pre-trained for the Legal Domain},
  author={Uğur, Özgür and Göksu, Mahmut and Çimen, Mahmut and Yılmaz, Musa and Şavirdi, Esra and Demir, Alp Talha and Güllüce, Rumeysa and Çetin, İclal and Sağbaş, Ömer Can},
  journal={arXiv preprint arXiv:2601.16018},
  year={2026},
  month={January},
  url={https://arxiv.org/abs/2601.16018},
  doi={10.48550/arXiv.2601.16018},
  eprint={2601.16018},
  archivePrefix={arXiv},
  primaryClass={cs.CL}
}

License

This dataset is released under the Apache 2.0 License.

Contact

For questions: info@newmind.ai

Downloads last month
266

Space using newmindai/regulation-retrieval 1

Collection including newmindai/regulation-retrieval

Paper for newmindai/regulation-retrieval