sudeshna84 commited on
Commit
4d2a7ca
·
verified ·
1 Parent(s): 2d9bdfd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +319 -1
README.md CHANGED
@@ -4,4 +4,322 @@ language:
4
  - hi
5
  - te
6
  pretty_name: Bengali-Hindi-Telugu
7
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  - hi
5
  - te
6
  pretty_name: Bengali-Hindi-Telugu
7
+ ---
8
+ # BHT25: Bengali-Hindi-Telugu Parallel Corpus for Literary Machine Translation
9
+
10
+ [![License: CC BY 4.0](https://img.shields.io/badge/License-CC%20BY%204.0-lightgrey.svg)](https://creativecommons.org/licenses/by/4.0/)
11
+ [![Dataset on HF](https://huggingface.co/datasets/huggingface/badges/resolve/main/dataset-on-hf-sm.svg)](https://huggingface.co/datasets/sudeshna84/BHT25)
12
+
13
+ ## Overview
14
+
15
+ BHT25 is a high-quality trilingual parallel corpus comprising 25,000 sentence triplets across Bengali (BN), Hindi (HI), and Telugu (TE) languages. This dataset fills a critical gap in resources for cross-family Indian language machine translation, particularly for literary and culturally rich content.
16
+
17
+ **Key Features:**
18
+ - 🌐 **Three Indian Languages**: Bengali (Indo-Aryan), Hindi (Indo-Aryan), Telugu (Dravidian)
19
+ - 📚 **Literary Domain**: Sourced from works of renowned authors including Rabindranath Tagore and Sarat Chandra Chattopadhyay
20
+ - ✨ **Archaic Varieties**: Includes traditional Bengali Sadhu Bhasha
21
+ - ✅ **Human-Verified**: 75.8% semantic alignment accuracy with expert validation
22
+ - 🔍 **Unique Identifiers**: Each triplet has a unique ID for reproducibility
23
+ - 📖 **High Quality**: Composite fluency score of 4.08/5.0
24
+
25
+ ## Dataset Description
26
+
27
+ ### Languages
28
+
29
+ - **Bengali (bn)**: Eastern Indo-Aryan language, ~265M speakers
30
+ - **Hindi (hi)**: Central Indo-Aryan language, ~600M speakers
31
+ - **Telugu (te)**: South-Central Dravidian language, ~95M speakers
32
+
33
+ ### Dataset Statistics
34
+
35
+ | Metric | Bengali | Hindi | Telugu |
36
+ |--------|---------|-------|--------|
37
+ | **Total Sentences** | 25,000 | 25,000 | 25,000 |
38
+ | **Total Tokens** | ~XXX,XXX | ~XXX,XXX | ~XXX,XXX |
39
+ | **Avg. Tokens/Sentence** | ~XX | ~XX | ~XX |
40
+ | **Vocabulary Size** | ~XX,XXX | ~XX,XXX | ~XX,XXX |
41
+ | **Min Sentence Length** | X tokens | X tokens | X tokens |
42
+ | **Max Sentence Length** | XXX tokens | XXX tokens | XXX tokens |
43
+
44
+ ### Content Characteristics
45
+
46
+ The corpus covers diverse literary genres:
47
+ - Short stories and narratives
48
+ - Poetry and prose
49
+ - Folk literature
50
+ - Contemporary literary works
51
+ - Traditional Sadhu Bhasha (classical Bengali)
52
+
53
+ ### Quality Metrics
54
+
55
+ - **Alignment Accuracy**: 75.8% (human-validated)
56
+ - **Semantic Consistency**: 0.873 (Bengali-Telugu), 0.71 (Hindi-based pairs via CLWE)
57
+ - **Translation Fluency**: 4.08/5.0 (composite expert score)
58
+ - **Inter-Annotator Agreement**: κ = 0.89
59
+
60
+ ## Dataset Structure
61
+
62
+ ### Data Format
63
+
64
+ The dataset is provided in Parquet format with the following schema:
65
+
66
+ ```python
67
+ {
68
+ 'id': string, # Unique identifier (BHT25_00001 to BHT25_25000)
69
+ 'bn': string, # Bengali sentence
70
+ 'hi': string, # Hindi sentence
71
+ 'te': string # Telugu sentence
72
+ }
73
+ ```
74
+
75
+ ### Example
76
+
77
+ ```python
78
+ {
79
+ 'id': 'BHT25_00001',
80
+ 'bn': 'আমি তোমাকে ভালোবাসি।',
81
+ 'hi': 'मैं तुमसे प्यार करता हूँ।',
82
+ 'te': 'నేను నిన్ను ప్రేమిస్తున్నాను।'
83
+ }
84
+ ```
85
+
86
+ ### Data Splits
87
+
88
+ The dataset is provided as a **single unified corpus** without pre-defined train/dev/test splits. This design choice allows researchers maximum flexibility to partition the data according to their specific experimental requirements. The unique identifiers enable deterministic and reproducible splitting across different studies.
89
+
90
+ **Suggested Split Strategy** (for standardization):
91
+ - Train: 80% (20,000 triplets)
92
+ - Development: 10% (2,500 triplets)
93
+ - Test: 10% (2,500 triplets)
94
+
95
+ Users can implement splits using the unique IDs with their preferred strategy (random, stratified, etc.).
96
+
97
+ ## Usage
98
+
99
+ ### Loading the Dataset
100
+
101
+ ```python
102
+ from datasets import load_dataset
103
+
104
+ # Load the full dataset
105
+ dataset = load_dataset("sudeshna84/BHT25")
106
+
107
+ # Access data
108
+ print(f"Total samples: {len(dataset['train'])}")
109
+ print(f"First example: {dataset['train'][0]}")
110
+ ```
111
+
112
+ ### Creating Train/Dev/Test Splits
113
+
114
+ ```python
115
+ from datasets import load_dataset
116
+
117
+ # Load dataset
118
+ dataset = load_dataset("sudeshna84/BHT25", split="train")
119
+
120
+ # Create 80-10-10 split
121
+ train_test = dataset.train_test_split(test_size=0.2, seed=42)
122
+ train_dataset = train_test['train']
123
+ temp_dataset = train_test['test']
124
+
125
+ # Further split test into dev and test
126
+ dev_test = temp_dataset.train_test_split(test_size=0.5, seed=42)
127
+ dev_dataset = dev_test['train']
128
+ test_dataset = dev_test['test']
129
+
130
+ print(f"Train: {len(train_dataset)}, Dev: {len(dev_dataset)}, Test: {len(test_dataset)}")
131
+ ```
132
+
133
+ ### Accessing Specific Language Pairs
134
+
135
+ ```python
136
+ # Extract Bengali-Hindi pairs
137
+ bn_hi_pairs = [(item['bn'], item['hi']) for item in dataset['train']]
138
+
139
+ # Extract Bengali-Telugu pairs
140
+ bn_te_pairs = [(item['bn'], item['te']) for item in dataset['train']]
141
+
142
+ # Extract Hindi-Telugu pairs
143
+ hi_te_pairs = [(item['hi'], item['te']) for item in dataset['train']]
144
+ ```
145
+
146
+ ### Integration with Translation Models
147
+
148
+ ```python
149
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
150
+ from datasets import load_dataset
151
+
152
+ # Load dataset
153
+ dataset = load_dataset("sudeshna84/BHT25", split="train")
154
+
155
+ # Load model (example: IndicTrans2)
156
+ tokenizer = AutoTokenizer.from_pretrained("ai4bharat/indictrans2-bn-hi")
157
+ model = AutoModelForSeq2SeqLM.from_pretrained("ai4bharat/indictrans2-bn-hi")
158
+
159
+ # Prepare data for training
160
+ def preprocess_function(examples):
161
+ inputs = examples['bn']
162
+ targets = examples['hi']
163
+ model_inputs = tokenizer(inputs, max_length=128, truncation=True, padding='max_length')
164
+ labels = tokenizer(targets, max_length=128, truncation=True, padding='max_length')
165
+ model_inputs["labels"] = labels["input_ids"]
166
+ return model_inputs
167
+
168
+ tokenized_dataset = dataset.map(preprocess_function, batched=True)
169
+ ```
170
+
171
+ ### Data Analysis Example
172
+
173
+ ```python
174
+ import pandas as pd
175
+ from datasets import load_dataset
176
+
177
+ # Load dataset
178
+ dataset = load_dataset("sudeshna84/BHT25", split="train")
179
+
180
+ # Convert to pandas for analysis
181
+ df = pd.DataFrame(dataset)
182
+
183
+ # Analyze sentence lengths
184
+ df['bn_length'] = df['bn'].str.split().str.len()
185
+ df['hi_length'] = df['hi'].str.split().str.len()
186
+ df['te_length'] = df['te'].str.split().str.len()
187
+
188
+ print("Average sentence lengths:")
189
+ print(f"Bengali: {df['bn_length'].mean():.2f} tokens")
190
+ print(f"Hindi: {df['hi_length'].mean():.2f} tokens")
191
+ print(f"Telugu: {df['te_length'].mean():.2f} tokens")
192
+
193
+ # Length distribution
194
+ import matplotlib.pyplot as plt
195
+ df[['bn_length', 'hi_length', 'te_length']].hist(bins=50, figsize=(15, 5))
196
+ plt.suptitle('Sentence Length Distribution')
197
+ plt.show()
198
+ ```
199
+
200
+ ## Applications
201
+
202
+ This corpus can be used for various NLP tasks:
203
+
204
+ 1. **Machine Translation**
205
+ - Neural machine translation training and evaluation
206
+ - Cross-family translation (Indo-Aryan ↔ Dravidian)
207
+ - Low-resource language pair modeling
208
+
209
+ 2. **Cross-Lingual Analysis**
210
+ - Cross-lingual word embeddings evaluation
211
+ - Semantic similarity studies
212
+ - Syntactic divergence analysis
213
+
214
+ 3. **Multilingual NLP**
215
+ - Multilingual language model fine-tuning
216
+ - Zero-shot translation experiments
217
+ - Transfer learning for Indian languages
218
+
219
+ 4. **Linguistic Research**
220
+ - Comparative morphology and syntax
221
+ - Literary translation analysis
222
+ - Archaic language preservation (Sadhu Bhasha)
223
+
224
+ 5. **Quality Estimation**
225
+ - Automatic alignment algorithm evaluation
226
+ - Translation quality metrics benchmarking
227
+
228
+ ## Methodology
229
+
230
+ ### Data Collection
231
+
232
+ The corpus was constructed through:
233
+ 1. **Source Selection**: Literary texts from renowned Indian authors
234
+ 2. **Sentence Extraction**: Language-specific boundary detection
235
+ 3. **Alignment**: Hybrid Gale-Church algorithm enhanced with cross-lingual word embeddings (CLWE)
236
+ 4. **Validation**: Manual verification by native speakers
237
+
238
+ ### Quality Assurance
239
+
240
+ Each sentence triplet underwent:
241
+ - Automated preprocessing and alignment scoring
242
+ - Manual verification for semantic equivalence
243
+ - Expert review for fluency and naturalness
244
+ - Inter-annotator agreement measurement (κ = 0.89)
245
+
246
+ ### Alignment Algorithm
247
+
248
+ A hybrid approach combining:
249
+ - Gale-Church length-based alignment
250
+ - Cross-lingual word embedding (FastText) similarity
251
+ - Threshold-based filtering (score ≥ 0.7)
252
+ - Human expert correction
253
+
254
+ ## Citation
255
+
256
+ If you use this dataset in your research, please cite:
257
+
258
+ ```bibtex
259
+ @article{sani2024bht25,
260
+ title={A Bengali-Hindi-Telugu Parallel Corpus for Enhanced Literary Machine Translation},
261
+ author={Sani, Sudeshna and Gangashetty, Suryakanth V and Samudravijaya, K and Nandi, Anik and Priya, Aruna and Kumar, Vineeth and Dubey, Akhilesh Kumar},
262
+ journal={Data in Brief},
263
+ year={2024},
264
+ publisher={Elsevier}
265
+ }
266
+ ```
267
+
268
+ **Related Research:**
269
+ ```bibtex
270
+ @article{sani2024esa,
271
+ title={Emotion-Semantic-Aware Neural Machine Translation for Bengali-Hindi-Telugu},
272
+ author={Sani, Sudeshna and others},
273
+ journal={IEEE Access},
274
+ year={2024},
275
+ doi={10.1109/ACCESS.2024.XXXXXXX}
276
+ }
277
+ ```
278
+
279
+ ## License
280
+
281
+ This dataset is released under the [Creative Commons Attribution 4.0 International License (CC BY 4.0)](https://creativecommons.org/licenses/by/4.0/).
282
+
283
+ **You are free to:**
284
+ - Share — copy and redistribute the material in any medium or format
285
+ - Adapt — remix, transform, and build upon the material for any purpose, even commercially
286
+
287
+ **Under the following terms:**
288
+ - Attribution — You must give appropriate credit, provide a link to the license, and indicate if changes were made
289
+
290
+ ## Ethical Considerations
291
+
292
+ - All source texts are either in the public domain or used with explicit permission
293
+ - Author attributions are maintained in metadata
294
+ - No personal or sensitive information is included
295
+ - The dataset supports linguistic diversity and cultural preservation
296
+
297
+ ## Authors and Affiliations
298
+
299
+ **Sudeshna Sani**¹, Suryakanth V Gangashetty¹, Samudravijaya K¹*, Anik Nandi², Aruna Priya³, Vineeth Kumar³, Akhilesh Kumar Dubey¹
300
+
301
+ ¹ Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Guntur, Andhra Pradesh, India
302
+ ² School of Business, Woxsen University, Hyderabad, Telangana, India
303
+ ³ School of Technology, Woxsen University, Hyderabad, Telangana, India
304
+
305
+ *Corresponding author
306
+
307
+ ## Contact
308
+
309
+ For questions, issues, or collaborations, please:
310
+ - Open an issue in this repository
311
+ - Contact: [your.email@klef.edu.in]
312
+
313
+ ## Acknowledgments
314
+
315
+ We thank the native speakers who participated in the manual verification process and the literary estates for permission to use source materials.
316
+
317
+ ## Dataset Card
318
+
319
+ For more detailed information about the dataset construction, evaluation metrics, and validation procedures, please refer to our paper in *Data in Brief*.
320
+
321
+ ---
322
+
323
+ **Last Updated**: December 2024
324
+ **Version**: 1.0
325
+ **DOI**: [Will be added upon publication]