davanstrien HF Staff commited on
Commit
6b40fbc
·
verified ·
1 Parent(s): 4669e6d

Add dataset card

Browse files
Files changed (1) hide show
  1. README.md +46 -17
README.md CHANGED
@@ -1,19 +1,48 @@
1
  ---
2
- dataset_info:
3
- features:
4
- - name: text
5
- dtype: string
6
- - name: label
7
- dtype: int64
8
- splits:
9
- - name: train
10
- num_bytes: 33231449
11
- num_examples: 24831
12
- download_size: 20827925
13
- dataset_size: 33231449
14
- configs:
15
- - config_name: default
16
- data_files:
17
- - split: train
18
- path: data/train-*
19
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ tags:
3
+ - deduplicated
4
+ - semhash
5
+ - semantic-deduplication
6
+ task_categories:
7
+ - text-generation
8
+ size_categories:
9
+ - 10K<n<100K
 
 
 
 
 
 
 
 
 
10
  ---
11
+
12
+ # Deduplicated imdb
13
+
14
+ This dataset is a deduplicated version of [imdb](https://huggingface.co/datasets/imdb)
15
+ using semantic deduplication with [SemHash](https://github.com/MinishLab/semhash).
16
+
17
+ ## Deduplication Details
18
+
19
+ - **Method**: deduplicate
20
+ - **Column**: `text`
21
+ - **Original size**: 25,000 samples
22
+ - **Deduplicated size**: 24,831 samples
23
+ - **Duplicate ratio**: 0.68%
24
+ - **Reduction**: 0.68%
25
+
26
+ - **Date processed**: 2025-06-27
27
+
28
+ ## How to use
29
+
30
+ ```python
31
+ from datasets import load_dataset
32
+
33
+ dataset = load_dataset("imdb-deduplicated")
34
+ ```
35
+
36
+ ## Processing script
37
+
38
+ This dataset was created using the following script:
39
+
40
+ ```bash
41
+ uv run dedupe-dataset.py imdb text <repo_id> --method deduplicate
42
+ ```
43
+
44
+ ## About semantic deduplication
45
+
46
+ Unlike exact deduplication, semantic deduplication identifies and removes samples that are
47
+ semantically similar even if they use different words. This helps create cleaner training
48
+ datasets and prevents data leakage between train/test splits.