Datasets:

Modalities:
Tabular
Text
Formats:
parquet
Languages:
Hungarian
License:
RabidUmarell commited on
Commit
d9ba748
·
verified ·
1 Parent(s): 1ecfda0

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +24 -25
README.md CHANGED
@@ -1,27 +1,26 @@
1
  ---
2
- dataset_info:
3
- features:
4
- - name: type
5
- dtype: large_string
6
- - name: subreddit
7
- dtype: large_string
8
- - name: text
9
- dtype: string
10
- - name: token_count
11
- dtype: int64
12
- - name: source_id
13
- dtype: int64
14
- - name: chunk_id
15
- dtype: string
16
- splits:
17
- - name: train
18
- num_bytes: 270685935
19
- num_examples: 1066356
20
- download_size: 140092107
21
- dataset_size: 270685935
22
- configs:
23
- - config_name: default
24
- data_files:
25
- - split: train
26
- path: data/train-*
27
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: cc-by-4.0
3
+ language:
4
+ - hu
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  ---
6
+
7
+ # Reddit Dataset (Semantic Chunks)
8
+
9
+ Hungarian Reddit conversations dataset preprocessed with semantic chunking.
10
+
11
+ ## Stats
12
+
13
+ | | |
14
+ |---|---|
15
+ | **Rows** | 1,066,356 |
16
+ | **Tokens** | 42,313,152 |
17
+ | **Tokenizer** | `magyar-nlp-szine-java/exotic_modernbert_128k_tokenizer_modified` |
18
+
19
+ ## Columns
20
+
21
+ - `text` - Chunked text content
22
+ - `token_count` - Token count per chunk
23
+ - `source_id` - Original source row index
24
+ - `chunk_id` - Unique chunk identifier
25
+ - `subreddit` - Source subreddit
26
+ - `type` - Submission or comment