izzako commited on
Commit
8b7c537
·
verified ·
1 Parent(s): 7e19088

Add dataset description with YAML metadata

Browse files
Files changed (1) hide show
  1. README.md +68 -31
README.md CHANGED
@@ -1,33 +1,70 @@
1
  ---
2
- dataset_info:
3
- features:
4
- - name: text_id
5
- dtype: int64
6
- - name: chunk_id
7
- dtype: int64
8
- - name: pixel_values
9
- sequence:
10
- sequence: uint8
11
- - name: grapheme_token_ids
12
- sequence: int64
13
- - name: llama_token_ids
14
- sequence: int64
15
- - name: text
16
- dtype: string
17
- splits:
18
- - name: train
19
- num_bytes: 14297819833
20
- num_examples: 54017
21
- - name: test
22
- num_bytes: 118272902
23
- num_examples: 450
24
- download_size: 322714534
25
- dataset_size: 14416092735
26
- configs:
27
- - config_name: default
28
- data_files:
29
- - split: train
30
- path: data/train-*
31
- - split: test
32
- path: data/test-*
33
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - ban
4
+ license: cc-by-4.0
5
+ task_categories:
6
+ - text-generation
7
+ - image-to-text
8
+ tags:
9
+ - pixelgpt
10
+ - balinese
11
+ - aksara
12
+ - multimodal
13
+ pretty_name: Balinese PixelGPT Dataset
14
+ size_categories:
15
+ - 10K<n<100K
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  ---
17
+
18
+ # Balinese PixelGPT Dataset
19
+
20
+ This dataset contains preprocessed Balinese text data for training PixelGPT models.
21
+
22
+ ## Dataset Statistics
23
+ - **Language**: Balinese (bali)
24
+ - **Total samples**: 54,467
25
+ - **Train samples**: 54,017
26
+ - **Test samples**: 450
27
+
28
+ ## Tokenizers
29
+ - **Grapheme tokenizer**: izzako/javanese-llama-tokenizer
30
+ - **LLaMA tokenizer**: ernie-research/DualGPT
31
+
32
+ ## Features
33
+ - `text_id`: Document identifier
34
+ - `chunk_id`: Chunk identifier within document
35
+ - `pixel_values`: Rendered pixel representation of aksara text
36
+ - `grapheme_token_ids`: Token IDs from grapheme-based tokenizer (with EOS token)
37
+ - `llama_token_ids`: Token IDs from LLaMA tokenizer (with EOS token)
38
+ - `text`: Original text chunk
39
+
40
+ ## Renderer Configuration
41
+ - **Renderer path**: ../../renderers/m4_renderer
42
+
43
+ ## Usage
44
+ ```python
45
+ from datasets import load_dataset
46
+
47
+ # Load the dataset
48
+ dataset = load_dataset("izzako/balinese-pixelgpt")
49
+
50
+ # Access train and test splits
51
+ train_data = dataset['train']
52
+ test_data = dataset['test']
53
+
54
+ # Example: Get first sample
55
+ sample = train_data[0]
56
+ print(sample['text'])
57
+ ```
58
+
59
+ ## Citation
60
+
61
+ If you use this dataset, please cite:
62
+ ```bibtex
63
+ @dataset{balinese_pixelgpt,
64
+ title = {Balinese PixelGPT Dataset},
65
+ author = {Musa Izzanardi Wijanarko},
66
+ year = {2025},
67
+ publisher = {Hugging Face},
68
+ url = {https://huggingface.co/datasets/izzako/balinese-pixelgpt}
69
+ }
70
+ ```