gtang666 commited on
Commit
57e66b3
·
verified ·
1 Parent(s): d85617d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -43
README.md CHANGED
@@ -6,73 +6,85 @@ task_categories:
6
  language:
7
  - zh
8
  - en
9
-
10
- dataset_info:
11
- - config_name: Full-page OCR
12
- features:
13
- - name: image_filename
14
- dtype: string
15
- - name: image
16
- dtype: image
17
- - name: label
18
- dtype: string
19
- splits:
20
- - name: test
21
-
22
  configs:
23
- - config_name: Full-page OCR # 例如:default
 
 
 
 
24
  data_files:
25
  - split: test
26
- path:
27
- - data/easy/*.parquet # 指定easy子文件夹下所有.parquet文件
28
- - data/medium/*.parquet # 指定medium子文件夹下所有.parquet文件
29
- - data/hard/*.parquet # 指定hard子文件夹下所有.parquet文件
 
30
  tags:
31
  - art
 
 
 
 
32
  size_categories:
33
  - 1K<n<10K
34
-
35
  ---
36
 
37
-
38
  # 🧠 CalliReader: Contextualizing Chinese Calligraphy via an Embedding-aligned Vision Language Model
 
39
  <div align="center">
40
- <a href="https://github.com/LoYuXr/CalliReader">📂 Code</a>
41
  <a href="https://arxiv.org/pdf/2503.06472">📄 Paper</a>
42
  </div>
43
 
44
- **CalliBench** is aimed to comprehensively evaluate VLMs' performance on the recognition and understanding of Chinese calligraphy.
 
 
 
 
 
 
45
 
 
 
 
 
46
 
 
 
 
 
47
 
48
- ## 📦 Dataset Summary
 
 
 
49
 
50
- * **Samples**: 3,192 image–annotation pairs
51
- * **Tasks**: **Full-page recognition** and **Contextual VQA** (choice of author/layout/style, bilingual interpretation, and intent analysis).
52
- * **Annotations**:
53
-
54
- * Metadata of author, layout, and style.
55
- * Fine-grained annotations of **character-wise bounding boxes and labels**.
56
- * Certain samples include **contextual VQA**.
57
 
58
- ## 🧪 How To Use
59
- All **.parqeut** files of different tiers can be found in the sub-folders of **data**. **Pandas** can be used to parse and further process those files.
 
 
60
 
61
- For example, to load a sample and convert its image into a .jpg file:
62
- ```
 
 
 
63
  import pandas as pd
64
  import io
65
  from PIL import Image
66
 
67
- df = pd.read_parquet('./data/hard/hard.parquet')
68
-
69
- image_data = df.iloc[0]['image']
70
- image = Image.open(io.BytesIO(image_data['bytes']))
71
-
72
- image.save('output_image.jpg')
73
 
74
- ```
 
75
 
76
- ## 🤗 License
 
77
 
78
- Apache 2.0 open for research and commercial use.
 
 
 
 
6
  language:
7
  - zh
8
  - en
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  configs:
10
+ - config_name: easy
11
+ data_files:
12
+ - split: test
13
+ path: data/easy/*.parquet
14
+ - config_name: medium
15
  data_files:
16
  - split: test
17
+ path: data/medium/*.parquet
18
+ - config_name: hard
19
+ data_files:
20
+ - split: test
21
+ path: data/hard/*.parquet
22
  tags:
23
  - art
24
+ - calligraphy
25
+ - chinese
26
+ - visual-language-model
27
+ - document-understanding
28
  size_categories:
29
  - 1K<n<10K
 
30
  ---
31
 
 
32
  # 🧠 CalliReader: Contextualizing Chinese Calligraphy via an Embedding-aligned Vision Language Model
33
+
34
  <div align="center">
35
+ <a href="https://github.com/LoYuXr/CalliReader">📂 Code</a>
36
  <a href="https://arxiv.org/pdf/2503.06472">📄 Paper</a>
37
  </div>
38
 
39
+ ## Dataset Description
40
+
41
+ **CalliBench** is a comprehensive benchmark designed to evaluate Vision-Language Models (VLMs) on the recognition and understanding of Chinese calligraphy.
42
+
43
+ ### Dataset Subsets
44
+
45
+ CalliBench contains three distinct subsets organized by difficulty level:
46
 
47
+ #### 🟢 Easy Subset
48
+ - **Samples**: ~1,000 image-annotation pairs
49
+ - **Characteristics**: Standard script styles, clear layouts, well-known works
50
+ - **Use Case**: Basic calligraphy recognition and transcription
51
 
52
+ #### 🟡 Medium Subset
53
+ - **Samples**: ~1,000 image-annotation pairs
54
+ - **Characteristics**: Mixed script styles, moderate complexity, historical variants
55
+ - **Use Case**: Intermediate-level analysis and contextual understanding
56
 
57
+ #### 🔴 Hard Subset
58
+ - **Samples**: ~1,192 image-annotation pairs
59
+ - **Characteristics**: Cursive scripts, complex layouts, rare characters, damaged works
60
+ - **Use Case**: Advanced contextual reasoning and expert-level interpretation
61
 
62
+ ### Dataset Summary
 
 
 
 
 
 
63
 
64
+ - **Total Samples**: 3,192 image–annotation pairs
65
+ - **Tasks**:
66
+ - **Full-page recognition**: Complete transcription of calligraphy works
67
+ - **Contextual VQA**: Author identification, layout analysis, style classification, bilingual interpretation, and intent analysis
68
 
69
+ ### Usage Examples
70
+
71
+ #### Loading Specific Subsets
72
+
73
+ ```python
74
  import pandas as pd
75
  import io
76
  from PIL import Image
77
 
78
+ # Load easy subset
79
+ df_easy = pd.read_parquet('./data/easy/easy.parquet')
 
 
 
 
80
 
81
+ # Load medium subset
82
+ df_medium = pd.read_parquet('./data/medium/medium.parquet')
83
 
84
+ # Load hard subset
85
+ df_hard = pd.read_parquet('./data/hard/hard.parquet')
86
 
87
+ # Access sample data
88
+ sample = df_hard.iloc[0]
89
+ image_data = sample['image']
90
+ image = Image.open(io.BytesIO(image_data['bytes']))