Improve dataset card: Add description, links, comprehensive usage, and update metadata

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +126 -5
README.md CHANGED
@@ -1,12 +1,133 @@
1
  ---
 
 
2
  license: mit
3
  task_categories:
 
4
  - text-generation
5
- language:
6
- - en
7
  tags:
8
  - code
9
  - cryptography
10
- size_categories:
11
- - 10M<n<100M
12
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
  license: mit
5
  task_categories:
6
+ - question-answering
7
  - text-generation
 
 
8
  tags:
9
  - code
10
  - cryptography
11
+ - benchmark
12
+ ---
13
+
14
+ # AICrypto: A Comprehensive Benchmark for Evaluating Cryptography Capabilities of Large Language Models
15
+
16
+ This repository contains **AICrypto**, the first comprehensive benchmark designed to evaluate the cryptography capabilities of Large Language Models (LLMs), as presented in the paper [AICrypto: A Comprehensive Benchmark for Evaluating Cryptography Capabilities of Large Language Models](https://huggingface.co/papers/2507.09580).
17
+
18
+ The benchmark comprises 135 multiple-choice questions, 150 capture-the-flag (CTF) challenges, and 18 proof problems, covering a broad range of skills from factual memorization to vulnerability exploitation and formal reasoning. All tasks are carefully reviewed or constructed by cryptography experts to ensure correctness and rigor.
19
+
20
+ - **Paper**: [AICrypto: A Comprehensive Benchmark for Evaluating Cryptography Capabilities of Large Language Models](https://huggingface.co/papers/2507.09580)
21
+ - **Project Page**: https://aicryptobench.github.io/
22
+ - **Code Repository**: https://github.com/wangyu-ovo/aicrypto-agent
23
+
24
+ ## Download Dataset
25
+
26
+ To use this dataset with the associated code, download its contents from this Hugging Face repository and place them in the `./data` directory within the code repository.
27
+
28
+ ## Setup
29
+
30
+ Before using the dataset with the provided code, you need to set up your environment as detailed in the [GitHub repository](https://github.com/wangyu-ovo/aicrypto-agent#setup).
31
+
32
+ ### Prerequisites
33
+
34
+ - Python 3.10.15
35
+ - SageMath 10.5
36
+ - yafu 1.34.5
37
+
38
+ ### Installation
39
+
40
+ 1. **Create conda environment:**
41
+ ```shell
42
+ conda env create -f environment.yml
43
+ conda activate crypto
44
+ ```
45
+
46
+ 2. **Install SageMath dependencies:**
47
+ ```shell
48
+ sage -pip install -r sage-requirements.txt
49
+ ```
50
+
51
+ 3. **Install additional tools:**
52
+ - Install flatter: https://github.com/keeganryan/flatter
53
+ - Install yafu: https://github.com/bbuhrow/yafu/tree/master
54
+
55
+ 4. **Configure API keys:**
56
+ Create a `.env` file with your API keys for the models you want to use.
57
+
58
+ ## Sample Usage
59
+
60
+ The associated code repository provides scripts to run and evaluate tasks.
61
+
62
+ ### Run Single CTF Challenge
63
+
64
+ To run a single CTF challenge:
65
+
66
+ ```shell
67
+ python run_single_ctf_task.py --task-path data/CTF/04-RSA/01-blue-hens-2023 --model gpt-4.1
68
+ ```
69
+
70
+ Results will be automatically saved in `./outputs/CTF-0/04-RSA/01-blue-hens-2023/gpt-4.1/run`
71
+
72
+ ### Run CTF Challenges in Parallel for Evaluation
73
+
74
+ To run multiple tasks simultaneously for evaluating multiple models:
75
+
76
+ ```shell
77
+ python run_ctf_parallel.py --jobs 4 --id 0
78
+ ```
79
+
80
+ This command uses 4 processes to run tasks with run ID 0. Results will be automatically saved in `./outputs/CTF-0`.
81
+
82
+ The script will run all models specified in `config/model.yaml`. You can customize the models by modifying `config/model.yaml` and the corresponding model implementations in the `src/model` directory.
83
+
84
+ ### Run Single MCQ Evaluation
85
+
86
+ Run MCQs with a single model:
87
+ ```shell
88
+ python run_choice_question.py --model gpt-4.1
89
+ ```
90
+
91
+ Results are saved to `./outputs/MultipleChoice/<model_name>/`.
92
+
93
+ ### Run MCQ Evaluation in Parallel
94
+
95
+ Run MCQs across multiple models in parallel:
96
+ ```shell
97
+ python batch_run_choice_question.py --parallel --jobs 4
98
+ ```
99
+ Optionally select specific models:
100
+ ```shell
101
+ python batch_run_choice_question.py --parallel --jobs 4 --models gpt-4.1 o3
102
+ ```
103
+
104
+ ### Run Single Proof Task
105
+
106
+ Run proof generation for a specific exam and model:
107
+ ```shell
108
+ python run_proof_task.py --exam 1 --model gpt-4.1
109
+ ```
110
+ Outputs:
111
+ - Proofs: `./outputs/proof/exam1/proof/gpt-4.1_proof_results.tex`
112
+ - Reasoning: `./outputs/proof/exam1/reasoning/gpt-4.1_reasoning_results.tex`
113
+ - Logs: `./outputs/proof/exam1/log/`
114
+
115
+ ### Run Proof Tasks in Parallel for Evaluation
116
+
117
+ Run multiple exams and models concurrently:
118
+ ```shell
119
+ python batch_run_proof_tasks.py --exam-values 1 2 3 --jobs 4
120
+ ```
121
+
122
+ ## Citation
123
+
124
+ If you find this repository useful, please consider citing:
125
+
126
+ ```bibtex
127
+ @article{wang2025aicrypto,
128
+ title={AICrypto: A Comprehensive Benchmark for Evaluating Cryptography Capabilities of Large Language Models},
129
+ author={Wang, Yu and Liu, Yijian and Ji, Liheng and Luo, Han and Li, Wenjie and Zhou, Xiaofei and Feng, Chiyun and Wang, Puji and Cao, Yuhan and Zhang, Geyuan and Li, Xiaojian and Xu, Rongwu and Chen, Yilei and He, Tianxing},
130
+ journal={arXiv preprint arXiv:2507.09580},
131
+ year={2025}
132
+ }
133
+ ```