zhber commited on
Commit
289f571
·
verified ·
1 Parent(s): 1fc8d73

Delete hf_README.md

Browse files
Files changed (1) hide show
  1. hf_README.md +0 -109
hf_README.md DELETED
@@ -1,109 +0,0 @@
1
- ---
2
- dataset_info:
3
- - config_name: release_v1
4
- features:
5
- - name: id
6
- dtype: int64
7
- - name: problem_id
8
- dtype: string
9
- - name: problem_desc
10
- dtype: string
11
- - name: time_limit_ms
12
- dtype: int64
13
- - name: memory_limit_MB
14
- dtype: int64
15
- - name: checker
16
- dtype: string
17
- - name: test_cases
18
- list:
19
- - name: input
20
- dtype: string
21
- - name: output
22
- dtype: string
23
- data_files:
24
- - split: test
25
- path: release_v1/test_50375e15.parquet
26
- license: cc-by-4.0
27
- language:
28
- - en
29
- tags:
30
- - benchmark
31
- - competitive-programming
32
- task_categories:
33
- - text-generation
34
- ---
35
- # CF-Div2-Step Evaluation Benchmark
36
-
37
- ## Introduction
38
-
39
- We introduce **CF-Div2-Stepfun**, a dataset curated to benchmark the competitive programming capabilities of Large Language Models (LLMs). We evaluate our proprietary **Step 3.5 Flash** ([visit website](https://static.stepfun.com/blog/step-3.5-flash/)) alongside several frontier models on this benchmark.
40
-
41
- The benchmark comprises 53 problems sourced from official CodeForces Division 2 contests held between September 2024 and February 2025. We develop an offline evaluation framework that utilizes a local grading mechanism as an alternative to real-time online submissions.
42
-
43
- The generated test cases consist of:
44
-
45
- - Small-scale test cases, for initial functional verification.
46
- - Randomized large-scale data, for performance and complexity verification.
47
- - Handcrafted edge cases, derived from common error patterns and "hacked" submissions from real contest participants.
48
- - Automated stress testing data, generated by stress testing technique, which keeps generating test cases until one can distinguish failed submissions from correct submissions.
49
-
50
- To validate the reliability of this benchmark, we run both correct and representative failed submissions from the original contests. Our evaluator correctly identify 100% of the accepted submissions as "Passed," while 92.45% of the failed submissions are accurately flagged.
51
-
52
- ## Quickstart
53
-
54
- ```python
55
- from datasets import load_dataset
56
-
57
- dataset = load_dataset("stepfun-ai/CF-Div2-Stepfun", name="release_v1", split="test")
58
- ```
59
-
60
- ## Evaluation Details
61
-
62
- The evaluation results are shown below.
63
-
64
- | Model | C++ (avg@8) | Python (avg@8) | Java (avg@8) | C++(pass@8 rating) |
65
- | Step 3.5 Flash | **86.1%** | **81.5%** | 77.1% | **2489** |
66
- | Gemini 3.0 Pro | 83.5% | 74.1% | **81.6%** | 2397 |
67
- | Deepseek V3.2 | 81.6% | 66.5% | 80.7% | 2319 |
68
- | GLM-4.7 | 74.1% | 63.0% | 70.5% | 2156 |
69
- | Claude Opus 4.5 | 72.2% | 68.4% | 68.9% | 2100 |
70
- | Kimi K2-Thinking | 67.9% | 60.4% | 58.5% | 1976 |
71
- | Minimax-M2.1 | 59.0% | 46.4% | 58.0% | 1869 |
72
- | Mimo-V2 Flash | 46.9% | 43.6% | 39.6% | 1658 |
73
-
74
- We use the following prompt for all model evaluations:
75
-
76
- ```
77
- You are a coding expert. Given a competition-level coding problem, you need to write a {LANGUAGE} program to solve it. You may start by outlining your thought process. In the end, please provide the complete code in a code block enclosed with ``` ```.
78
- {question}
79
- ```
80
-
81
- The compilation and execution commands for C++, Python, Java are given below:
82
-
83
- ```
84
- g++ -std=c++20 -fno-asm -fsanitize=bounds -fno-sanitize-recover=bounds –static -O2 -DONLINE_JUDGE -o code.exe code.cpp
85
- ./code.exe
86
- ```
87
-
88
- ```
89
- python3 code.py
90
- ```
91
-
92
- ```
93
- javac -J-Xmx544m {JAVA_CLASS_NAME}.java
94
- java -XX:+UseSerialGC -Xmx544m -Xss64m -DONLINE_JUDGE {JAVA_CLASS_NAME}
95
- ```
96
-
97
- For Python and Java evaluation, we use a double time limit.
98
-
99
- The benchmark kits follow the [testlib](https://github.com/MikeMirzayanov/testlib) pipeline in validation and evaluation. There is a validator for each problem to check test case integrity, and a specific checker to verify output correctness.
100
-
101
- The rating evaluation follows [CodeELO](https://arxiv.org/abs/2501.01257) methodology. For pass@8 metrics, we calculate the expected score with a fail-penalty but no time-penalty.
102
-
103
- ## License
104
-
105
- We are releasing the benchmark under the Creative Commons Attribution 4.0 International (CC-BY-4.0) license.
106
-
107
- ## Citation
108
-
109
- If you find CF-Div2-Step useful, please consider citing it as: