| | """Preprocessing functions for various benchmark datasets. |
| | |
| | This module provides data loading and prompt formatting functions for: |
| | - Math benchmarks: MATH, GSM8K, AIME, Minerva Math, OmniMath, etc. |
| | - Coding benchmarks: HumanEval, LiveCodeBench, MBPP |
| | - Multiple-choice: MMLU, MMLU Pro, GPQA |
| | - Instruction following: IFEval, IFBench, MT-Bench |
| | - General: AlpacaEval, Arena-Hard |
| | """ |
| |
|
| | import json |
| | import pandas |
| | import os |
| |
|
| |
|
| | def preprocess_gpqa_chatml_template(data_file, use_r1=False, think=True): |
| | """Preprocess GPQA dataset with ChatML template formatting. |
| | |
| | Args: |
| | data_file: Path to GPQA JSON file |
| | use_r1: Whether to use DeepSeek R1-style prompting (default: False) |
| | think: Whether to enable thinking mode (default: True) |
| | |
| | Returns: |
| | list: Formatted prompts with ChatML template |
| | """ |
| | if use_r1: |
| | QUERY_TEMPLATE_MULTICHOICE = "{Question}\n\n\nA. {choice1}\nB. {choice2}\nC. {choice3}\nD. {choice4}\n\nPlease reason step-by-step and put your choice letter without any other text with \\boxed{{}} in the end. Let's think step by step and output the final answer within \\boxed{{}}." |
| | else: |
| | QUERY_TEMPLATE_MULTICHOICE = "Return your final response within \\boxed{{}} and only include the letter choice (e.g., A, B, C, or D) as your final response.\n\n{Question}\n\nAnswer Choices:\n(A) {choice1}\n(B) {choice2}\n(C) {choice3}\n(D) {choice4}" |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant.<|im_end|>\n" |
| |
|
| | with open(data_file, "r") as f: |
| | data_list = json.load(f) |
| |
|
| | prompt_list = [] |
| | for item in data_list: |
| | choices_dict = dict( |
| | Question=item['question'].strip(), |
| | choice1=item['choice_A'].strip(), |
| | choice2=item['choice_B'].strip(), |
| | choice3=item['choice_C'].strip(), |
| | choice4=item['choice_D'].strip() |
| | ) |
| | final_question = QUERY_TEMPLATE_MULTICHOICE.format(**choices_dict) |
| | if use_r1: |
| | final_prompt = """<|begin▁of▁sentence|><|User|>{question}.<|Assistant|><think>\n""".format(question=final_question) |
| | else: |
| | if think: |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + " /think<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| | else: |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + " /no_think<|im_end|>\n<|im_start|>assistant\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_gpqa_raw_template(data_file, use_r1=False, think=True): |
| | """Preprocess GPQA dataset with raw (no template) formatting. |
| | |
| | Args: |
| | data_file: Path to GPQA JSON file |
| | use_r1: Whether to use DeepSeek R1-style prompting (default: False) |
| | think: Whether to enable thinking mode (default: True) |
| | |
| | Returns: |
| | list: Raw formatted prompts without chat template |
| | """ |
| | if use_r1: |
| | QUERY_TEMPLATE_MULTICHOICE = "{Question}\n\n\nA. {choice1}\nB. {choice2}\nC. {choice3}\nD. {choice4}\n\nPlease reason step-by-step and put your choice letter without any other text with \\boxed{{}} in the end. Let's think step by step and output the final answer within \\boxed{{}}." |
| | else: |
| | QUERY_TEMPLATE_MULTICHOICE = "Return your final response within \\boxed{{}} and only include the letter choice (e.g., A, B, C, or D) as your final response.\n\n{Question}\n\nAnswer Choices:\n(A) {choice1}\n(B) {choice2}\n(C) {choice3}\n(D) {choice4}" |
| |
|
| | with open(data_file, "r") as f: |
| | data_list = json.load(f) |
| |
|
| | prompt_list = [] |
| | for item in data_list: |
| | choices_dict = dict( |
| | Question=item['question'].strip(), |
| | choice1=item['choice_A'].strip(), |
| | choice2=item['choice_B'].strip(), |
| | choice3=item['choice_C'].strip(), |
| | choice4=item['choice_D'].strip() |
| | ) |
| | final_question = QUERY_TEMPLATE_MULTICHOICE.format(**choices_dict) |
| | prompt_list.append(final_question) |
| |
|
| | return prompt_list |
| |
|
| |
|
| |
|
| |
|
| | def preprocess_gsm8k_zeroshot_chatml_template(data_file): |
| | """Preprocess GSM8K dataset with zero-shot ChatML template. |
| | |
| | Args: |
| | data_file: Path to GSM8K JSON file |
| | |
| | Returns: |
| | list: Formatted prompts with ChatML template and thinking enabled |
| | """ |
| | with open(data_file, "r") as f: |
| | data_list = json.load(f) |
| |
|
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| |
|
| | prompt_list = [] |
| | for item in data_list: |
| | final_question = item['question'].strip() |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| | def preprocess_gsm8k_zeroshot_raw(data_file): |
| | """Preprocess GSM8K dataset with zero-shot raw formatting. |
| | |
| | Args: |
| | data_file: Path to GSM8K JSON file |
| | |
| | Returns: |
| | list: Raw question prompts without chat template |
| | """ |
| | with open(data_file, "r") as f: |
| | data_list = json.load(f) |
| |
|
| | prompt_list = [] |
| | for item in data_list: |
| | final_question = item['question'].strip() |
| | final_prompt = final_question |
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_humaneval_raw(data_file): |
| | """Preprocess HumanEval code generation dataset. |
| | |
| | Args: |
| | data_file: Path to HumanEval JSON file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Code completion prompts with instructions |
| | - qid_list: Task IDs |
| | """ |
| | qid_list = [] |
| | prompt_list = [] |
| | instruction = "Read the following function signature and docstring, and fully implement the function described. Your response should only contain the code for this function.\n" |
| | with open(data_file, "r") as f: |
| | data_dict = json.load(f) |
| | for key, values in data_dict.items(): |
| | qid_list.append(key) |
| | prompt = instruction + values['prompt'] |
| | prompt_list.append(prompt) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_math_zeroshot_chatml_template(data_file): |
| | """Preprocess MATH dataset with zero-shot ChatML template. |
| | |
| | Args: |
| | data_file: Path to MATH CSV file |
| | |
| | Returns: |
| | list: Formatted prompts with ChatML template and thinking enabled |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| | df = pandas.read_csv(data_file) |
| | test_list = [row.to_dict() for _, row in df.iterrows()] |
| | prompt_list = [] |
| | for item in test_list: |
| | final_question = item['Question'].strip() |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_math500_zeroshot_chatml_template(data_file, use_r1=False): |
| | """Preprocess MATH-500 dataset with zero-shot prompting. |
| | |
| | Args: |
| | data_file: Path to MATH-500 JSONL file |
| | use_r1: Whether to use DeepSeek R1-style prompting (default: False) |
| | |
| | Returns: |
| | list: Formatted prompts with boxed answer instruction |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| |
|
| | prompt_list = [] |
| | with open(data_file, "r") as f: |
| | for line in f: |
| | data_dict = json.loads(line) |
| | final_question = data_dict['problem'].strip() |
| | if use_r1: |
| | final_prompt = """<|begin▁of▁sentence|><|User|>{question}\nPlease reason step by step, and put your final answer within \boxed{{}}.<|Assistant|><think>\n""".format(question=final_question) |
| | else: |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "\n\nPlease place your final answer inside \\boxed{}." + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_minerva_math_chatml_template(data_file): |
| | """Preprocess Minerva Math dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to Minerva Math JSONL file |
| | |
| | Returns: |
| | list: Formatted prompts with boxed answer instruction |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| | prompt_list = [] |
| | with open(data_file, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | final_question = item['problem'].strip() |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "\n\nPlease place your final answer inside \\boxed{}." + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_gaokao2023en_chatml_template(data_file): |
| | """Preprocess Gaokao 2023 English dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to Gaokao 2023 English JSONL file |
| | |
| | Returns: |
| | list: Formatted prompts with boxed answer instruction |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| | prompt_list = [] |
| | with open(data_file, "r", encoding="utf-8") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | final_question = item['question'].strip() |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "\n\nPlease place your final answer inside \\boxed{}." + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_olympiadbench_chatml_template(data_file): |
| | """Preprocess OlympiadBench dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to OlympiadBench JSONL file |
| | |
| | Returns: |
| | list: Formatted prompts with boxed answer instruction |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| | prompt_list = [] |
| | with open(data_file, "r", encoding="utf-8") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | final_question = item['question'].strip() |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "\n\nPlease place your final answer inside \\boxed{}." + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_collegemath_chatml_template(data_file): |
| | """Preprocess College Math dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to College Math JSONL file |
| | |
| | Returns: |
| | list: Formatted prompts with boxed answer instruction |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| | prompt_list = [] |
| | with open(data_file, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | final_question = item['question'].strip() |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "\n\nPlease place your final answer inside \\boxed{}." + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_aime24_chatml_template(data_file): |
| | """Preprocess AIME 2024 dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to AIME 2024 JSONL file |
| | |
| | Returns: |
| | list: Formatted prompts with thinking enabled |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| | prompt_list = [] |
| | with open(data_file, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | final_question = item['question'].strip() |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_aime25_chatml_template(data_file): |
| | """Preprocess AIME 2025 dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to AIME 2025 JSONL file |
| | |
| | Returns: |
| | list: Formatted prompts with thinking enabled |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| | prompt_list = [] |
| | with open(data_file, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | final_question = item['problem'].strip() |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| |
|
| | def preprocess_aime24_raw(data_file): |
| | """Preprocess AIME 2024 dataset with raw formatting. |
| | |
| | Args: |
| | data_file: Path to AIME 2024 JSONL file |
| | |
| | Returns: |
| | list: Raw prompts with boxed answer instruction |
| | """ |
| | prompt_list = [] |
| | with open(data_file, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | final_question = item['question'].strip() |
| | final_prompt = "{question}\nPlease reason step by step, and put your final answer within \\boxed{{}}.".format(question=final_question) |
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_aime25_raw(data_file): |
| | """Preprocess AIME 2025 dataset with raw formatting. |
| | |
| | Args: |
| | data_file: Path to AIME 2025 JSONL file |
| | |
| | Returns: |
| | list: Raw prompts with boxed answer instruction |
| | """ |
| | prompt_list = [] |
| | with open(data_file, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | final_question = item['problem'].strip() |
| | final_prompt = "{question}\nPlease reason step by step, and put your final answer within \\boxed{{}}.".format(question=final_question) |
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_amc23_chatml_template(data_file): |
| | """Preprocess AMC 2023 dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to AMC 2023 JSONL file |
| | |
| | Returns: |
| | list: Formatted prompts with thinking enabled |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| | prompt_list = [] |
| | with open(data_file, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | final_question = item['question'].strip() |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_omnimath_chatml_template(data_file): |
| | """Preprocess OmniMath dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to OmniMath JSONL file |
| | |
| | Returns: |
| | list: Formatted prompts with thinking enabled |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| | prompt_list = [] |
| |
|
| | with open(data_file, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | final_question = item['problem'].strip() |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_ifeval_chatml_template(data_file): |
| | """Preprocess IFEval instruction-following dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to IFEval JSONL file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Formatted prompts with ChatML template |
| | - qid_list: Task IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data = f.readlines() |
| | data_list = [json.loads(x) for x in data] |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for item in data_list: |
| | qid_list.append(item['key']) |
| | first_question = item['prompt'] |
| | final_prompt = "<|im_start|>user\n" + first_question + "<|im_end|>\n<|im_start|>assistant\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_ifeval_raw(data_file): |
| | """Preprocess IFEval instruction-following dataset with raw formatting. |
| | |
| | Args: |
| | data_file: Path to IFEval JSONL file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Raw instruction prompts |
| | - qid_list: Task IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data = f.readlines() |
| | data_list = [json.loads(x) for x in data] |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for item in data_list: |
| | qid_list.append(item['key']) |
| | final_prompt = item['prompt'] |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_ifbench_raw(data_file): |
| | """Preprocess IFBench instruction-following dataset with raw formatting. |
| | |
| | Args: |
| | data_file: Path to IFBench JSONL file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Raw instruction prompts |
| | - qid_list: Task IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data = f.readlines() |
| | data_list = [json.loads(x) for x in data] |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for item in data_list: |
| | qid_list.append(item['key']) |
| | final_prompt = item['prompt'] |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_arena_hard_chatml_template(data_file): |
| | """Preprocess Arena-Hard dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to Arena-Hard JSONL file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Formatted prompts with ChatML template |
| | - qid_list: Question IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data = f.readlines() |
| | data_list = [json.loads(x) for x in data] |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for item in data_list: |
| | qid_list.append(item['question_id']) |
| | first_question = item['turns'][0]['content'] |
| | final_prompt = "<|im_start|>user\n" + first_question + "<|im_end|>\n<|im_start|>assistant\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list, qid_list |
| |
|
| | def preprocess_arena_hard_raw(data_file): |
| | """Preprocess Arena-Hard dataset with raw formatting. |
| | |
| | Args: |
| | data_file: Path to Arena-Hard JSONL file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Raw question prompts |
| | - qid_list: Question IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data = f.readlines() |
| | data_list = [json.loads(x) for x in data] |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for item in data_list: |
| | qid_list.append(item['question_id']) |
| | first_question = item['turns'][0]['content'] |
| |
|
| | prompt_list.append(first_question) |
| |
|
| | return prompt_list, qid_list |
| |
|
| | def preprocess_arena_hard_v2_raw(data_file): |
| | """Preprocess Arena-Hard v2.0 dataset with raw formatting. |
| | |
| | Args: |
| | data_file: Path to Arena-Hard v2.0 JSONL file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Raw question prompts |
| | - qid_list: Unique IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data = f.readlines() |
| | data_list = [json.loads(x) for x in data] |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for item in data_list: |
| | qid_list.append(item['uid']) |
| | first_question = item['prompt'] |
| |
|
| | prompt_list.append(first_question) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_alpaca_eval_raw(data_file): |
| | """Preprocess AlpacaEval dataset with raw formatting. |
| | |
| | Args: |
| | data_file: Path to AlpacaEval JSON file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Raw instruction prompts |
| | - qid_list: Sequential question IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data_list = json.load(f) |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for i, item in enumerate(data_list): |
| | qid_list.append(i) |
| | final_prompt = item['instruction'] |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_alpaca_eval_chatml_template(data_file): |
| | """Preprocess AlpacaEval dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to AlpacaEval JSON file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Formatted prompts with ChatML template |
| | - qid_list: Sequential question IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data_list = json.load(f) |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for i, item in enumerate(data_list): |
| | qid_list.append(i) |
| | first_question = item['instruction'] |
| | final_prompt = "<|im_start|>user\n" + first_question + "<|im_end|>\n<|im_start|>assistant\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| |
|
| | def preprocess_mtbench_firstturn(data_file): |
| | """Preprocess MT-Bench first turn with ChatML template. |
| | |
| | Args: |
| | data_file: Path to MT-Bench JSONL file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: First turn prompts with ChatML template |
| | - qid_list: Question IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data = f.readlines() |
| | data_list = [json.loads(x) for x in data] |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for item in data_list: |
| | qid_list.append(item['question_id']) |
| | first_question = item['turns'][0] |
| | final_prompt = "<|im_start|>user\n" + first_question + "<|im_end|>\n<|im_start|>assistant\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_mtbench_firstturn_raw(data_file): |
| | """Preprocess MT-Bench first turn with raw formatting. |
| | |
| | Args: |
| | data_file: Path to MT-Bench JSONL file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: First turn raw prompts |
| | - qid_list: Question IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data = f.readlines() |
| | data_list = [json.loads(x) for x in data] |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for item in data_list: |
| | qid_list.append(item['question_id']) |
| | first_question = item['turns'][0] |
| |
|
| | prompt_list.append(first_question) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_mtbench_secondturn(data_file, output_file): |
| | """Preprocess MT-Bench second turn with ChatML template. |
| | |
| | Args: |
| | data_file: Path to MT-Bench JSONL file |
| | output_file: Model output file for the first turn of MT-Bench |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Second turn prompts with conversation history |
| | - qid_list: Question IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data = f.readlines() |
| | data_list = [json.loads(x) for x in data] |
| |
|
| | id2output = {} |
| | with open(output_file, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | id2output[item['task_id']] = item['output'] |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for item in data_list: |
| | qid_list.append(item['question_id']) |
| | first_question = item['turns'][0] |
| | second_question = item['turns'][1] |
| |
|
| | model_output = id2output[item['question_id']] |
| |
|
| | final_prompt = "<|im_start|>user\n" + first_question + "<|im_end|>\n<|im_start|>assistant\n" + model_output + "<|im_end|>\n" + \ |
| | "<|im_start|>user\n" + second_question + "<|im_end|>\n<|im_start|>assistant\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_mtbench_secondturn_raw(data_file, output_file): |
| | """Preprocess MT-Bench second turn with raw formatting. |
| | |
| | Args: |
| | data_file: Path to MT-Bench JSONL file |
| | output_file: Model output file for the first turn of MT-Bench |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Second turn prompts as chat message lists |
| | - qid_list: Question IDs |
| | """ |
| | with open(data_file, "r") as f: |
| | data = f.readlines() |
| | data_list = [json.loads(x) for x in data] |
| |
|
| | id2output = {} |
| | with open(output_file, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | id2output[item['task_id']] = (item['output'], item['reason_text']) |
| |
|
| | qid_list = [] |
| | prompt_list = [] |
| | for item in data_list: |
| | qid_list.append(item['question_id']) |
| | first_question = item['turns'][0] |
| | second_question = item['turns'][1] |
| |
|
| | output, reason = id2output[item['question_id']] |
| | model_output = output |
| |
|
| | chat = [ |
| | {'role': 'user', 'content': first_question}, |
| | {'role': 'assistant', 'content': model_output}, |
| | {'role': 'user', 'content': second_question} |
| | ] |
| |
|
| | prompt_list.append(chat) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_mmlu_chatml_template(data_file): |
| | """Preprocess MMLU dataset with few-shot ChatML template. |
| | |
| | Args: |
| | data_file: Path to MMLU CSV file |
| | |
| | Returns: |
| | list: Formatted prompts with few-shot examples and ChatML template |
| | """ |
| | def _load_mmlu_cot_fewshot_examples(): |
| | import yaml |
| | current_folder = os.path.dirname(os.path.abspath(__file__)) |
| | fewshot_folder = os.path.join(current_folder, "flan_cot_fewshot") |
| | file_list = os.listdir(fewshot_folder) |
| | fewshot_dict = {} |
| | for filename in file_list: |
| | with open(os.path.join(fewshot_folder, filename)) as f: |
| | data = yaml.safe_load(f) |
| | dataset_name = data['dataset_name'].strip() |
| | description = data['description'].strip() |
| | sample_list = data['fewshot_config']["samples"] |
| |
|
| | prompt = description |
| | for sample in sample_list: |
| | prompt += "\n\n" |
| | prompt += "Q: " + sample['question'].strip() + "\n" + "A: " + sample['target'].strip() |
| |
|
| | fewshot_dict[dataset_name] = prompt |
| |
|
| | return fewshot_dict |
| |
|
| | fewshot_dict = _load_mmlu_cot_fewshot_examples() |
| | df = pandas.read_csv(data_file) |
| | test_list = [row.to_dict() for _, row in df.iterrows()] |
| |
|
| | prompt_list = [] |
| | for item in test_list: |
| | subject = item['Subject'] |
| | fewshot_prompt = fewshot_dict[subject] |
| | question = item['Question'] |
| | choice_a = str(item['A']).strip() |
| | choice_b = str(item['B']).strip() |
| | choice_c = str(item['C']).strip() |
| | choice_d = str(item['D']).strip() |
| |
|
| | final_question = fewshot_prompt + "\n\n" + "Q: " + question + "\n" |
| | final_question += "(A) " + choice_a + " (B) " + choice_b + " (C) " + choice_c + " (D) " + choice_d + "\n" |
| | final_question += "A: " |
| |
|
| | final_prompt = "<|im_start|>user\n" + final_question + "<|im_end|>\n<|im_start|>assistant\n" |
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_mmlu_raw_template(data_file): |
| | """Preprocess MMLU dataset with few-shot raw formatting. |
| | |
| | Args: |
| | data_file: Path to MMLU CSV file |
| | |
| | Returns: |
| | list: Raw prompts with few-shot examples |
| | """ |
| | def _load_mmlu_cot_fewshot_examples(): |
| | import yaml |
| | current_folder = os.path.dirname(os.path.abspath(__file__)) |
| | fewshot_folder = os.path.join(current_folder, "flan_cot_fewshot") |
| | file_list = os.listdir(fewshot_folder) |
| | fewshot_dict = {} |
| | for filename in file_list: |
| | with open(os.path.join(fewshot_folder, filename)) as f: |
| | data = yaml.safe_load(f) |
| | dataset_name = data['dataset_name'].strip() |
| | description = data['description'].strip() |
| | sample_list = data['fewshot_config']["samples"] |
| |
|
| | prompt = description |
| | for sample in sample_list: |
| | prompt += "\n\n" |
| | prompt += "Q: " + sample['question'].strip() + "\n" + "A: " + sample['target'].strip() |
| |
|
| | fewshot_dict[dataset_name] = prompt |
| |
|
| | return fewshot_dict |
| |
|
| | fewshot_dict = _load_mmlu_cot_fewshot_examples() |
| | df = pandas.read_csv(data_file) |
| | test_list = [row.to_dict() for _, row in df.iterrows()] |
| |
|
| | prompt_list = [] |
| | for item in test_list: |
| | subject = item['Subject'] |
| | fewshot_prompt = fewshot_dict[subject] |
| | question = item['Question'] |
| | choice_a = str(item['A']).strip() |
| | choice_b = str(item['B']).strip() |
| | choice_c = str(item['C']).strip() |
| | choice_d = str(item['D']).strip() |
| |
|
| | final_question = fewshot_prompt + "\n\n" + "Q: " + question + "\n" |
| | final_question += "(A) " + choice_a + " (B) " + choice_b + " (C) " + choice_c + " (D) " + choice_d + "\n" |
| | final_question += "A: " |
| |
|
| | prompt_list.append(final_question) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | QUERY_TEMPLATE_MULTICHOICE = """ |
| | Answer the following multiple choice question. The last line of your response should be of the following format: 'Answer: $LETTER' (without quotes) where LETTER is one of ABCD. Think step by step before answering. |
| | |
| | {Question} |
| | |
| | A) {A} |
| | B) {B} |
| | C) {C} |
| | D) {D} |
| | """.strip() |
| |
|
| | def preprocess_mmlu_r1_raw_template(data_file): |
| | """Preprocess MMLU dataset for R1-style models with zero-shot prompting. |
| | |
| | Args: |
| | data_file: Path to MMLU CSV file |
| | |
| | Returns: |
| | list: Formatted prompts for R1-style reasoning models |
| | """ |
| | df = pandas.read_csv(data_file) |
| | test_list = [row.to_dict() for _, row in df.iterrows()] |
| |
|
| | prompt_list = [] |
| |
|
| |
|
| |
|
| | for item in test_list: |
| | subject = item['Subject'] |
| | question = item['Question'] |
| | choice_a = str(item['A']).strip() |
| | choice_b = str(item['B']).strip() |
| | choice_c = str(item['C']).strip() |
| | choice_d = str(item['D']).strip() |
| |
|
| | |
| | |
| | |
| | final_question = QUERY_TEMPLATE_MULTICHOICE.format(Question=question, A=choice_a, B=choice_b, C=choice_c, D=choice_d) |
| |
|
| | prompt_list.append(final_question) |
| |
|
| | return prompt_list |
| |
|
| | def preprocess_mmlu_r1_raw_template_wdai(data_file): |
| | """Preprocess MMLU dataset with boxed answer format (alternative R1 style). |
| | |
| | Args: |
| | data_file: Path to MMLU CSV file |
| | |
| | Returns: |
| | list: Formatted prompts with boxed answer instruction |
| | """ |
| | MMLU_QUERY_TEMPLATE_MULTICHOICE = "Answer the following multiple-choice question. At the end of your response, conclude with the sentence `The answer is \\boxed{{X}}.`, replacing X with the correct capital letter of your choice.\n\n{Question}\n\nAnswer Choices:\n(A) {choice1}\n(B) {choice2}\n(C) {choice3}\n(D) {choice4}" |
| | df = pandas.read_csv(data_file) |
| | test_list = [row.to_dict() for _, row in df.iterrows()] |
| |
|
| | prompt_list = [] |
| | for item in test_list: |
| | choices_dict = dict( |
| | Question=item['Question'].strip(), |
| | choice1=str(item['A']).strip(), |
| | choice2=str(item['B']).strip(), |
| | choice3=str(item['C']).strip(), |
| | choice4=str(item['D']).strip() |
| | ) |
| | final_question = MMLU_QUERY_TEMPLATE_MULTICHOICE.format(**choices_dict) |
| |
|
| | prompt_list.append(final_question) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_mmlu_pro_chatml_template(data_file, fewshot_file): |
| | """Preprocess MMLU-Pro dataset with 5-shot ChatML template. |
| | |
| | Args: |
| | data_file: Path to MMLU-Pro test JSON file |
| | fewshot_file: Path to MMLU-Pro validation JSON file (for few-shot examples) |
| | |
| | Returns: |
| | list: Formatted prompts with 5-shot examples and ChatML template |
| | """ |
| | def _preprocess(data_list): |
| | output_list = [] |
| | for item in data_list: |
| | options = [] |
| | for opt in item["options"]: |
| | if opt == "N/A": |
| | continue |
| | options.append(opt) |
| | item["options"] = options |
| | output_list.append(item) |
| | return output_list |
| |
|
| | def _categorize_basedon_subject(data_list): |
| | fewshot_dict = {} |
| | for item in data_list: |
| | subject = item['category'] |
| | if subject in fewshot_dict: |
| | fewshot_dict[subject].append(item) |
| | else: |
| | fewshot_dict[subject] = [item] |
| | return fewshot_dict |
| |
|
| | def _format_each_sample(sample, is_test): |
| | choices = ["A", "B", "C", "D", "E", "F", "G", "H", "I", "J", "K", "L", "M", "N", "O", "P"] |
| |
|
| | question = sample['question'] |
| | options = sample['options'] |
| |
|
| | sample_prompt = "Q: " + question + "\n" |
| | for i, opt in enumerate(options): |
| | sample_prompt += "(%s) %s " % (choices[i], opt) |
| | sample_prompt = sample_prompt.strip() + "\n" |
| |
|
| | if is_test: |
| | sample_prompt += "A: " |
| | else: |
| | sample_prompt += sample['cot_content'].strip() |
| |
|
| | return sample_prompt |
| |
|
| | def _get_fewshot_prompt(fewshot_samples, test_sample, subject): |
| | description = "The following are multiple choice questions (with answers) about %s." % subject |
| |
|
| | final_question = description + "\n\n" |
| | for sample in fewshot_samples: |
| | sample_prompt = _format_each_sample(sample, is_test=False) |
| | final_question += sample_prompt + "\n\n" |
| |
|
| | test_prompt = _format_each_sample(test_sample, is_test=True) |
| | final_question += test_prompt |
| |
|
| | return final_question |
| |
|
| | with open(fewshot_file, "r") as f: |
| | fewshot_list = json.load(f) |
| | with open(data_file, "r") as f: |
| | test_list = json.load(f) |
| |
|
| | fewshot_list = _preprocess(fewshot_list) |
| | fewshot_dict = _categorize_basedon_subject(fewshot_list) |
| | test_list = _preprocess(test_list) |
| |
|
| | prompt_list = [] |
| | for test_sample in test_list: |
| | subject = test_sample['category'] |
| | fewshot_samples = fewshot_dict[subject] |
| | |
| | assert len(fewshot_samples) == 5 |
| | final_question = _get_fewshot_prompt(fewshot_samples, test_sample, subject) |
| | final_prompt = "<|im_start|>user\n" + final_question + "<|im_end|>\n<|im_start|>assistant\n" |
| |
|
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_mmlu_pro_zero_shot_chatml_template(data_file, think=True): |
| | """Preprocess MMLU-Pro dataset with zero-shot ChatML template. |
| | |
| | Args: |
| | data_file: Path to MMLU-Pro test JSON file |
| | think: Whether to enable thinking mode (default: True) |
| | |
| | Returns: |
| | list: Formatted prompts with ChatML template and boxed answer instruction |
| | """ |
| | def _preprocess(data_list): |
| | output_list = [] |
| | for item in data_list: |
| | options = [] |
| | for opt in item["options"]: |
| | if opt == "N/A": |
| | continue |
| | options.append(opt) |
| | item["options"] = options |
| | output_list.append(item) |
| | return output_list |
| |
|
| | def _format_each_sample(sample): |
| | choices = ["A", "B", "C", "D", "E", "F", "G", "H", "I", "J", "K", "L", "M", "N", "O", "P"] |
| |
|
| | question = sample['question'] |
| | options = sample['options'] |
| |
|
| | |
| | |
| | |
| | |
| | |
| |
|
| | sample_prompt = "Question:\n" + question + "\n\nAnswer Choices:" |
| | for i, opt in enumerate(options): |
| | sample_prompt += "\n(%s) %s" % (choices[i], opt) |
| | sample_prompt += "\n\nConclude your response with the sentence `The answer is \\boxed{{X}}.`, in which X is the correct capital letter of your choice." |
| | sample_prompt = sample_prompt.strip() + "\n" |
| |
|
| | return sample_prompt |
| |
|
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant.<|im_end|>\n" |
| |
|
| | with open(data_file, "r") as f: |
| | test_list = json.load(f) |
| | test_list = _preprocess(test_list) |
| |
|
| | prompt_list = [] |
| | for test_sample in test_list: |
| | test_prompt = _format_each_sample(test_sample) |
| | if think: |
| | final_prompt = instruction + "<|im_start|>user\n" + test_prompt + "\n /think<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| | else: |
| | final_prompt = instruction + "<|im_start|>user\n" + test_prompt + "\n /no_think<|im_end|>\n<|im_start|>assistant\n" |
| | prompt_list.append(final_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_mmlu_pro_zero_shot_raw_template(data_file, think=True): |
| | """Preprocess MMLU-Pro dataset with zero-shot raw formatting. |
| | |
| | Args: |
| | data_file: Path to MMLU-Pro test JSON file |
| | think: Whether to enable thinking mode (default: True, currently unused) |
| | |
| | Returns: |
| | list: Raw prompts with boxed answer instruction |
| | """ |
| | def _preprocess(data_list): |
| | output_list = [] |
| | for item in data_list: |
| | options = [] |
| | for opt in item["options"]: |
| | if opt == "N/A": |
| | continue |
| | options.append(opt) |
| | item["options"] = options |
| | output_list.append(item) |
| | return output_list |
| |
|
| | def _format_each_sample(sample): |
| | choices = ["A", "B", "C", "D", "E", "F", "G", "H", "I", "J", "K", "L", "M", "N", "O", "P"] |
| |
|
| | question = sample['question'] |
| | options = sample['options'] |
| |
|
| | |
| | |
| | |
| | |
| | |
| |
|
| | sample_prompt = "Question:\n" + question + "\n\nAnswer Choices:" |
| | for i, opt in enumerate(options): |
| | sample_prompt += "\n(%s) %s" % (choices[i], opt) |
| | sample_prompt += "\n\nConclude your response with the sentence `The answer is \\boxed{{X}}.`, in which X is the correct capital letter of your choice." |
| | sample_prompt = sample_prompt.strip() + "\n" |
| |
|
| | return sample_prompt |
| |
|
| | with open(data_file, "r") as f: |
| | test_list = json.load(f) |
| | test_list = _preprocess(test_list) |
| |
|
| | prompt_list = [] |
| | for test_sample in test_list: |
| | test_prompt = _format_each_sample(test_sample) |
| | prompt_list.append(test_prompt) |
| |
|
| | return prompt_list |
| |
|
| |
|
| | def preprocess_livecodebench_chatml_template(data_file): |
| | """Preprocess LiveCodeBench dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to LiveCodeBench JSON file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Formatted coding prompts with ChatML template |
| | - qid_list: Question IDs |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| |
|
| | code_instruction_nostartercode = """Write Python code to solve the problem. Please place the solution code in the following format:\n```python\n# Your solution code here\n```""" |
| | code_instruction_hasstartercode = """Please place the solution code in the following format:\n```python\n# Your solution code here\n```""" |
| |
|
| | with open(data_file, "r") as f: |
| | data_list = json.load(f) |
| |
|
| | prompt_list = [] |
| | qid_list = [] |
| | for item in data_list: |
| | question = item['question_content'].strip() |
| | if item['starter_code'] != "": |
| | question += "\n\n" + "Solve the problem starting with the provided function header.\n\nFunction header:\n" + "```\n" + item['starter_code'] + "\n```" |
| | question += "\n\n" + code_instruction_hasstartercode |
| | else: |
| | question += "\n\n" + code_instruction_nostartercode |
| |
|
| | final_prompt = instruction + "<|im_start|>user\n" + question + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| |
|
| | prompt_list.append(final_prompt) |
| | qid_list.append(item['question_id']) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_livecodebench_raw(data_file): |
| | """Preprocess LiveCodeBench dataset with raw formatting. |
| | |
| | Args: |
| | data_file: Path to LiveCodeBench JSON file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Raw coding prompts |
| | - qid_list: Question IDs |
| | """ |
| | code_instruction_nostartercode = """Write Python code to solve the problem. Please place the solution code in the following format:\n```python\n# Your solution code here\n```""" |
| | code_instruction_hasstartercode = """Please place the solution code in the following format:\n```python\n# Your solution code here\n```""" |
| |
|
| | with open(data_file, "r") as f: |
| | data_list = json.load(f) |
| |
|
| | prompt_list = [] |
| | qid_list = [] |
| | for item in data_list: |
| | question = item['question_content'].strip() |
| | if item['starter_code'] != "": |
| | question += "\n\n" + "Solve the problem starting with the provided function header.\n\nFunction header:\n" + "```\n" + \ |
| | item['starter_code'] + "\n```" |
| | question += "\n\n" + code_instruction_hasstartercode |
| | else: |
| | question += "\n\n" + code_instruction_nostartercode |
| |
|
| | final_prompt = question |
| |
|
| | prompt_list.append(final_prompt) |
| | qid_list.append(item['question_id']) |
| |
|
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_mbpp_chatml_template(data_file): |
| | """Preprocess MBPP (Mostly Basic Python Problems) dataset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to MBPP JSON file |
| | |
| | Returns: |
| | tuple: (prompt_list, qid_list) |
| | - prompt_list: Formatted code generation prompts with ChatML template |
| | - qid_list: Task IDs |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| | |
| | with open(data_file, "r") as f: |
| | data_dict = json.load(f) |
| | |
| | prompt_list = [] |
| | qid_list = [] |
| | |
| | for key, value in data_dict.items(): |
| | qid_list.append(key) |
| | question = value.get('text', value.get('prompt', '')) |
| | final_prompt = instruction + "<|im_start|>user\n" + question + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| | prompt_list.append(final_prompt) |
| | |
| | return prompt_list, qid_list |
| |
|
| |
|
| | def preprocess_mmlu_stem_chatml_template(data_file): |
| | """Preprocess MMLU STEM subset with ChatML template. |
| | |
| | Args: |
| | data_file: Path to MMLU STEM JSON file |
| | |
| | Returns: |
| | list: Formatted prompts with ChatML template |
| | """ |
| | instruction = "<|im_start|>system\nYou are a helpful and harmless assistant. You should think step-by-step.<|im_end|>\n" |
| | |
| | with open(data_file, "r") as f: |
| | data_list = json.load(f) |
| | |
| | prompt_list = [] |
| | for item in data_list: |
| | question = item['question'].strip() |
| | choices = item.get('choices', []) |
| | |
| | final_question = question + "\n" |
| | if choices: |
| | for i, choice in enumerate(choices): |
| | final_question += f"({chr(65+i)}) {choice}\n" |
| | |
| | final_prompt = instruction + "<|im_start|>user\n" + final_question + "<|im_end|>\n<|im_start|>assistant\n<think>\n" |
| | prompt_list.append(final_prompt) |
| | |
| | return prompt_list |
| |
|
| |
|