Racktic's picture
Upload folder using huggingface_hub
b5beb60 verified
import argparse
import json
import logging
import os
from tqdm import tqdm
from .utils import *
import re
import time
def fast_extract_answer(response) :
response = response.strip()
response = process_answer(response)
# Direct Strategy Multi-Choice
# A / A: / A.
for ch in 'ABCDEFGH':
if response.upper() == ch or response.startswith(f'{ch}:') or response.startswith(f'{ch}.'):
return ch
# Direct Strategy Open-ended
# 1
if is_number(response):
return response
# CoT strategy
if 'boxed{' in response:
try:
model_answers = extract_full_boxed_content(response)
if model_answers:
# for coding
# \\boxed{\\text{}}
try:
text_content = re.findall(r'\\text{(.*?)}', model_answers[-1])
if text_content:
return text_content[-1].strip()
except Exception:
pass
return model_answers[-1].strip()
except Exception:
pass
# for Coding
# the correct answer is\n D.
for flag in ['final answer is', 'correct answer is', 'answer should be', 'answer is', 'answer:']:
if flag in response.lower():
try:
model_answer = response.lower().split(flag)[-1].strip()
return model_answer.split('\n')[0].split('.')[0]
except Exception:
pass
return ""
def create_test_prompt(score_prompt, problem, label):
score_prompt = score_prompt.strip()
response = problem[label]
answer = problem['answer']
full_prompt = f'{score_prompt}\n' + f'Response: {response}\n' + f'Answer: {answer}\n' + 'Correct_or_not:'
return full_prompt
def call_gpt(client, model, user_prompt):
attempt = 0
while attempt < 5:
try:
response = client.chat.completions.create(
model=model,
messages=[
{"role": "user", "content": user_prompt}
]
)
return response.choices[0].message.content.strip()
except Exception as e:
logging.error(f"Attempt {attempt + 1} failed: {e}")
if 'error' in str(e) and 'message' in str(e):
error_message = str(e)
if 'The server had an error processing your request.' in error_message:
sleep_time = 30
logging.error(f"Server error, retrying in {sleep_time}s...")
time.sleep(sleep_time)
elif 'Please try again in ' in error_message:
sleep_time = float(error_message.split('Please try again in ')[1].split('s.')[0])
logging.error(f"Rate limit exceeded, retrying in {sleep_time * 2}s...")
time.sleep(sleep_time * 2)
else:
print("Unknown error, skipping this request.")
break
attempt += 1
def gen_true_false(answer_file, response_label='response', gpt_eval=False, model="", api_key="", rerun=True, save_every=20, logger=logging.getLogger(__name__)):
logger.info(f"Reading {answer_file}.....")
label = response_label
if gpt_eval:
from openai import OpenAI
client = OpenAI(api_key=api_key)
with open(answer_file, "r") as f:
results = json.load(f)
full_pids = list(results.keys())
skip_pids = []
# for pid, problem in results.items():
# flag = problem.get('true_false')
# if flag is not None:
# skip_pids.append(problem['pid'])
if rerun:
test_pids = full_pids
else:
if len(skip_pids) > 0:
logger.info(
f"Found existing results file with {len(skip_pids)} problems with valid responses. Skipping these problems..."
)
test_pids = [pid for pid in full_pids if pid not in skip_pids]
logger.info(f"Number of test problems to run: {len(test_pids)}")
for i, pid in enumerate(tqdm(test_pids)):
problem = results[pid]
flag = False
if label not in problem or not problem[label]:
results[pid]['extraction'] = None
results[pid]['true_false'] = False
continue
if gpt_eval:
user_prompt = create_test_prompt(score_demo_prompt, problem, label)
flag_cache = call_gpt(client, model, user_prompt)
results[pid]['gpt_eval'] = flag_cache
if flag_cache.lower() == 'correct':
flag = True
else:
flag = False
else:
model_answer = fast_extract_answer(problem[label])
results[pid]['extraction'] = model_answer
if is_equal(model_answer, results[pid]['answer']) or is_equal(model_answer, results[pid]['gt_content']):
flag = True
results[pid]['true_false'] = flag
if (i % save_every == 0 and i > 0) or i == len(test_pids) - 1:
with open(answer_file, "w") as f:
f.write(json.dumps(results, indent=2))
logger.info(f"Saved results to {answer_file}")
with open(answer_file, "w") as f:
f.write(json.dumps(results, indent=2))
logger.info(f"Saved results to {answer_file}")
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--results_dir', type=str, default='')
parser.add_argument('--response_label', type=str, default='response', help='response label for the input file')
parser.add_argument('--rerun', action='store_true', help='rerun the answer extraction')
parser.add_argument('--save_every', type=int, default=10, help='save every n problems')
parser.add_argument('--gpt_eval', action='store_true', help='use gpt to evaluate')
parser.add_argument('--api_key', type=str, default="")
parser.add_argument('--model', type=str, default="chatgpt-4o-latest")
args = parser.parse_args()
logging.info("Starting to extract answers.......")
for root, dirs, files in os.walk(args.results_dir):
for file in files:
if file.endswith(".json") and not file.endswith("_result.json"):
gen_true_false(os.path.join(root, file), args)
if __name__ == "__main__":
logging.basicConfig(
level=os.environ.get("LOGLEVEL", "INFO").upper(),
format="[%(name)s] %(message)s",
datefmt="[%X]"
)
logger_blocklist = [
"asyncio",
"azure",
"azureml",
"datasets",
"httpx",
"httpcore",
"filelock",
"fsspec",
"msal",
"msrest",
"openai",
"PIL",
"urllib3",
]
for module in logger_blocklist:
logging.getLogger(module).setLevel(logging.WARNING)
main()