from ...smp import * from ...utils import can_infer import timeout_decorator try: from latex2sympy2 import latex2sympy except Exception as e: logging.critical(f'{type(e)}: {e}') logging.critical('Please install latex2sympy2 by running "pip install latex2sympy2"') FAIL_MSG = 'Failed to obtain answer via API.' @timeout_decorator.timeout(30, use_signals=False) def is_equal(asw: str, gt_asw: str) -> bool: if not isinstance(asw, str) != str or not isinstance(gt_asw, str): print('Warning: input is not string') print(asw, gt_asw) asw = str(asw).lower().strip() gt_asw = str(gt_asw).lower().strip() if gt_asw == asw: return True try: a = eval(gt_asw) b = eval(asw) if abs(a - b) < 1e-6: return True except: pass try: a = latex2sympy(gt_asw) b = latex2sympy(asw) if abs(eval(str(a)) - eval(str(b))) < 1e-6: return True if abs(a - b) < 1e-6: return True except: pass return False def get_gpt4_ICE(): example_1 = """ Hint: Please answer the question and provide the final answer at the end.\n Question: Which number is missing?\n Model response: The number missing in the sequence is 14.\n Extracted answer: 14 """ example_2 = """ Hint: Please answer the question and provide the final answer at the end.\n Question: What is the fraction of females facing the camera?\n Model response: The fraction of females facing the camera is 0.6, which means that six out of ten females in the group are facing the camera.\n Extracted answer: 0.6 """ example_3 = """ Hint: Please answer the question and provide the final answer at the end.\n Question: How much money does Luca need to buy a sour apple candy and a butter-scotch candy? (Unit: $)\n Model response: Luca needs $1.45 to buy a sour apple candy and a butterscotch candy.\n Extracted answer: 1.45 """ example_4 = """ Hint: Please answer the question and provide the final answer at the end.\n Question: Between which two years does the line graph saw its maximum peak?\n Model response: The line graph saw its maximum peak between 2007 and 2008.\n Extracted answer: [2007, 2008] """ example_5 = """ Hint: Please answer the question and provide the correct option letter, e.g., A, B, C, D, at the end.\n Question: What fraction of the shape is blue?\n Choices: (A) 3/11 (B) 8/11 (C) 6/11 (D) 3/5\n Model response: The correct answer is (B) 8/11.\n Extracted answer: B """ return [example_1, example_2, example_3, example_4, example_5] def build_mathv_gpt4_prompt(line): task_description = """ Please read the following example. Then extract the answer from the model response and type it at the end of the prompt.\n """ question = line['question'] prediction = str(line['prediction']) prompt = task_description examples = get_gpt4_ICE() for example in examples: prompt += example + '\n' prompt += question + '\n' prompt += 'Model respone: ' + prediction prompt += 'Extracted answer:' return prompt def list_to_dict(lst): return {chr(65 + i): val for i, val in enumerate(lst)} def post_check(line, prefetch=False): res = None ans = line['answer'] response = line['prediction'] if prefetch else line['res'] try: if len(eval(line['choices'])) > 0: ans = line['answer'] choices = list_to_dict(eval(line['choices'])) res = can_infer(response, choices) if prefetch: return res else: res = str(response) ans = str(ans) except ValueError: pass try: if is_equal(res, ans): return res if prefetch else True else: return False except Exception as err: logging.warning(f'{type(err)}: {err}') return False def MATH_V_auxeval(model, line): prompt = build_mathv_gpt4_prompt(line) log = '' retry = 5 if post_check(line, prefetch=True): res = post_check(line, prefetch=True) return dict(log='Prefetch succeed', res=res) for i in range(retry): prediction = line['prediction'] res = model.generate(prompt, temperature=i * 0.5) if FAIL_MSG in res: log += f'Try {i}: output is {prediction}, failed to parse.\n' else: log += 'Succeed' return dict(log=log, res=res) log += 'All 5 retries failed.\n' return dict(log=log, res='') def MATH_V_acc(result_file): data = load(result_file) tot = defaultdict(lambda: 0) fetch = defaultdict(lambda: 0) hit = defaultdict(lambda: 0) lt = len(data) from tqdm import tqdm for i in tqdm(range(lt)): item = data.iloc[i] cate = item['category'] tot['Overall'] += 1 tot[cate] += 1 if item['log'] == 'Prefetch succeed': fetch['Overall'] += 1 fetch[cate] += 1 if post_check(item, prefetch=False): hit['Overall'] += 1 hit[cate] += 1 res = defaultdict(list) for k in tot.keys(): res['Subject'].append(k) res['tot'].append(tot[k]) res['prefetch'].append(fetch[k]) res['hit'].append(hit[k]) res['prefetch_rate'].append(fetch[k] / tot[k] * 100) res['acc'].append(hit[k] / tot[k] * 100) res = pd.DataFrame(res).sort_values('Subject', ignore_index=True) return res