NoFunEval / src /classification_generation.py
ManavSinghal157's picture
Editing dataset
2b42f63 verified
import os
import pandas as pd
import argparse
from transformers import AutoTokenizer
import jsonlines
from tqdm import tqdm
from vllm import LLM, SamplingParams
#Input all the arguments
parser = argparse.ArgumentParser()
parser.add_argument("--data_subset", type = str, default = "latency", help = "type of non-func requirement")
parser.add_argument("--temperature", type = float, default = 0.0, help = "temperature")
parser.add_argument("--max_new_tokens", type = int, default = 8, help = "max length of tokens")
parser.add_argument("--top_p", type = float, default = 0.95, help = "top_p")
parser.add_argument("--prompt", type = str, default = "base_prompt", help = "type of prompt")
parser.add_argument("--num_samples", type = int, default = 1, help = "number of samples")
parser.add_argument("--model_path", type = str, required = True, help = "HF path for OS models")
parser.add_argument("--load_in_8bit", action = "store_true", help = "Load model in 8bit")
parser.add_argument("--load_in_4bit", action = "store_true", help = "Load model in 4bit")
parser.add_argument("--precision", type = str, default = "fp16", help = "Model precision, from: fp32, fp16 or bf16")
parser.add_argument("--tensor_parallel_size", type = int, default = 1, help = "Tensor parallel size")
parser.add_argument("--swap_space", type = int, default = 4, help = "The size (GiB) of CPU memory per GPU to use as swap space.")
parser.add_argument("--batch_size", type = int, default = 1, help = "Number of examples to send to llm engine at once.")
args = parser.parse_args()
argsdict = vars(args)
# Function to extract the classification prediction
def extract_single_predictions(input_string):
if input_string.strip().split()[0].lower() == "A".lower():
return "A"
elif input_string.strip().split()[0].lower() == "B".lower():
return "B"
return None
model_basename = args.model_path.split("/")[-1]
llm_tokenizer = AutoTokenizer.from_pretrained(
args.model_path,
truncation_side="left",
padding_side="right", # padding on the right is needed to cut off padding in `complete_code`
trust_remote_code=True,
)
GREEDY = True
sampling_params = SamplingParams(
n = 1, # for multisamples we sample multiple times
temperature = args.temperature if not GREEDY else 0.0,
top_p = args.top_p if not GREEDY else 1.0,
top_k = 50 if not GREEDY else -1,
max_tokens = args.max_new_tokens,
stop_token_ids = [llm_tokenizer.eos_token_id])
llm = LLM(model = args.model_path,
tensor_parallel_size = args.tensor_parallel_size,
swap_space = args.swap_space,
trust_remote_code = True)
# Initializing variables
dataset_path = os.path.join("datasets", f"{args.data_subset}.jsonl")
args.num_samples = 1
data = []
max_tokens = []
generations = []
left_prompts = []
right_prompts = []
generations = []
with jsonlines.open(dataset_path) as data_file:
for data_item in data_file:
data.append(data_item)
left_prompts.append(data_item["classification_left_prompt"])
right_prompts.append(data_item["classification_right_prompt"])
print("Starting model inference...")
left_llm_outputs = llm.generate(left_prompts, sampling_params)
left_predictions = [extract_single_predictions(output.outputs[0].text) for output in left_llm_outputs]
right_llm_outputs = llm.generate(right_prompts, sampling_params)
right_predictions = [extract_single_predictions(output.outputs[0].text) for output in right_llm_outputs]
for i, data_item in tqdm(enumerate(left_predictions)):
curr_sample = data[i]
curr_sample["left_output"] = left_predictions[i]
curr_sample["right_output"] = right_predictions[i]
for prompt in ["base_prompt", "coding_concepts", "chain_of_thought", "one_shot", "classification_left_prompt", "classification_right_prompt"]:
if(prompt in curr_sample):
del curr_sample[prompt]
generations.append(curr_sample)
# Saving the generations
generations = pd.DataFrame(generations)
path = os.path.join("generations", "classification", args.data_subset, os.path.split(args.model_path)[1], args.prompt, f"{args.num_samples}_samples")
if not os.path.exists(path):
os.makedirs(path)
path = os.path.join(path, "generated_outputs.jsonl")
generations.to_json(path, orient = "records", lines = True)