NoFunEval / src /nofunedit_generation.py
ManavSinghal157's picture
Editing dataset
2b42f63 verified
import os
import time
import argparse
import jsonlines
from tqdm import tqdm
import pandas as pd
from transformers import AutoTokenizer
from vllm import LLM, SamplingParams
#Input all the arguments
parser = argparse.ArgumentParser()
parser.add_argument("--data_subset", type = str, default = "latency", help = "type of non-func requirement")
parser.add_argument("--temperature", type = float, default = 0.0, help = "temperature")
parser.add_argument("--max_new_tokens", type = int, default = 5192, help = "max length of tokens")
parser.add_argument("--top_p", type = float, default = 0.95, help = "top_p")
parser.add_argument("--prompt", type = str, default = "base_prompt", help = "type of prompt")
parser.add_argument("--num_samples", type = int, default = 1, help = "number of samples")
parser.add_argument("--model_path", type = str, required=True, help = "HF path for OS models")
parser.add_argument("--load_in_8bit", action="store_true", help = "Load model in 8bit")
parser.add_argument("--load_in_4bit", action="store_true", help = "Load model in 4bit")
parser.add_argument("--precision", type = str, default = "fp16", help = "Model precision, from: fp32, fp16 or bf16")
parser.add_argument("--tensor_parallel_size", type = int, default = 1, help = "Tensor parallel size")
parser.add_argument("--swap_space", type = int, default = 4, help = "The size (GiB) of CPU memory per GPU to use as swap space.")
parser.add_argument("--batch_size", type = int, default = 1, help = "Number of examples to send to llm engine at once.")
args = parser.parse_args()
argsdict = vars(args)
def model_query(all_messages, batch_size = 1):
all_messages = [messages[0]["content"] for messages in all_messages]
llm_tokenizer = AutoTokenizer.from_pretrained(
args.model_path,
truncation_side="left",
padding_side="right", # padding on the right is needed to cut off padding in `complete_code`
)
if args.num_samples == 1:
GREEDY = True
else:
GREEDY = False
assert args.num_samples % batch_size == 0, "num_samples must be divisible by batch_size"
sampling_params = SamplingParams(
n = batch_size, # for multisamples we sample multiple times
temperature = args.temperature if not GREEDY else 0.0,
top_p = args.top_p if not GREEDY else 1.0,
top_k = 50 if not GREEDY else -1,
max_tokens = args.max_new_tokens,
stop_token_ids = [llm_tokenizer.eos_token_id])
llm = LLM(model = args.model_path,
tensor_parallel_size = args.tensor_parallel_size,
swap_space = args.swap_space)
start_time = time.time()
for turn_id in tqdm(range(0, args.num_samples//batch_size)):
llm_outputs = llm.generate(all_messages, sampling_params)
if turn_id == 0:
all_generated_answers = [[llm_output.prompt + llm_gen.text
for llm_gen in llm_output.outputs]
for llm_output in llm_outputs]
else:
for idx, llm_output in enumerate(llm_outputs):
all_generated_answers[idx].extend([llm_output.prompt + llm_gen.text
for llm_gen in llm_output.outputs])
total_time = time.time() - start_time
avg_times = [total_time / len(all_messages)] * len(all_messages)
return all_generated_answers, avg_times
dataset_path = os.path.join("datasets", f"{args.data_subset}.jsonl")
data = []
max_tokens = []
generations = []
all_messages = []
with jsonlines.open(dataset_path) as data_file:
for data_item in data_file:
data.append(data_item)
content = data_item[args.prompt]
messages=[{"role": "user", "content": content}]
all_messages.append(messages)
print("Starting model inference...")
all_generated_answers, all_inference_times = model_query(all_messages = all_messages, batch_size = args.batch_size)
for i, data_item in tqdm(enumerate(data)):
generated_answers = all_generated_answers[i]
inference_time = all_inference_times[i]
curr_sample = data_item
curr_sample["inference_time"] = inference_time
curr_sample["generated_answers"] = generated_answers
for prompt in ["base_prompt", "coding_concepts", "chain_of_thought", "one_shot"]:
del curr_sample[prompt]
generations.append(curr_sample)
generations = pd.DataFrame(generations)
path = os.path.join("generations", "edit", args.data_subset, os.path.split(args.model_path)[1], args.prompt, f"{args.num_samples}_samples")
if not os.path.exists(path):
os.makedirs(path)
path=os.path.join(path, "generated_outputs.jsonl")
generations.to_json(path, orient = "records", lines=True)