File size: 4,938 Bytes
3f53269
 
 
 
 
2b42f63
 
3f53269
 
 
 
 
2b42f63
 
 
 
 
 
 
 
 
 
 
 
 
3f53269
 
 
 
2b42f63
 
3f53269
2b42f63
3f53269
 
 
 
 
2b42f63
3f53269
 
 
 
2b42f63
3f53269
2b42f63
3f53269
2b42f63
 
 
 
 
 
 
 
 
 
 
3f53269
2b42f63
3f53269
2b42f63
3f53269
2b42f63
3f53269
 
 
 
 
 
 
 
2b42f63
3f53269
 
2b42f63
3f53269
 
 
2b42f63
3f53269
 
2b42f63
 
 
 
3f53269
2b42f63
3f53269
2b42f63
3f53269
 
 
 
 
 
2b42f63
3f53269
 
2b42f63
 
 
3f53269
 
 
2b42f63
 
3f53269
2b42f63
3f53269
 
 
2b42f63
3f53269
 
 
2b42f63
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
import os
import time
import argparse
import jsonlines
from tqdm import tqdm
import pandas as pd
from transformers import AutoTokenizer
from vllm import LLM, SamplingParams


#Input all the arguments
parser = argparse.ArgumentParser()
parser.add_argument("--data_subset", type = str, default = "latency", help = "type of non-func requirement")
parser.add_argument("--temperature", type = float, default = 0.0, help = "temperature")
parser.add_argument("--max_new_tokens", type = int, default = 5192, help = "max length of tokens")
parser.add_argument("--top_p", type = float, default = 0.95, help = "top_p")
parser.add_argument("--prompt", type = str, default = "base_prompt", help = "type of prompt")
parser.add_argument("--num_samples", type = int, default = 1, help = "number of samples")
parser.add_argument("--model_path", type = str, required=True, help = "HF path for OS models")
parser.add_argument("--load_in_8bit", action="store_true", help = "Load model in 8bit")
parser.add_argument("--load_in_4bit", action="store_true", help = "Load model in 4bit")
parser.add_argument("--precision", type = str, default = "fp16", help = "Model precision, from: fp32, fp16 or bf16")
parser.add_argument("--tensor_parallel_size", type = int, default = 1, help = "Tensor parallel size")
parser.add_argument("--swap_space", type = int, default = 4, help = "The size (GiB) of CPU memory per GPU to use as swap space.")
parser.add_argument("--batch_size", type = int, default = 1, help = "Number of examples to send to llm engine at once.")
args = parser.parse_args()
argsdict = vars(args)


def model_query(all_messages, batch_size = 1):

    all_messages = [messages[0]["content"] for messages in all_messages]
    
    llm_tokenizer = AutoTokenizer.from_pretrained(
        args.model_path,
        truncation_side="left",
        padding_side="right",  # padding on the right is needed to cut off padding in `complete_code`
    )
    
    if args.num_samples == 1:
        GREEDY = True
    else:
        GREEDY = False
    
    assert args.num_samples % batch_size == 0, "num_samples must be divisible by batch_size"
    
    sampling_params = SamplingParams(
                                    n = batch_size, # for multisamples we sample multiple times
                                    temperature = args.temperature if not GREEDY else 0.0, 
                                    top_p = args.top_p if not GREEDY else 1.0,
                                    top_k = 50 if not GREEDY else -1,
                                    max_tokens = args.max_new_tokens,
                                    stop_token_ids = [llm_tokenizer.eos_token_id])
    
    llm = LLM(model = args.model_path,
            tensor_parallel_size = args.tensor_parallel_size,
            swap_space = args.swap_space)

    start_time = time.time()

    for turn_id in tqdm(range(0, args.num_samples//batch_size)):

        llm_outputs = llm.generate(all_messages, sampling_params)

        if turn_id == 0:
            all_generated_answers = [[llm_output.prompt + llm_gen.text 
                                    for llm_gen in llm_output.outputs]   
                                    for llm_output in llm_outputs]
        else:
            for idx, llm_output in enumerate(llm_outputs):
                all_generated_answers[idx].extend([llm_output.prompt + llm_gen.text 
                                                for llm_gen in llm_output.outputs])
                
    total_time = time.time() - start_time
    avg_times = [total_time / len(all_messages)] * len(all_messages)
    
    return all_generated_answers, avg_times


dataset_path = os.path.join("datasets", f"{args.data_subset}.jsonl")

data = []
max_tokens = []
generations = []
all_messages = []

with jsonlines.open(dataset_path) as data_file:

    for data_item in data_file:

        data.append(data_item)
        content = data_item[args.prompt]
        messages=[{"role": "user", "content": content}]
        all_messages.append(messages)

print("Starting model inference...")
all_generated_answers, all_inference_times = model_query(all_messages = all_messages, batch_size = args.batch_size)    

for i, data_item in tqdm(enumerate(data)):

    generated_answers = all_generated_answers[i]
    inference_time = all_inference_times[i]
    curr_sample = data_item
    curr_sample["inference_time"] = inference_time
    curr_sample["generated_answers"] = generated_answers

    for prompt in ["base_prompt", "coding_concepts", "chain_of_thought", "one_shot"]:
        del curr_sample[prompt]
    
    generations.append(curr_sample)   
    
generations = pd.DataFrame(generations)
path = os.path.join("generations", "edit", args.data_subset, os.path.split(args.model_path)[1], args.prompt, f"{args.num_samples}_samples")
if not os.path.exists(path):
    os.makedirs(path)
path=os.path.join(path, "generated_outputs.jsonl")
generations.to_json(path, orient = "records", lines=True)