|
|
import os |
|
|
import torch |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer, Trainer, TrainingArguments, DataCollatorForLanguageModeling |
|
|
import datasets |
|
|
from datasets import load_dataset |
|
|
from datasets import Dataset, DatasetDict |
|
|
from peft import get_peft_model, LoraConfig, TaskType |
|
|
import json |
|
|
from tqdm import tqdm |
|
|
import pandas as pd |
|
|
from functools import partial |
|
|
import argparse |
|
|
import re |
|
|
|
|
|
import matplotlib |
|
|
matplotlib.use('Agg') |
|
|
import matplotlib.pyplot as plt |
|
|
|
|
|
parser = argparse.ArgumentParser() |
|
|
parser.add_argument('--model', type=str, default='160m',help='model name') |
|
|
parser.add_argument('--epoch', type=int, default=3,help='model name') |
|
|
parser.add_argument('--subname', type=str, default='arxiv',help='model name') |
|
|
parser.add_argument('--size', type=int, default=600 ,help='model name') |
|
|
parser.add_argument('--lr', type=float, default=2e-5, help='learning rate') |
|
|
parser.add_argument('--temp', type=float, default=0.0, help='generation temperature') |
|
|
parser.add_argument('--topp', type=float, default=1.0, help='generation top_p') |
|
|
parser.add_argument('--candidate', type=str, default='member', help='learning rate') |
|
|
args = parser.parse_args() |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
os.environ["WANDB_DISABLED"] = "true" |
|
|
|
|
|
|
|
|
|
|
|
loss_file = f'/workspace/{args.subname}_dataset/output_ft_more_layers_{args.subname}_epoch_{args.epoch}_mlp/pythia-{args.model}-{args.candidate}-{args.model}-epoch-{args.epoch}-pile-full-{args.size}-subsets-{args.subname}-{args.lr}/checkpoint-675/trainer_state.json' |
|
|
|
|
|
loss_datafile = json.load(open(loss_file))['log_history'] |
|
|
|
|
|
loss_l = [] |
|
|
|
|
|
for i in range(len(loss_datafile)): |
|
|
try: |
|
|
loss_data = loss_datafile[i]['loss'] |
|
|
loss_l.append(loss_data) |
|
|
except: |
|
|
continue |
|
|
|
|
|
model_name = f'pythia-{args.model}' |
|
|
|
|
|
model_name_hf_ori = f"/workspace/{model_name}" |
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name_hf_ori) |
|
|
tokenizer.padding_side = "left" |
|
|
|
|
|
if tokenizer.pad_token is None: |
|
|
tokenizer.pad_token = tokenizer.eos_token |
|
|
tokenizer.pad_token_id = tokenizer.eos_token_id |
|
|
data_files = f"/workspace/dataset_inference/{args.subname}_train.jsonl" |
|
|
raw_train_data_df = pd.read_json(data_files, lines=True) |
|
|
|
|
|
|
|
|
val_data_files = f"/workspace/dataset_inference/{args.subname}_val.jsonl" |
|
|
raw_val_data_df = pd.read_json(val_data_files, lines=True) |
|
|
|
|
|
tds=Dataset.from_pandas(raw_train_data_df) |
|
|
vds=Dataset.from_pandas(raw_val_data_df) |
|
|
|
|
|
raw_data = DatasetDict() |
|
|
|
|
|
raw_data['train'] = tds |
|
|
raw_data['validation'] = vds |
|
|
|
|
|
|
|
|
|
|
|
def tokenize_function(examples,max_length=384): |
|
|
tokens = tokenizer(examples["text"], padding="max_length", truncation=True, max_length=max_length) |
|
|
|
|
|
return tokens |
|
|
|
|
|
data_num = 1000 |
|
|
A_members = raw_data['train'].shuffle(seed=42).select(range(0, args.size)).map(partial(tokenize_function,max_length=512), batched=True, remove_columns=["text"]) |
|
|
A_nonmembers = raw_data['validation'].shuffle(seed=42).select(range(0, args.size)).map(partial(tokenize_function,max_length=512), batched=True, remove_columns=["text"]) |
|
|
|
|
|
B_members = raw_data['train'].shuffle(seed=42).select(range(data_num, data_num*2)).map(tokenize_function, batched=True, remove_columns=["text"]) |
|
|
B_nonmembers = raw_data['validation'].shuffle(seed=42).select(range(data_num, data_num*2)).map(tokenize_function, batched=True, remove_columns=["text"]) |
|
|
|
|
|
def get_num_from_directory(directory_path): |
|
|
|
|
|
|
|
|
|
|
|
numbers = [] |
|
|
|
|
|
|
|
|
for filename in os.listdir(directory_path): |
|
|
|
|
|
match = re.search(r'checkpoint-(\d+)', filename) |
|
|
if match: |
|
|
|
|
|
numbers.append(int(match.group(1))) |
|
|
|
|
|
return numbers |
|
|
|
|
|
def load_jsonl(file_path): |
|
|
data = [] |
|
|
with open(file_path, 'r') as file: |
|
|
for line in file: |
|
|
data.append(json.loads(line.strip())) |
|
|
return data |
|
|
|
|
|
def dump_jsonl(data, file_path): |
|
|
with open(file_path, 'w') as file: |
|
|
for item in data: |
|
|
json.dump(item, file) |
|
|
file.write('\n') |
|
|
|
|
|
def generate_responses(model,ds): |
|
|
response_list = [] |
|
|
for item in tqdm(ds): |
|
|
input_ids = torch.tensor(item['input_ids']).reshape(1,-1).to(model.device) |
|
|
input_len = input_ids.shape[1] |
|
|
pred = model.generate(input_ids, max_new_tokens=100) |
|
|
input_text = tokenizer.decode(pred[0][:input_len], skip_special_tokens=True) |
|
|
output_text = tokenizer.decode(pred[0][input_len:], skip_special_tokens=True) |
|
|
response_list.append({'output_text':output_text,'input_text':input_text}) |
|
|
return response_list |
|
|
|
|
|
def generate_responses(model,ds,temperature,top_p): |
|
|
model.eval() |
|
|
|
|
|
|
|
|
inputs = torch.tensor([item['input_ids'] for item in ds]).to(model.device) |
|
|
masks = torch.tensor([item['attention_mask'] for item in ds]).to(model.device) |
|
|
num_input,input_len = inputs.shape |
|
|
input_text = [] |
|
|
output_text = [] |
|
|
bs = 10 |
|
|
for i in tqdm(range(0,num_input,bs)): |
|
|
pred = model.generate(inputs=inputs[i:i+bs], attention_mask=masks[i:i+bs],max_new_tokens=100, temperature=temperature, top_p=top_p).detach() |
|
|
input_text += tokenizer.batch_decode(pred[:,:input_len], skip_special_tokens=True) |
|
|
output_text += tokenizer.batch_decode(pred[:,input_len:], skip_special_tokens=True) |
|
|
|
|
|
return [{'output_text':a,'input_text':b} for a,b in zip(output_text,input_text)] |
|
|
|
|
|
def run(train_dataset,eval_dataset,log_str, loss_l, args): |
|
|
directory_path = f"/workspace/{args.subname}_dataset/output_ft_more_layers_{args.subname}_epoch_{args.epoch}_mlp/pythia-{args.model}-{args.candidate}-{args.model}-epoch-{args.epoch}-pile-full-{args.size}-subsets-{args.subname}-{args.lr}" |
|
|
numbers = get_num_from_directory(directory_path) |
|
|
min_loss_index = loss_l.index(min(loss_l)) |
|
|
os.makedirs(f'responses_ft_more_layers_{args.subname}_epoch_{args.epoch}_mlp/all_checkpoint', exist_ok=True) |
|
|
for num in numbers: |
|
|
model_name_hf = f"/workspace/{args.subname}_dataset/output_ft_more_layers_{args.subname}_epoch_{args.epoch}_mlp/pythia-{args.model}-{args.candidate}-{args.model}-epoch-{args.epoch}-pile-full-{args.size}-subsets-{args.subname}-{args.lr}/checkpoint-{num}" |
|
|
if torch.cuda.is_available(): |
|
|
device = torch.device("cuda") |
|
|
model = AutoModelForCausalLM.from_pretrained(model_name_hf).to(device) |
|
|
|
|
|
|
|
|
model.eval() |
|
|
response_list = generate_responses(model,eval_dataset, args.temp, args.topp) |
|
|
if num == numbers[min_loss_index]: |
|
|
dump_jsonl(response_list,f'responses_ft_more_layers_{args.subname}_epoch_{args.epoch}_mlp/all_checkpoint/{model_name}-{log_str}-{num}-ft.jsonl') |
|
|
else: |
|
|
dump_jsonl(response_list,f'responses_ft_more_layers_{args.subname}_epoch_{args.epoch}_mlp/all_checkpoint/{model_name}-{log_str}-{num}-ft.jsonl') |
|
|
|
|
|
|
|
|
run(A_members,B_members,f'member-{args.model}-epoch-{args.epoch}-pile-full-{args.size}-subsets-{args.subname}-{args.lr}', loss_l, args) |
|
|
run(A_nonmembers,B_nonmembers,f'nonmember-{args.model}-epoch-{args.epoch}-pile-full-{args.size}-subsets-{args.subname}-{args.lr}', loss_l, args) |