import os import sys import argparse import logging import time import json #os.environ["CUDA_VISIBLE_DEVICES"] = "1" from peft import PeftModel,LoraConfig, get_peft_model, prepare_model_for_kbit_training from transformers import AutoTokenizer, AutoModelForCausalLM, Trainer, TrainingArguments, BitsAndBytesConfig from datasets import load_dataset,Dataset import torch from utils.warp import Warp,WarpLJP from utils.dataset import DataCollatorForReward from utils.trainer import PRGTrainer from tree.base import Tree,Node,I_policy from configs.hyperparametric import Reward_config,Tree_config from model.logitsprocessor import OutputControlLogitsProcessor,RewardControlLogitsProcessor from tree.asts import AST from utils.model_generate import generate_string,generate_score tree_config = Tree_config().to_dict() #reward_config = Reward_config().to_dict() import torch #torch.cuda.set_device(0) #TASKS = ['ecthr_a','ecthr_b'] TASKS = ['ljp',] def get_args(): parser = argparse.ArgumentParser() ## ___datasets___ #parser.add_argument('--data_path',default='lex_glue',type=str, help='Path containing dataset') parser.add_argument('--train_path',default='',type=str, help='Path containing dataset') parser.add_argument('--eval_path',default='',type=str, help='Path containing dataset') parser.add_argument('--test_path',default='',type=str, help='Path containing dataset') parser.add_argument('--dataset',default='ljp',type=str, help='Dataset of choice in data_path') parser.add_argument('--save_data_path',default='',type=str, help='The path used to save dataset') parser.add_argument('--output_path',default='',type=str, help='The path used to save outputs') parser.add_argument('--sample_path',default='',type=str, help='The path used to samples') parser.add_argument('--control_file',default='./codekey_proofread.txt',type=str, help='The path used to output control') ## ___model___ parser.add_argument('--generate_model_path',default='',type=str, help='Path containing model') parser.add_argument('--reward_model_path',default='',type=str, help='Path containing model') parser.add_argument('--reward_save_path',default='./output/reward',type=str, help='Path containing model') parser.add_argument('--reward_lora_path',default='',type=str,) parser.add_argument('--per_device_train_batch_size',default=2,type=int) parser.add_argument('--gradient_accumulation_steps',default=2,type=int) parser.add_argument('--learning_rate',default=1e-3,type=float) parser.add_argument('--num_train_epochs',default=10,type=int) parser.add_argument('--logging_steps',default=200,type=int) parser.add_argument('--save_strategy',default='epoch',type=str,) parser.add_argument('--fp16',action='store_true',default=True,) parser.add_argument('--optim',default='paged_adamw_8bit',type=str,) parser.add_argument('--lora_rank',default=64,type=int) parser.add_argument('--lora_alpha',default=16,type=int) parser.add_argument('--lora_dropout',default=0.1,type=float) ## ___pipline___ parser.add_argument('--do_train',action='store_true',default=False, help='Training or not') parser.add_argument('--do_test',action='store_true',default=True, help='Eval or not') ## ___parameter___ parser.add_argument('--budget',default=20,type=int, help='iterations of search') parser.add_argument('--reward_funcation',default='leaf',type=str,choices=['random','reward','leaf'], help='iterations of search') parser.add_argument('--iteration',default=3,type=int, help='iterations of sample') ## ___special___ parser.add_argument('--ljp_mode',default='p',type=str,choices=['p','pd','pdf']) parser.add_argument('--logits_control',action='store_true',default=False, help='Training or not') parser.add_argument('--add_reward',action='store_true',default=False,) parser.add_argument('--inference_mode',default='zeroshot',type=str,choices=['zeroshot','fewshot','cot']) return parser.parse_args() def get_logger(path='./'): log_path = os.path.join(path,"log_%s.txt"%(time.strftime("%Y-%m-%d-%H-%M-%S",time.localtime()))) logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", handlers=[logging.StreamHandler(sys.stdout), logging.FileHandler(log_path)], ) logger = logging.getLogger(__name__) logger.setLevel(level=logging.DEBUG) return logger def load_data(args): if os.path.isdir(args.train_path): save_path = os.path.join(args.save_data_path,args.dataset) if not os.path.exists(save_path): dataset = load_dataset(path=args.train_path,name=args.dataset) dataset.save_to_disk(save_path) else: dataset = load_dataset(save_path) if os.path.isfile(args.train_path): data_files = {mode:path for mode,path in zip(['train','validation','test'],[args.train_path,args.eval_path,args.test_path]) if path} dataset = load_dataset('json',data_files=data_files) return dataset def load_samples(sample_path,): path_list = os.listdir(sample_path) samples = [] for path in path_list: path = os.path.join(sample_path,path) with open(path,'r') as f: for l in f.readlines(): sample = json.loads(l) samples.append(sample) return samples def train(args,warp,dataset,): # init TrainingArgument training_args = TrainingArguments( output_dir=os.path.join(args.reward_save_path,'reward_%s'%(time.strftime("%Y-%m-%d-%H-%M-%S",time.localtime()))), per_device_train_batch_size=args.per_device_train_batch_size, gradient_accumulation_steps=args.gradient_accumulation_steps, learning_rate=args.learning_rate, num_train_epochs=args.num_train_epochs, logging_steps=args.logging_steps, save_strategy=args.save_strategy, fp16=args.fp16, optim=args.optim, remove_unused_columns=False ) bnb_config = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type="nf4", llm_int8_threshold=6.0, llm_int8_has_fp16_weight=False, ) peft_config = LoraConfig( r=args.lora_rank, lora_alpha=args.lora_alpha, lora_dropout=args.lora_dropout, target_modules=["q_proj", "k_proj", "v_proj", "o_proj"], bias="none", task_type="CAUSAL_LM" ) # init model if not warp.reward_model: warp.load_reward_model(bnb_config=bnb_config) model = warp.reward_model tokenizer = warp.reward_tokenizer model = prepare_model_for_kbit_training(model) model = get_peft_model(model, peft_config) # init collator collator = DataCollatorForReward(tokenizer=tokenizer) logits_processor = RewardControlLogitsProcessor(tokenizer=tokenizer) # init dataset trainset = dataset['0'] # init trainer trainer = PRGTrainer( tokenizer=tokenizer, model=model, args=training_args, train_dataset=trainset, data_collator=collator, logits_processor=logits_processor ) # clean memory warp.generate_model = None torch.cuda.empty_cache() # training logger.info('start training..') for i,trainset in dataset.items(): if i != '0': trainer.train_dataset = trainset trainer.train() logger.info('training end..') warp.reward_model = model warp.reward_tokenizer = tokenizer def evaluate(args,warp,dataset,): if warp.generate_model == None: warp.load_generate_model() if args.add_reward: if args.reward_lora_path != '': bnb_config = BitsAndBytesConfig(load_in_4bit=True, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type="nf4", llm_int8_threshold=6.0, llm_int8_has_fp16_weight=False, ) warp.load_reward_model(bnb_config=bnb_config) warp.reward_model = PeftModel.from_pretrained(warp.reward_model, args.reward_lora_path) else: warp.load_reward_model() rewarder = {'model':warp.reward_model,'tokenizer':warp.reward_tokenizer,} if args.logits_control: rewarder['reward_processor'] = RewardControlLogitsProcessor(tokenizer=rewarder['tokenizer']) else: rewarder = {} model = warp.generate_model tokenizer = warp.generate_tokenizer def get_response(x,a,tokenizer,model,rewarder={}): if rewarder == {}: inputs = warp.prompt_to_crime(x,a,bos=tokenizer.bos_token,eos=tokenizer.eos_token) if args.logits_control: outputs = generate_string(inputs,tokenizer=tokenizer,model=model,logits_processor=warp.logits_processor) else: outputs = generate_string(inputs,tokenizer=tokenizer,model=model,) response = warp.step_from_response(outputs) return response if 'reward_processor' not in rewarder.keys(): rewarder['reward_processor'] = None for i in range(tree_config['branch']): inputs = warp.prompt_to_crime(x,a,bos=tokenizer.bos_token,eos=tokenizer.eos_token) if args.logits_control: outputs = generate_string(inputs,tokenizer=tokenizer,model=model,logits_processor=warp.logits_processor) else: outputs = generate_string(inputs,tokenizer=tokenizer,model=model,) response = warp.step_from_response(outputs) thought = warp.prompt_to_value(x,a+response,bos=rewarder['tokenizer'].bos_token,eos=rewarder['tokenizer'].eos_token) reward = generate_score(thought, tokenizer=rewarder['tokenizer'], model=rewarder['model'], reward_processor=rewarder['reward_processor'] ) r = warp.value_from_response(reward) if '拒绝' in r: continue else: break return response logger.info('start eval..') time_start = time.time() reward_control = 'rewardcontrol' if args.add_reward else 'un-reward' task_name = args.test_path.split('/')[-2] save_path = os.path.join(args.output_path,'eval',"%s_%s_%s_%s.json"%(task_name,args.inference_mode,reward_control,time.strftime("%Y-%m-%d-%H-%M-%S",time.localtime()))) preds = [] for i,data in enumerate(dataset): x,a,y = warp.processing_single(data) if args.inference_mode == 'fewshot': x = '这是一个例子:根据案情描述和已有步骤仅给出一个推理。如果是结论则直接输出,例如盗窃罪。如果是步骤则直接输出

,例如

步骤1:…

\n案情描述:2013年下半年至2015年10月26日,被告人张和菊利用担任山东泰开电力建设工程有限公司、山东泰开国际工程技术有限公司现金出纳的职务便利,多次将公司的资金共计4472572.91元挪出,用于其在深圳石油化工交易所、天津渤海商品交易所的投资交易,已全部亏损。2015年10月26日,张和菊从公司提取现金26万元后,携款潜逃至济南市长清区租房处藏匿。2015年11月6日,张和菊被公安机关抓获。\n已有推理步骤:\n挪用资金罪\n这是问题:\n根据案情描述和已有步骤仅给出一个推理。如果是结论则直接输出,例如盗窃罪。如果是步骤则直接输出

,例如

步骤1:…

\n案情描述:'+x elif args.inference_mode == 'cot': x = '一步步思考并回答,' + x y = y['crime'] a = '' for d in range(tree_config['max_depth']): try: response = get_response(x,a,tokenizer,model,rewarder=rewarder) except Exception as E: response = '' if '' in response: break if '

' in response: a += response y_ = response preds.append({'x':x,'y':y,'pred':y_}) if i % args.logging_steps == 0: logger.info('{x}'.format(x=str({'x':x,'y':y,'pred':y_}))) logger.info('eval: save...') with open(save_path,'w') as file: for l in preds: line = json.dumps(l,ensure_ascii=False) file.write(line) file.write('\n') time_end = time.time() logger.info('reward_eval : {x} '.format(x=args.reward_model_path)) logger.info('save_eval : {x} '.format(x=save_path)) logger.info('running time: {x}'.format(x=time_end-time_start)) logger.info('eval: fin') def sample(args,warp): # load datset logger.info('load datset..') dataset = load_data(args) logger.info('start sampling..') samples = {} for iter in range(args.iteration): logger.info('iter_{x}'.format(x=iter)) train_samples = [] sample_path = 'branch{b}_deep{d}_budget{g}_iter{i}'.format(b=tree_config['branch'],d=tree_config['max_depth'],g=args.budget,i=iter) if args.sample_path != '' and sample_path in os.listdir(args.sample_path): sample_path = os.path.join(args.sample_path,sample_path) train_samples += load_samples(sample_path) else: save_path = os.path.join(args.output_path,'data', 'branch{b}_deep{d}_budget{g}_iter{i}'.format(b=tree_config['branch'],d=tree_config['max_depth'],g=args.budget,i=iter), ) os.makedirs(save_path) if not os.path.exists(save_path) else None for i,sample in enumerate(dataset['train']): time_start = time.time() tree_of_sample = Tree(sample=sample,warp=warp) tree_of_sample.monte_carlo_tree_search(budget=args.budget,reward_funcation=args.reward_funcation) #train_samples += tree_of_sample.sample(attribute='positive') save_path = os.path.join(args.output_path,'data', 'branch{b}_deep{d}_budget{g}_iter{i}'.format(b=tree_config['branch'],d=tree_config['max_depth'],g=args.budget,i=iter), 'samples' + time.strftime("-%Y-%m-%d-%H:%M:%S", time.localtime()) + '.json') train_samples += tree_of_sample.save(path=save_path) time_end = time.time() logger.info('\nrunning time: {x}'.format(x=time_end-time_start)) _example = tree_of_sample.root.x[:50] if len(tree_of_sample.root.x) > 50 else tree_of_sample.root.x logger.info('{i}-th sample: {x}'.format(i=i,x=_example)) train_samples = Dataset.from_list(train_samples) samples[str(iter)] = train_samples return samples def run(args): # create framework if args.dataset in ['ljp']: warp = WarpLJP(args=args) warp.load_generate_model() # load training data # # data collection if args.do_train: trainsets = sample(args=args,warp=warp) #raise ValueError train(args,warp,trainsets) # test if args.do_test: datasets = load_data(args) evaluate(args,warp,datasets['test']) if __name__ == "__main__": args = get_args() logger = get_logger(args.output_path) loginfo = '\n'.join(['{k}: {v}'.format(k=k,v=v) for k,v in vars(args).items()]) logger.info(loginfo) try: run(args) except Exception as E: logger.exception('{x}'.format(x=E))