diff --git "a/Ins/log.txt" "b/Ins/log.txt" new file mode 100644--- /dev/null +++ "b/Ins/log.txt" @@ -0,0 +1,2821 @@ +W0404 07:59:56.954000 140117560674112 torch/distributed/run.py:779] +W0404 07:59:56.954000 140117560674112 torch/distributed/run.py:779] ***************************************** +W0404 07:59:56.954000 140117560674112 torch/distributed/run.py:779] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. +W0404 07:59:56.954000 140117560674112 torch/distributed/run.py:779] ***************************************** +{'seed': 42, 'base_model': '/home/jovyan/workspace/Llama-7b', 'output_dir': './Ins', 'optim': 'adamw_torch', 'epochs': 1, 'learning_rate': 0.0005, 'per_device_batch_size': 8, 'gradient_accumulation_steps': 2, 'logging_step': 10, 'model_max_length': 1024, 'weight_decay': 0.01, 'lora_r': 8, 'lora_alpha': 32, 'lora_dropout': 0.05, 'lora_target_modules': 'q_proj,v_proj,k_proj,o_proj,gate_proj,down_proj,up_proj', 'lora_modules_to_save': 'embed_tokens,lm_head', 'resume_from_checkpoint': None, 'warmup_ratio': 0.01, 'lr_scheduler_type': 'cosine', 'save_and_eval_strategy': 'epoch', 'save_and_eval_steps': 1000, 'fp16': False, 'bf16': True, 'deepspeed': './config/ds_z2_bf16.json', 'remove_unused_columns': False, 'data_path': '/home/jovyan/workspace', 'tasks': 'seqrec,itemsearch,inters2title,inters2description,preferenceobtain,item2index,index2item,intertitles2item,query2item', 'train_data_sample_num': '0,0,0,0,0,0,0,0,0', 'dataset': 'Instruments', 'index_file': '.index.json', 'dataloader_num_workers': 4, 'dataloader_prefetch_factor': 2, 'max_his_len': 20, 'add_prefix': False, 'his_sep': ', ', 'only_train_response': True, 'train_prompt_sample_num': '1,1,1,1,1,1,1,1,1', 'valid_prompt_id': 0, 'sample_valid': True, 'valid_prompt_sample_num': 2, 'lr': 0.001, 'batch_size': 1024, 'num_workers': 4, 'eval_step': 50, 'learner': 'AdamW', 'dropout_prob': 0.0, 'bn': False, 'loss_type': 'mse', 'kmeans_init': False, 'kmeans_iters': 100, 'sk_epsilons': [0.0, 0.0, 0.0, 0.0], 'sk_iters': 50, 'device': 'cuda:1', 'num_emb_list': [256, 256, 256, 256], 'e_dim': 32, 'quant_loss_weight': 1.0, 'layers': [2048, 1024, 512, 256, 128, 64], 'ckpt_dir': '', 'warmup': 5, 'rqvae_model': '/home/jovyan/workspace/LC-Rec/index/Ins/Apr-04-2025_07-12-04/best_collision_model.pth'} +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message + Loading checkpoint shards: 0%| | 0/2 [00:00. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message + Loading checkpoint shards: 0%| | 0/2 [00:00