tnnanh1005 commited on
Commit
c35ad83
·
verified ·
1 Parent(s): 04c8d77

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +18 -0
  2. data/thinking_dataset_merged.jsonl +3 -0
  3. data/thinking_dataset_v2_swift.jsonl +3 -0
  4. data/wandb/debug-internal.log +12 -0
  5. data/wandb/debug.log +22 -0
  6. data/wandb/run-20260304_035219-bzc9z48s/files/config.yaml +632 -0
  7. data/wandb/run-20260304_035219-bzc9z48s/files/output.log +0 -0
  8. data/wandb/run-20260304_035219-bzc9z48s/files/requirements.txt +159 -0
  9. data/wandb/run-20260304_035219-bzc9z48s/files/wandb-metadata.json +110 -0
  10. data/wandb/run-20260304_035219-bzc9z48s/files/wandb-summary.json +1 -0
  11. data/wandb/run-20260304_035219-bzc9z48s/logs/debug-core.log +15 -0
  12. data/wandb/run-20260304_035219-bzc9z48s/logs/debug-internal.log +14 -0
  13. data/wandb/run-20260304_035219-bzc9z48s/logs/debug.log +24 -0
  14. data/wandb/run-20260304_035219-bzc9z48s/run-bzc9z48s.wandb +3 -0
  15. data/wandb/run-20260306_073928-dpxzw953/files/output.log +32 -0
  16. data/wandb/run-20260306_073928-dpxzw953/files/requirements.txt +241 -0
  17. data/wandb/run-20260306_073928-dpxzw953/files/wandb-metadata.json +97 -0
  18. data/wandb/run-20260306_073928-dpxzw953/logs/debug-core.log +8 -0
  19. data/wandb/run-20260306_073928-dpxzw953/logs/debug-internal.log +6 -0
  20. data/wandb/run-20260306_073928-dpxzw953/logs/debug.log +22 -0
  21. data/wandb/run-20260306_073928-dpxzw953/run-dpxzw953.wandb +3 -0
  22. data/wandb/run-20260306_105746-8846dbl6/files/output.log +0 -0
  23. data/wandb/run-20260306_105746-8846dbl6/files/requirements.txt +243 -0
  24. data/wandb/run-20260306_105746-8846dbl6/files/wandb-metadata.json +97 -0
  25. data/wandb/run-20260306_105746-8846dbl6/logs/debug-core.log +8 -0
  26. data/wandb/run-20260306_105746-8846dbl6/logs/debug-internal.log +12 -0
  27. data/wandb/run-20260306_105746-8846dbl6/logs/debug.log +22 -0
  28. data/wandb/run-20260306_105746-8846dbl6/run-8846dbl6.wandb +3 -0
  29. output/Qwen3-1.7B/v0-20260304-031440/args.json +346 -0
  30. output/Qwen3-1.7B/v0-20260304-031440/images/train_epoch.png +0 -0
  31. output/Qwen3-1.7B/v0-20260304-031440/images/train_grad_norm.png +0 -0
  32. output/Qwen3-1.7B/v0-20260304-031440/images/train_learning_rate.png +0 -0
  33. output/Qwen3-1.7B/v0-20260304-031440/images/train_loss.png +0 -0
  34. output/Qwen3-1.7B/v0-20260304-031440/images/train_token_acc.png +0 -0
  35. output/Qwen3-1.7B/v0-20260304-031440/logging.jsonl +17 -0
  36. output/Qwen3-1.7B/v0-20260304-031440/runs/Mar04_03-19-12_ubuntu/events.out.tfevents.1772594352.ubuntu.2490446.0 +3 -0
  37. output/proposals_RM_gpt_oss_120b_ensemble/polaris/polaris-adme-fang-hclint-1_scored_16.json +0 -0
  38. output/proposals_RM_gpt_oss_120b_ensemble/polaris/polaris-adme-fang-hppb-1_scored_16.json +0 -0
  39. output/proposals_RM_gpt_oss_120b_ensemble/polaris/polaris-adme-fang-solu-1_scored_16.json +0 -0
  40. output/proposals_RM_gpt_oss_120b_ensemble/polaris/tdcommons-bbb-martins_scored_16.json +0 -0
  41. output/proposals_RM_gpt_oss_120b_ensemble/polaris/tdcommons-cyp2d6-substrate-carbonmangels_scored_16.json +0 -0
  42. output/proposals_RM_gpt_oss_120b_ensemble/polaris/tdcommons-herg_scored_16.json +0 -0
  43. output/proposals_RM_gpt_oss_120b_ensemble/polaris/tdcommons-lipophilicity-astrazeneca_scored_16.json +0 -0
  44. output/proposals_RM_gpt_oss_20b_ensemble/polaris/polaris-adme-fang-hclint-1_scored_16.json +0 -0
  45. output/proposals_RM_gpt_oss_20b_ensemble/polaris/polaris-adme-fang-hppb-1_scored_16.json +0 -0
  46. output/proposals_RM_gpt_oss_20b_ensemble/polaris/polaris-adme-fang-solu-1_scored_16.json +0 -0
  47. output/proposals_RM_gpt_oss_20b_ensemble/polaris/tdcommons-bbb-martins_scored_16.json +0 -0
  48. output/proposals_RM_gpt_oss_20b_ensemble/polaris/tdcommons-cyp2d6-substrate-carbonmangels_scored_16.json +0 -0
  49. output/proposals_RM_gpt_oss_20b_ensemble/polaris/tdcommons-herg_scored_16.json +0 -0
  50. output/proposals_RM_gpt_oss_20b_ensemble/polaris/tdcommons-lipophilicity-astrazeneca_scored_16.json +0 -0
.gitattributes CHANGED
@@ -33,3 +33,21 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ data/thinking_dataset_merged.jsonl filter=lfs diff=lfs merge=lfs -text
37
+ data/thinking_dataset_v2_swift.jsonl filter=lfs diff=lfs merge=lfs -text
38
+ data/wandb/run-20260304_035219-bzc9z48s/run-bzc9z48s.wandb filter=lfs diff=lfs merge=lfs -text
39
+ data/wandb/run-20260306_073928-dpxzw953/run-dpxzw953.wandb filter=lfs diff=lfs merge=lfs -text
40
+ data/wandb/run-20260306_105746-8846dbl6/run-8846dbl6.wandb filter=lfs diff=lfs merge=lfs -text
41
+ output/reward/ensemble_comparison_gpt_oss_120b_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
42
+ output/reward/ensemble_comparison_gpt_oss_20b_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
43
+ output/reward/ensemble_comparison_qwen3_1,7b_OG_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
44
+ output/reward/ensemble_comparison_qwen3_1,7b_sft_thinking_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
45
+ output/reward/ensemble_convergence_gpt_oss_120b_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
46
+ output/reward/ensemble_convergence_gpt_oss_20b_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
47
+ output/reward/ensemble_convergence_qwen3_1,7b_OG_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
48
+ output/reward/ensemble_convergence_qwen3_1,7b_sft_thinking_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
49
+ output/reward/ensemble_separate_gpt_oss_120b_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
50
+ output/reward/ensemble_separate_gpt_oss_20b_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
51
+ output/reward/ensemble_separate_qwen3_1,7b_OG_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
52
+ output/reward/ensemble_separate_qwen3_1,7b_sft_thinking_ensemble_temp0,8.png filter=lfs diff=lfs merge=lfs -text
53
+ train/output/v0-20260304-035120/checkpoint-4374-merged/tokenizer.json filter=lfs diff=lfs merge=lfs -text
data/thinking_dataset_merged.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc3de53a26565bfce6ea455e6dd280984c40f68e79e1a0acefeefd1f2259897e
3
+ size 337772135
data/thinking_dataset_v2_swift.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cc9cf7992b7f889abba2bf9ac64e1ff44d1dc2b5f498c2208e231c7d7cba59a
3
+ size 185136875
data/wandb/debug-internal.log ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-06T10:57:46.486394548Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-03-06T10:57:46.705840771Z","level":"INFO","msg":"stream: created new stream","id":"8846dbl6"}
3
+ {"time":"2026-03-06T10:57:46.706184448Z","level":"INFO","msg":"handler: started","stream_id":"8846dbl6"}
4
+ {"time":"2026-03-06T10:57:46.706376796Z","level":"INFO","msg":"stream: started","id":"8846dbl6"}
5
+ {"time":"2026-03-06T10:57:46.70647031Z","level":"INFO","msg":"writer: started","stream_id":"8846dbl6"}
6
+ {"time":"2026-03-06T10:57:46.706495415Z","level":"INFO","msg":"sender: started","stream_id":"8846dbl6"}
7
+ {"time":"2026-03-06T19:23:17.154701842Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/nhatanh10102005-vinuniversity/ms-swift/8846dbl6/file_stream","body":"\n<html><head>\n<meta http-equiv=\"content-type\" content=\"text/html;charset=utf-8\">\n<title>502 Server Error</title>\n</head>\n<body text=#000000 bgcolor=#ffffff>\n<h1>Error: Server Error</h1>\n<h2>The server encountered a temporary error and could not complete your request.<p>Please try again in 30 seconds.</h2>\n<h2></h2>\n</body></html>\n"}
8
+ {"time":"2026-03-07T01:23:02.15976294Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/nhatanh10102005-vinuniversity/ms-swift/8846dbl6/file_stream","body":"\n<html><head>\n<meta http-equiv=\"content-type\" content=\"text/html;charset=utf-8\">\n<title>502 Server Error</title>\n</head>\n<body text=#000000 bgcolor=#ffffff>\n<h1>Error: Server Error</h1>\n<h2>The server encountered a temporary error and could not complete your request.<p>Please try again in 30 seconds.</h2>\n<h2></h2>\n</body></html>\n"}
9
+ {"time":"2026-03-07T02:55:05.824285344Z","level":"INFO","msg":"flowcontrol: backed up, offloading to disk","recordNumber":15580}
10
+ {"time":"2026-03-07T02:55:05.824494952Z","level":"INFO","msg":"flowcontrol: unblocked","totalOffloaded":1}
11
+ {"time":"2026-03-07T02:55:05.825190016Z","level":"INFO","msg":"flowcontrol: backed up, offloading to disk","recordNumber":15613}
12
+ {"time":"2026-03-07T02:55:05.840550203Z","level":"INFO","msg":"flowcontrol: unblocked","totalOffloaded":904}
data/wandb/debug.log ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-06 10:57:46,235 INFO MainThread:1654770 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-03-06 10:57:46,236 INFO MainThread:1654770 [wandb_setup.py:_flush():81] Configure stats pid to 1654770
3
+ 2026-03-06 10:57:46,236 INFO MainThread:1654770 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-03-06 10:57:46,236 INFO MainThread:1654770 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /root/code/na/Idea-filtering-training/data/wandb/run-20260306_105746-8846dbl6/logs/debug.log
5
+ 2026-03-06 10:57:46,236 INFO MainThread:1654770 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /root/code/na/Idea-filtering-training/data/wandb/run-20260306_105746-8846dbl6/logs/debug-internal.log
6
+ 2026-03-06 10:57:46,236 INFO MainThread:1654770 [wandb_init.py:init():844] calling init triggers
7
+ 2026-03-06 10:57:46,237 INFO MainThread:1654770 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'_wandb': {}}
9
+ 2026-03-06 10:57:46,237 INFO MainThread:1654770 [wandb_init.py:init():892] starting backend
10
+ 2026-03-06 10:57:46,468 INFO MainThread:1654770 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-03-06 10:57:46,475 INFO MainThread:1654770 [wandb_init.py:init():903] backend started and connected
12
+ 2026-03-06 10:57:46,480 INFO MainThread:1654770 [wandb_init.py:init():973] updated telemetry
13
+ 2026-03-06 10:57:46,491 INFO MainThread:1654770 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-03-06 10:57:47,094 INFO MainThread:1654770 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-03-06 10:57:47,290 INFO MainThread:1654770 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-03-06 10:57:47,290 INFO MainThread:1654770 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-03-06 10:57:47,290 INFO MainThread:1654770 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-03-06 10:57:47,290 INFO MainThread:1654770 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-03-06 10:57:47,295 INFO MainThread:1654770 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-03-06 10:57:47,296 INFO MainThread:1654770 [wandb_run.py:_config_callback():1403] config_cb None None {'peft_config': {'default': {'task_type': 'CAUSAL_LM', 'peft_type': 'LORA', 'auto_mapping': None, 'peft_version': '0.18.1', 'base_model_name_or_path': '/root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e', 'revision': None, 'inference_mode': False, 'r': 32, 'target_modules': ['gate_proj', 'up_proj', 'v_proj,o_proj', 'down_proj', 'k_proj', 'q_proj'], 'exclude_modules': None, 'lora_alpha': 64, 'lora_dropout': 0.05, 'fan_in_fan_out': False, 'bias': 'none', 'use_rslora': False, 'modules_to_save': [], 'init_lora_weights': True, 'layers_to_transform': None, 'layers_pattern': None, 'rank_pattern': {}, 'alpha_pattern': {}, 'megatron_config': None, 'megatron_core': 'megatron.core', 'trainable_token_indices': None, 'loftq_config': {}, 'eva_config': None, 'corda_config': None, 'use_dora': False, 'alora_invocation_tokens': None, 'use_qalora': False, 'qalora_group_size': 16, 'layer_replication': None, 'runtime_config': {'ephemeral_gpu_offload': False}, 'lora_bias': False, 'target_parameters': None, 'arrow_config': None, 'ensure_weight_tying': False, 'lora_dtype': None, 'lorap_lr_ratio': None, 'lorap_emb_lr': 1e-06}}, 'vocab_size': 151936, 'max_position_embeddings': 40960, 'hidden_size': 2048, 'intermediate_size': 6144, 'num_hidden_layers': 28, 'num_attention_heads': 16, 'use_sliding_window': False, 'sliding_window': None, 'max_window_layers': 28, 'num_key_value_heads': 8, 'head_dim': 128, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-06, 'use_cache': False, 'attention_bias': False, 'attention_dropout': 0.0, 'layer_types': ['full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention'], 'pad_token_id': 151643, 'bos_token_id': None, 'eos_token_id': 151645, 'tie_word_embeddings': True, 'rope_parameters': {'rope_theta': 1000000, 'rope_type': 'default'}, 'return_dict': True, 'output_hidden_states': False, 'dtype': 'bfloat16', 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'architectures': ['Qwen3ForCausalLM'], 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'problem_type': None, '_name_or_path': '/root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e', 'transformers_version': '5.3.0', 'model_type': 'qwen3', 'output_attentions': False, 'output_dir': '/root/code/na/Idea-filtering-training/train/output/v5-20260306-105729', 'per_device_train_batch_size': 1, 'num_train_epochs': 3.0, 'max_steps': -1, 'learning_rate': 0.0002, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': None, 'warmup_steps': 0.05, 'optim': 'adamw_torch_fused', 'optim_args': None, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.95, 'adam_epsilon': 1e-08, 'optim_target_modules': None, 'gradient_accumulation_steps': 16, 'average_tokens_across_devices': True, 'max_grad_norm': 1.0, 'label_smoothing_factor': 0.0, 'bf16': True, 'fp16': False, 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'use_liger_kernel': False, 'liger_kernel_config': None, 'neftune_noise_alpha': None, 'torch_empty_cache_steps': None, 'auto_find_batch_size': False, 'logging_strategy': 'steps', 'logging_steps': 5, 'logging_first_step': True, 'log_on_each_node': True, 'logging_nan_inf_filter': True, 'include_num_input_tokens_seen': 'no', 'log_level': 'passive', 'log_level_replica': 'warning', 'disable_tqdm': False, 'report_to': ['wandb'], 'run_name': '/root/code/na/Idea-filtering-training/train/output/v5-20260306-105729', 'project': 'huggingface', 'trackio_space_id': 'trackio', 'eval_strategy': 'no', 'eval_steps': 500.0, 'eval_delay': 0, 'per_device_eval_batch_size': 1, 'prediction_loss_only': False, 'eval_on_start': False, 'eval_do_concat_batches': True, 'eval_use_gather_object': False, 'eval_accumulation_steps': None, 'include_for_metrics': [], 'batch_eval_metrics': False, 'save_only_model': False, 'save_strategy': 'steps', 'save_steps': 1000, 'save_on_each_node': False, 'save_total_limit': None, 'enable_jit_checkpoint': False, 'push_to_hub': False, 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_always_push': False, 'hub_revision': None, 'load_best_model_at_end': False, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'restore_callback_states_from_checkpoint': False, 'full_determinism': False, 'seed': 42, 'data_seed': 42, 'use_cpu': False, 'accelerator_config': {'split_batches': False, 'dispatch_batches': False, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'parallelism_config': None, 'dataloader_drop_last': False, 'dataloader_num_workers': 1, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'dataloader_prefetch_factor': 2, 'remove_unused_columns': False, 'label_names': None, 'train_sampling_strategy': 'random', 'length_column_name': 'length', 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'ddp_backend': None, 'ddp_timeout': 18000000, 'fsdp': [], 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'deepspeed': None, 'debug': [], 'skip_memory_metrics': True, 'do_train': False, 'do_eval': False, 'do_predict': False, 'resume_from_checkpoint': None, 'warmup_ratio': 0.05, 'logging_dir': '/root/code/na/Idea-filtering-training/train/output/v5-20260306-105729/runs', 'local_rank': -1, 'sortish_sampler': False, 'predict_with_generate': False, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None, 'tuner_backend': 'peft', 'vit_gradient_checkpointing': True, 'router_aux_loss_coef': 0.0, 'enable_dft_loss': False, 'enable_channel_loss': False, 'check_model': True, 'acc_strategy': 'token', 'train_dataloader_shuffle': True, 'group_by_length': False, 'max_epochs': 3, 'aligner_lr': None, 'vit_lr': None, 'use_logits_to_keep': None, 'ds3_gather_for_generation': True, 'resume_only_model': False, 'optimizer': None, 'loss_type': None, 'eval_metric': None, 'callbacks': [], 'early_stop_interval': None, 'eval_use_evalscope': False, 'eval_dataset': [], 'eval_dataset_args': None, 'eval_limit': None, 'eval_generation_config': None, 'extra_eval_args': None, 'tuner_type': 'lora', 'use_galore': False, 'galore_target_modules': None, 'galore_rank': 128, 'galore_update_proj_gap': 50, 'galore_scale': 1.0, 'galore_proj_type': 'std', 'galore_optim_per_parameter': False, 'galore_with_embedding': False, 'galore_quantization': False, 'galore_proj_quant': False, 'galore_proj_bits': 4, 'galore_proj_group_size': 256, 'galore_cos_threshold': 0.4, 'galore_gamma_proj': 2, 'galore_queue_size': 5, 'lisa_activated_layers': 0, 'lisa_step_interval': 20, 'use_flash_ckpt': False}
21
+ 2026-03-06 10:57:47,308 INFO MainThread:1654770 [wandb_config.py:__setitem__():155] [no run ID] config set model/num_parameters = 2060182528 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7ffc51dc5100>>
22
+ 2026-03-06 10:57:47,308 INFO MainThread:1654770 [wandb_run.py:_config_callback():1403] config_cb model/num_parameters 2060182528 None
data/wandb/run-20260304_035219-bzc9z48s/files/config.yaml ADDED
@@ -0,0 +1,632 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _name_or_path:
2
+ value: /root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e
3
+ _wandb:
4
+ value:
5
+ cli_version: 0.25.0
6
+ e:
7
+ cu7bd4tvbr27kn8zb6sicf7uvdvg0f2r:
8
+ args:
9
+ - --model
10
+ - Qwen/Qwen3-1.7B
11
+ - --use_hf
12
+ - "true"
13
+ - --dataset
14
+ - /root/code/na/Idea-filtering-training/data/thinking_dataset_v2_swift.jsonl
15
+ - --train_type
16
+ - lora
17
+ - --lora_rank
18
+ - "32"
19
+ - --lora_alpha
20
+ - "64"
21
+ - --lora_dropout
22
+ - "0.05"
23
+ - --target_modules
24
+ - q_proj
25
+ - k_proj
26
+ - v_proj
27
+ - o_proj
28
+ - gate_proj
29
+ - up_proj
30
+ - down_proj
31
+ - --lora_ga_batch_size
32
+ - "1"
33
+ - --lora_ga_iters
34
+ - "2"
35
+ - --lora_ga_max_length
36
+ - "1024"
37
+ - --lora_ga_direction
38
+ - ArB2r
39
+ - --lora_ga_scale
40
+ - stable
41
+ - --lora_ga_stable_gamma
42
+ - "16"
43
+ - --per_device_train_batch_size
44
+ - "1"
45
+ - --gradient_accumulation_steps
46
+ - "16"
47
+ - --max_length
48
+ - "2600"
49
+ - --learning_rate
50
+ - "2e-4"
51
+ - --warmup_ratio
52
+ - "0.05"
53
+ - --max_epochs
54
+ - "3"
55
+ - --attn_impl
56
+ - eager
57
+ - --output_dir
58
+ - /root/code/na/Idea-filtering-training/train/output
59
+ - --save_strategy
60
+ - steps
61
+ - --save_steps
62
+ - "1000"
63
+ - --eval_steps
64
+ - "500"
65
+ - --dataloader_num_workers
66
+ - "1"
67
+ - --dataset_num_proc
68
+ - "16"
69
+ - --load_from_cache_file
70
+ - "true"
71
+ - --model_author
72
+ - swift
73
+ - --model_name
74
+ - swift-robot
75
+ - --loss_scale
76
+ - ignore_empty_think
77
+ - --report_to
78
+ - wandb
79
+ cpu_count: 49
80
+ cpu_count_logical: 49
81
+ cudaVersion: "13.0"
82
+ disk:
83
+ /:
84
+ total: "960312868864"
85
+ used: "865124937728"
86
+ email: nhatanh10102005@gmail.com
87
+ executable: /root/miniconda3/envs/RM/bin/python3
88
+ git:
89
+ commit: 0dbe0bbbe696b2e49ffefa83123579a73fab61c6
90
+ remote: https://tnnanh1010:@github.com/Drylab-AI/Idea-filtering-training.git
91
+ gpu: NVIDIA GeForce RTX 3060
92
+ gpu_count: 1
93
+ gpu_nvidia:
94
+ - architecture: Ampere
95
+ cudaCores: 3584
96
+ memoryTotal: "12884901888"
97
+ name: NVIDIA GeForce RTX 3060
98
+ uuid: GPU-f370a18a-d749-7bf1-c827-82726afbb6cf
99
+ host: ubuntu
100
+ memory:
101
+ total: "628090474496"
102
+ os: Linux-6.8.0-87-generic-x86_64-with-glibc2.35
103
+ program: /root/miniconda3/envs/RM/lib/python3.12/site-packages/swift/cli/sft.py
104
+ python: CPython 3.12.12
105
+ root: /root/code/na/Idea-filtering-training/data
106
+ startedAt: "2026-03-04T03:52:19.561578Z"
107
+ writerId: cu7bd4tvbr27kn8zb6sicf7uvdvg0f2r
108
+ m:
109
+ - "1": train/global_step
110
+ "6":
111
+ - 3
112
+ "7": []
113
+ - "2": '*'
114
+ "5": 1
115
+ "6":
116
+ - 1
117
+ "7": []
118
+ python_version: 3.12.12
119
+ t:
120
+ "1":
121
+ - 1
122
+ - 11
123
+ - 49
124
+ - 51
125
+ - 71
126
+ - 84
127
+ - 98
128
+ - 105
129
+ "2":
130
+ - 1
131
+ - 11
132
+ - 49
133
+ - 51
134
+ - 71
135
+ - 84
136
+ - 98
137
+ - 105
138
+ "3":
139
+ - 7
140
+ - 13
141
+ - 19
142
+ - 62
143
+ - 66
144
+ "4": 3.12.12
145
+ "5": 0.25.0
146
+ "6": 5.2.0
147
+ "9":
148
+ "1": transformers_trainer
149
+ "12": 0.25.0
150
+ "13": linux-x86_64
151
+ acc_strategy:
152
+ value: token
153
+ accelerator_config:
154
+ value:
155
+ dispatch_batches: false
156
+ even_batches: true
157
+ gradient_accumulation_kwargs: null
158
+ non_blocking: false
159
+ split_batches: false
160
+ use_seedable_sampler: true
161
+ adam_beta1:
162
+ value: 0.9
163
+ adam_beta2:
164
+ value: 0.95
165
+ adam_epsilon:
166
+ value: 1e-08
167
+ aligner_lr:
168
+ value: null
169
+ architectures:
170
+ value:
171
+ - Qwen3ForCausalLM
172
+ attention_bias:
173
+ value: false
174
+ attention_dropout:
175
+ value: 0
176
+ auto_find_batch_size:
177
+ value: false
178
+ average_tokens_across_devices:
179
+ value: true
180
+ batch_eval_metrics:
181
+ value: false
182
+ bf16:
183
+ value: true
184
+ bf16_full_eval:
185
+ value: false
186
+ bos_token_id:
187
+ value: null
188
+ callbacks:
189
+ value: []
190
+ check_model:
191
+ value: true
192
+ chunk_size_feed_forward:
193
+ value: 0
194
+ data_seed:
195
+ value: 42
196
+ dataloader_drop_last:
197
+ value: false
198
+ dataloader_num_workers:
199
+ value: 1
200
+ dataloader_persistent_workers:
201
+ value: false
202
+ dataloader_pin_memory:
203
+ value: true
204
+ dataloader_prefetch_factor:
205
+ value: 2
206
+ ddp_backend:
207
+ value: null
208
+ ddp_broadcast_buffers:
209
+ value: null
210
+ ddp_bucket_cap_mb:
211
+ value: null
212
+ ddp_find_unused_parameters:
213
+ value: null
214
+ ddp_timeout:
215
+ value: 18000000
216
+ debug:
217
+ value: []
218
+ deepspeed:
219
+ value: null
220
+ disable_tqdm:
221
+ value: false
222
+ do_eval:
223
+ value: false
224
+ do_predict:
225
+ value: false
226
+ do_train:
227
+ value: false
228
+ ds3_gather_for_generation:
229
+ value: true
230
+ dtype:
231
+ value: bfloat16
232
+ early_stop_interval:
233
+ value: null
234
+ enable_channel_loss:
235
+ value: false
236
+ enable_dft_loss:
237
+ value: false
238
+ enable_jit_checkpoint:
239
+ value: false
240
+ eos_token_id:
241
+ value: 151645
242
+ eval_accumulation_steps:
243
+ value: null
244
+ eval_dataset:
245
+ value: []
246
+ eval_dataset_args:
247
+ value: null
248
+ eval_delay:
249
+ value: 0
250
+ eval_do_concat_batches:
251
+ value: true
252
+ eval_generation_config:
253
+ value: null
254
+ eval_limit:
255
+ value: null
256
+ eval_metric:
257
+ value: null
258
+ eval_on_start:
259
+ value: false
260
+ eval_steps:
261
+ value: 500
262
+ eval_strategy:
263
+ value: "no"
264
+ eval_use_evalscope:
265
+ value: false
266
+ eval_use_gather_object:
267
+ value: false
268
+ extra_eval_args:
269
+ value: null
270
+ fp16:
271
+ value: false
272
+ fp16_full_eval:
273
+ value: false
274
+ fsdp:
275
+ value: []
276
+ fsdp_config:
277
+ value:
278
+ min_num_params: 0
279
+ xla: false
280
+ xla_fsdp_grad_ckpt: false
281
+ xla_fsdp_v2: false
282
+ full_determinism:
283
+ value: false
284
+ galore_cos_threshold:
285
+ value: 0.4
286
+ galore_gamma_proj:
287
+ value: 2
288
+ galore_optim_per_parameter:
289
+ value: false
290
+ galore_proj_bits:
291
+ value: 4
292
+ galore_proj_group_size:
293
+ value: 256
294
+ galore_proj_quant:
295
+ value: false
296
+ galore_proj_type:
297
+ value: std
298
+ galore_quantization:
299
+ value: false
300
+ galore_queue_size:
301
+ value: 5
302
+ galore_rank:
303
+ value: 128
304
+ galore_scale:
305
+ value: 1
306
+ galore_target_modules:
307
+ value: null
308
+ galore_update_proj_gap:
309
+ value: 50
310
+ galore_with_embedding:
311
+ value: false
312
+ generation_config:
313
+ value: null
314
+ generation_max_length:
315
+ value: null
316
+ generation_num_beams:
317
+ value: null
318
+ gradient_accumulation_steps:
319
+ value: 16
320
+ gradient_checkpointing:
321
+ value: false
322
+ gradient_checkpointing_kwargs:
323
+ value: null
324
+ greater_is_better:
325
+ value: false
326
+ group_by_length:
327
+ value: false
328
+ head_dim:
329
+ value: 128
330
+ hidden_act:
331
+ value: silu
332
+ hidden_size:
333
+ value: 2048
334
+ hub_always_push:
335
+ value: false
336
+ hub_model_id:
337
+ value: null
338
+ hub_private_repo:
339
+ value: null
340
+ hub_revision:
341
+ value: null
342
+ hub_strategy:
343
+ value: every_save
344
+ hub_token:
345
+ value: <HUB_TOKEN>
346
+ id2label:
347
+ value:
348
+ "0": LABEL_0
349
+ "1": LABEL_1
350
+ ignore_data_skip:
351
+ value: false
352
+ include_for_metrics:
353
+ value: []
354
+ include_num_input_tokens_seen:
355
+ value: "no"
356
+ initializer_range:
357
+ value: 0.02
358
+ intermediate_size:
359
+ value: 6144
360
+ is_encoder_decoder:
361
+ value: false
362
+ label_names:
363
+ value: null
364
+ label_smoothing_factor:
365
+ value: 0
366
+ label2id:
367
+ value:
368
+ LABEL_0: 0
369
+ LABEL_1: 1
370
+ layer_types:
371
+ value:
372
+ - full_attention
373
+ - full_attention
374
+ - full_attention
375
+ - full_attention
376
+ - full_attention
377
+ - full_attention
378
+ - full_attention
379
+ - full_attention
380
+ - full_attention
381
+ - full_attention
382
+ - full_attention
383
+ - full_attention
384
+ - full_attention
385
+ - full_attention
386
+ - full_attention
387
+ - full_attention
388
+ - full_attention
389
+ - full_attention
390
+ - full_attention
391
+ - full_attention
392
+ - full_attention
393
+ - full_attention
394
+ - full_attention
395
+ - full_attention
396
+ - full_attention
397
+ - full_attention
398
+ - full_attention
399
+ - full_attention
400
+ learning_rate:
401
+ value: 0.0002
402
+ length_column_name:
403
+ value: length
404
+ liger_kernel_config:
405
+ value: null
406
+ lisa_activated_layers:
407
+ value: 0
408
+ lisa_step_interval:
409
+ value: 20
410
+ load_best_model_at_end:
411
+ value: false
412
+ local_rank:
413
+ value: -1
414
+ log_level:
415
+ value: passive
416
+ log_level_replica:
417
+ value: warning
418
+ log_on_each_node:
419
+ value: true
420
+ logging_dir:
421
+ value: /root/code/na/Idea-filtering-training/train/output/v0-20260304-035120/runs
422
+ logging_first_step:
423
+ value: true
424
+ logging_nan_inf_filter:
425
+ value: true
426
+ logging_steps:
427
+ value: 5
428
+ logging_strategy:
429
+ value: steps
430
+ loss_type:
431
+ value: null
432
+ lr_scheduler_kwargs:
433
+ value: null
434
+ lr_scheduler_type:
435
+ value: cosine
436
+ max_epochs:
437
+ value: 3
438
+ max_grad_norm:
439
+ value: 1
440
+ max_position_embeddings:
441
+ value: 40960
442
+ max_steps:
443
+ value: -1
444
+ max_window_layers:
445
+ value: 28
446
+ metric_for_best_model:
447
+ value: loss
448
+ model/num_parameters:
449
+ value: 2066605056
450
+ model_type:
451
+ value: qwen3
452
+ neftune_noise_alpha:
453
+ value: null
454
+ num_attention_heads:
455
+ value: 16
456
+ num_hidden_layers:
457
+ value: 28
458
+ num_key_value_heads:
459
+ value: 8
460
+ num_train_epochs:
461
+ value: 3
462
+ optim:
463
+ value: adamw_torch_fused
464
+ optim_args:
465
+ value: null
466
+ optim_target_modules:
467
+ value: null
468
+ optimizer:
469
+ value: null
470
+ output_attentions:
471
+ value: false
472
+ output_dir:
473
+ value: /root/code/na/Idea-filtering-training/train/output/v0-20260304-035120
474
+ output_hidden_states:
475
+ value: false
476
+ pad_token_id:
477
+ value: 151643
478
+ parallelism_config:
479
+ value: null
480
+ peft_config:
481
+ value:
482
+ default:
483
+ alora_invocation_tokens: null
484
+ arrow_config: null
485
+ auto_mapping: null
486
+ base_model_name_or_path: /root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e
487
+ bias: none
488
+ corda_config: null
489
+ ensure_weight_tying: false
490
+ eva_config: null
491
+ exclude_modules: null
492
+ fan_in_fan_out: false
493
+ inference_mode: false
494
+ init_lora_weights: true
495
+ layer_replication: null
496
+ layers_pattern: null
497
+ layers_to_transform: null
498
+ lora_alpha: 64
499
+ lora_bias: false
500
+ lora_dropout: 0.05
501
+ lora_dtype: null
502
+ lorap_emb_lr: 1e-06
503
+ lorap_lr_ratio: null
504
+ megatron_config: null
505
+ megatron_core: megatron.core
506
+ modules_to_save: []
507
+ peft_type: LORA
508
+ peft_version: 0.18.1
509
+ qalora_group_size: 16
510
+ r: 32
511
+ revision: null
512
+ runtime_config:
513
+ ephemeral_gpu_offload: false
514
+ target_modules:
515
+ - o_proj
516
+ - q_proj
517
+ - down_proj
518
+ - v_proj
519
+ - gate_proj
520
+ - k_proj
521
+ - up_proj
522
+ target_parameters: null
523
+ task_type: CAUSAL_LM
524
+ trainable_token_indices: null
525
+ use_dora: false
526
+ use_qalora: false
527
+ use_rslora: false
528
+ per_device_eval_batch_size:
529
+ value: 1
530
+ per_device_train_batch_size:
531
+ value: 1
532
+ predict_with_generate:
533
+ value: false
534
+ prediction_loss_only:
535
+ value: false
536
+ problem_type:
537
+ value: null
538
+ project:
539
+ value: huggingface
540
+ push_to_hub:
541
+ value: false
542
+ remove_unused_columns:
543
+ value: false
544
+ report_to:
545
+ value:
546
+ - wandb
547
+ restore_callback_states_from_checkpoint:
548
+ value: false
549
+ resume_from_checkpoint:
550
+ value: null
551
+ resume_only_model:
552
+ value: false
553
+ return_dict:
554
+ value: true
555
+ rms_norm_eps:
556
+ value: 1e-06
557
+ rope_parameters:
558
+ value:
559
+ rope_theta: 1000000
560
+ rope_type: default
561
+ router_aux_loss_coef:
562
+ value: 0
563
+ run_name:
564
+ value: /root/code/na/Idea-filtering-training/train/output/v0-20260304-035120
565
+ save_on_each_node:
566
+ value: false
567
+ save_only_model:
568
+ value: false
569
+ save_steps:
570
+ value: 1000
571
+ save_strategy:
572
+ value: steps
573
+ save_total_limit:
574
+ value: null
575
+ seed:
576
+ value: 42
577
+ skip_memory_metrics:
578
+ value: true
579
+ sliding_window:
580
+ value: null
581
+ sortish_sampler:
582
+ value: false
583
+ tf32:
584
+ value: null
585
+ tie_word_embeddings:
586
+ value: true
587
+ torch_compile:
588
+ value: false
589
+ torch_compile_backend:
590
+ value: null
591
+ torch_compile_mode:
592
+ value: null
593
+ torch_empty_cache_steps:
594
+ value: null
595
+ trackio_space_id:
596
+ value: trackio
597
+ train_dataloader_shuffle:
598
+ value: true
599
+ train_sampling_strategy:
600
+ value: random
601
+ transformers_version:
602
+ value: 5.2.0
603
+ tuner_backend:
604
+ value: peft
605
+ tuner_type:
606
+ value: lora
607
+ use_cache:
608
+ value: false
609
+ use_cpu:
610
+ value: false
611
+ use_flash_ckpt:
612
+ value: false
613
+ use_galore:
614
+ value: false
615
+ use_liger_kernel:
616
+ value: false
617
+ use_logits_to_keep:
618
+ value: null
619
+ use_sliding_window:
620
+ value: false
621
+ vit_gradient_checkpointing:
622
+ value: true
623
+ vit_lr:
624
+ value: null
625
+ vocab_size:
626
+ value: 151936
627
+ warmup_ratio:
628
+ value: 0.05
629
+ warmup_steps:
630
+ value: 0.05
631
+ weight_decay:
632
+ value: 0.1
data/wandb/run-20260304_035219-bzc9z48s/files/output.log ADDED
The diff for this file is too large to render. See raw diff
 
data/wandb/run-20260304_035219-bzc9z48s/files/requirements.txt ADDED
@@ -0,0 +1,159 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trl==0.28.0
2
+ pyarrow==23.0.1
3
+ cffi==2.0.0
4
+ aiosignal==1.4.0
5
+ ms_swift==4.0.0
6
+ aliyun-python-sdk-kms==2.16.5
7
+ nvidia-cublas-cu12==12.8.4.1
8
+ markdown-it-py==4.0.0
9
+ nltk==3.9.3
10
+ cuda-pathfinder==1.4.0
11
+ sentencepiece==0.2.1
12
+ pycryptodome==3.23.0
13
+ hf-xet==1.3.2
14
+ multiprocess==0.70.16
15
+ wandb==0.25.0
16
+ uv==0.10.8
17
+ rich==14.3.3
18
+ modelscope==1.34.0
19
+ websockets==15.0.1
20
+ binpacking==2.0.1
21
+ mpmath==1.3.0
22
+ certifi==2026.2.25
23
+ triton==3.6.0
24
+ nvidia-nvshmem-cu12==3.4.5
25
+ python-multipart==0.0.22
26
+ peft==0.18.1
27
+ httpx==0.28.1
28
+ numpy==2.4.2
29
+ pillow==11.3.0
30
+ wheel==0.46.3
31
+ MarkupSafe==3.0.3
32
+ ruff==0.15.4
33
+ frozenlist==1.8.0
34
+ xxhash==3.6.0
35
+ psutil==7.2.2
36
+ pytz==2026.1.post1
37
+ platformdirs==4.9.2
38
+ joblib==1.5.3
39
+ nvidia-cuda-nvrtc-cu12==12.8.93
40
+ safehttpx==0.1.7
41
+ datasets==3.6.0
42
+ tqdm==4.67.3
43
+ openai==2.24.0
44
+ nvidia-nccl-cu12==2.27.5
45
+ sympy==1.14.0
46
+ Werkzeug==3.1.6
47
+ sentry-sdk==2.54.0
48
+ attrdict==2.0.1
49
+ nvidia-nvtx-cu12==12.8.90
50
+ PyYAML==6.0.3
51
+ requests==2.32.5
52
+ importlib_metadata==8.7.1
53
+ h11==0.16.0
54
+ json_repair==0.58.3
55
+ typer-slim==0.24.0
56
+ pandas==2.3.3
57
+ absl-py==2.4.0
58
+ tomlkit==0.13.3
59
+ nvidia-cuda-cupti-cu12==12.8.90
60
+ gitdb==4.0.12
61
+ tiktoken==0.12.0
62
+ cpm-kernels==1.0.11
63
+ tensorboard-data-server==0.7.2
64
+ sniffio==1.3.1
65
+ sortedcontainers==2.4.0
66
+ nvidia-curand-cu12==10.3.9.90
67
+ aiohttp==3.13.3
68
+ transformers-stream-generator==0.0.5
69
+ gradio==5.50.0
70
+ pycparser==3.0
71
+ yarl==1.23.0
72
+ Pygments==2.19.2
73
+ ffmpy==1.0.0
74
+ brotli==1.2.0
75
+ idna==3.11
76
+ Jinja2==3.1.6
77
+ pip==26.0.1
78
+ jieba==0.42.1
79
+ typing_extensions==4.15.0
80
+ torch==2.10.0
81
+ attrs==25.4.0
82
+ cuda-bindings==12.9.4
83
+ filelock==3.25.0
84
+ fonttools==4.61.1
85
+ accelerate==1.12.0
86
+ httpcore==1.0.9
87
+ charset-normalizer==3.4.4
88
+ regex==2026.2.28
89
+ simplejson==3.20.2
90
+ tensorboard==2.20.0
91
+ python-dateutil==2.9.0.post0
92
+ tzdata==2025.3
93
+ typer==0.24.1
94
+ pydantic==2.12.3
95
+ jiter==0.13.0
96
+ annotated-doc==0.0.4
97
+ aliyun-python-sdk-core==2.16.0
98
+ oss2==2.19.1
99
+ fastapi==0.135.1
100
+ shellingham==1.5.4
101
+ addict==2.4.0
102
+ omegaconf==2.3.0
103
+ distro==1.9.0
104
+ fsspec==2025.3.0
105
+ cycler==0.12.1
106
+ rouge==1.0.1
107
+ grpcio==1.78.0
108
+ kiwisolver==1.4.9
109
+ Markdown==3.10.2
110
+ pydub==0.25.1
111
+ protobuf==6.33.5
112
+ crcmod==1.7
113
+ typing-inspection==0.4.2
114
+ zstandard==0.25.0
115
+ gradio_client==1.14.0
116
+ tokenizers==0.22.2
117
+ packaging==25.0
118
+ zipp==3.23.0
119
+ mdurl==0.1.2
120
+ contourpy==1.3.3
121
+ aiohappyeyeballs==2.6.1
122
+ transformers==5.2.0
123
+ nvidia-cuda-runtime-cu12==12.8.90
124
+ semantic-version==2.10.0
125
+ dacite==1.9.2
126
+ antlr4-python3-runtime==4.9.3
127
+ anyio==4.12.1
128
+ aiofiles==24.1.0
129
+ nvidia-cusparse-cu12==12.5.8.93
130
+ networkx==3.6.1
131
+ cryptography==46.0.5
132
+ uvicorn==0.41.0
133
+ nvidia-cusolver-cu12==11.7.3.90
134
+ pydantic_core==2.41.4
135
+ nvidia-cufile-cu12==1.13.1.3
136
+ pyparsing==3.3.2
137
+ nvidia-cusparselt-cu12==0.7.1
138
+ annotated-types==0.7.0
139
+ propcache==0.4.1
140
+ click==8.3.1
141
+ starlette==0.52.1
142
+ orjson==3.11.7
143
+ dill==0.3.8
144
+ groovy==0.1.2
145
+ safetensors==0.7.0
146
+ GitPython==3.1.46
147
+ smmap==5.0.2
148
+ nvidia-nvjitlink-cu12==12.8.93
149
+ einops==0.8.2
150
+ nvidia-cufft-cu12==11.3.3.83
151
+ multidict==6.7.1
152
+ matplotlib==3.10.8
153
+ six==1.17.0
154
+ jmespath==0.10.0
155
+ nvidia-cudnn-cu12==9.10.2.21
156
+ urllib3==2.6.3
157
+ setuptools==80.10.2
158
+ scipy==1.17.1
159
+ huggingface_hub==1.5.0
data/wandb/run-20260304_035219-bzc9z48s/files/wandb-metadata.json ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-87-generic-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.12.12",
4
+ "startedAt": "2026-03-04T03:52:19.561578Z",
5
+ "args": [
6
+ "--model",
7
+ "Qwen/Qwen3-1.7B",
8
+ "--use_hf",
9
+ "true",
10
+ "--dataset",
11
+ "/root/code/na/Idea-filtering-training/data/thinking_dataset_v2_swift.jsonl",
12
+ "--train_type",
13
+ "lora",
14
+ "--lora_rank",
15
+ "32",
16
+ "--lora_alpha",
17
+ "64",
18
+ "--lora_dropout",
19
+ "0.05",
20
+ "--target_modules",
21
+ "q_proj",
22
+ "k_proj",
23
+ "v_proj",
24
+ "o_proj",
25
+ "gate_proj",
26
+ "up_proj",
27
+ "down_proj",
28
+ "--lora_ga_batch_size",
29
+ "1",
30
+ "--lora_ga_iters",
31
+ "2",
32
+ "--lora_ga_max_length",
33
+ "1024",
34
+ "--lora_ga_direction",
35
+ "ArB2r",
36
+ "--lora_ga_scale",
37
+ "stable",
38
+ "--lora_ga_stable_gamma",
39
+ "16",
40
+ "--per_device_train_batch_size",
41
+ "1",
42
+ "--gradient_accumulation_steps",
43
+ "16",
44
+ "--max_length",
45
+ "2600",
46
+ "--learning_rate",
47
+ "2e-4",
48
+ "--warmup_ratio",
49
+ "0.05",
50
+ "--max_epochs",
51
+ "3",
52
+ "--attn_impl",
53
+ "eager",
54
+ "--output_dir",
55
+ "/root/code/na/Idea-filtering-training/train/output",
56
+ "--save_strategy",
57
+ "steps",
58
+ "--save_steps",
59
+ "1000",
60
+ "--eval_steps",
61
+ "500",
62
+ "--dataloader_num_workers",
63
+ "1",
64
+ "--dataset_num_proc",
65
+ "16",
66
+ "--load_from_cache_file",
67
+ "true",
68
+ "--model_author",
69
+ "swift",
70
+ "--model_name",
71
+ "swift-robot",
72
+ "--loss_scale",
73
+ "ignore_empty_think",
74
+ "--report_to",
75
+ "wandb"
76
+ ],
77
+ "program": "/root/miniconda3/envs/RM/lib/python3.12/site-packages/swift/cli/sft.py",
78
+ "git": {
79
+ "remote": "https://tnnanh1010:@github.com/Drylab-AI/Idea-filtering-training.git",
80
+ "commit": "0dbe0bbbe696b2e49ffefa83123579a73fab61c6"
81
+ },
82
+ "email": "nhatanh10102005@gmail.com",
83
+ "root": "/root/code/na/Idea-filtering-training/data",
84
+ "host": "ubuntu",
85
+ "executable": "/root/miniconda3/envs/RM/bin/python3",
86
+ "cpu_count": 49,
87
+ "cpu_count_logical": 49,
88
+ "gpu": "NVIDIA GeForce RTX 3060",
89
+ "gpu_count": 1,
90
+ "disk": {
91
+ "/": {
92
+ "total": "960312868864",
93
+ "used": "865124937728"
94
+ }
95
+ },
96
+ "memory": {
97
+ "total": "628090474496"
98
+ },
99
+ "gpu_nvidia": [
100
+ {
101
+ "name": "NVIDIA GeForce RTX 3060",
102
+ "memoryTotal": "12884901888",
103
+ "cudaCores": 3584,
104
+ "architecture": "Ampere",
105
+ "uuid": "GPU-f370a18a-d749-7bf1-c827-82726afbb6cf"
106
+ }
107
+ ],
108
+ "cudaVersion": "13.0",
109
+ "writerId": "cu7bd4tvbr27kn8zb6sicf7uvdvg0f2r"
110
+ }
data/wandb/run-20260304_035219-bzc9z48s/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/token_acc":0.713329659081962,"train/grad_norm":0.028987018391489983,"_step":875,"train_samples_per_second":0.588,"train/epoch":3,"train_loss":1.1254750251988053,"_wandb":{"runtime":118958},"total_flos":1.168560397849688e+18,"_runtime":118958,"train_steps_per_second":0.037,"train_runtime":118959.0608,"train/learning_rate":4.573485906322894e-10,"_timestamp":1.772715298382969e+09,"train/loss":0.9691848754882812,"train/global_step":4374}
data/wandb/run-20260304_035219-bzc9z48s/logs/debug-core.log ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-04T03:52:19.636010859Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpizzvshbt/port-2895203.txt","pid":2895203,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2026-03-04T03:52:19.640913014Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":2895203}
3
+ {"time":"2026-03-04T03:52:19.64082463Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2895203-2907053-1415758021/socket","Net":"unix"}}
4
+ {"time":"2026-03-04T03:52:19.788521881Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2026-03-04T03:52:19.800324062Z","level":"INFO","msg":"handleInformInit: received","streamId":"bzc9z48s","id":"1(@)"}
6
+ {"time":"2026-03-04T03:52:20.001980096Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"bzc9z48s","id":"1(@)"}
7
+ {"time":"2026-03-04T03:52:25.452788253Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"rpf9en822nne"}
8
+ {"time":"2026-03-05T12:54:58.719926201Z","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
9
+ {"time":"2026-03-05T12:54:58.720391834Z","level":"INFO","msg":"server is shutting down"}
10
+ {"time":"2026-03-05T12:54:58.720342858Z","level":"INFO","msg":"connection: closing","id":"1(@)"}
11
+ {"time":"2026-03-05T12:54:58.720715555Z","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
12
+ {"time":"2026-03-05T12:54:58.720699881Z","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2895203-2907053-1415758021/socket","Net":"unix"}}
13
+ {"time":"2026-03-05T12:54:59.799535118Z","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
14
+ {"time":"2026-03-05T12:54:59.79964598Z","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
15
+ {"time":"2026-03-05T12:54:59.799674575Z","level":"INFO","msg":"server is closed"}
data/wandb/run-20260304_035219-bzc9z48s/logs/debug-internal.log ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-04T03:52:19.801424661Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-03-04T03:52:20.001672605Z","level":"INFO","msg":"stream: created new stream","id":"bzc9z48s"}
3
+ {"time":"2026-03-04T03:52:20.001883943Z","level":"INFO","msg":"handler: started","stream_id":"bzc9z48s"}
4
+ {"time":"2026-03-04T03:52:20.001944483Z","level":"INFO","msg":"stream: started","id":"bzc9z48s"}
5
+ {"time":"2026-03-04T03:52:20.001991456Z","level":"INFO","msg":"sender: started","stream_id":"bzc9z48s"}
6
+ {"time":"2026-03-04T03:52:20.001995704Z","level":"INFO","msg":"writer: started","stream_id":"bzc9z48s"}
7
+ {"time":"2026-03-04T22:00:05.351127978Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/nhatanh10102005-vinuniversity/ms-swift/bzc9z48s/file_stream","body":"\n<html><head>\n<meta http-equiv=\"content-type\" content=\"text/html;charset=utf-8\">\n<title>502 Server Error</title>\n</head>\n<body text=#000000 bgcolor=#ffffff>\n<h1>Error: Server Error</h1>\n<h2>The server encountered a temporary error and could not complete your request.<p>Please try again in 30 seconds.</h2>\n<h2></h2>\n</body></html>\n"}
8
+ {"time":"2026-03-04T22:06:05.347371295Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/nhatanh10102005-vinuniversity/ms-swift/bzc9z48s/file_stream","body":"\n<html><head>\n<meta http-equiv=\"content-type\" content=\"text/html;charset=utf-8\">\n<title>502 Server Error</title>\n</head>\n<body text=#000000 bgcolor=#ffffff>\n<h1>Error: Server Error</h1>\n<h2>The server encountered a temporary error and could not complete your request.<p>Please try again in 30 seconds.</h2>\n<h2></h2>\n</body></html>\n"}
9
+ {"time":"2026-03-05T02:39:18.098277092Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/nhatanh10102005-vinuniversity/ms-swift/bzc9z48s/file_stream","body":"\n<html><head>\n<meta http-equiv=\"content-type\" content=\"text/html;charset=utf-8\">\n<title>502 Server Error</title>\n</head>\n<body text=#000000 bgcolor=#ffffff>\n<h1>Error: Server Error</h1>\n<h2>The server encountered a temporary error and could not complete your request.<p>Please try again in 30 seconds.</h2>\n<h2></h2>\n</body></html>\n"}
10
+ {"time":"2026-03-05T12:54:58.720317991Z","level":"INFO","msg":"stream: closing","id":"bzc9z48s"}
11
+ {"time":"2026-03-05T12:54:59.333556397Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
12
+ {"time":"2026-03-05T12:54:59.784047742Z","level":"INFO","msg":"handler: closed","stream_id":"bzc9z48s"}
13
+ {"time":"2026-03-05T12:54:59.784388437Z","level":"INFO","msg":"sender: closed","stream_id":"bzc9z48s"}
14
+ {"time":"2026-03-05T12:54:59.784422878Z","level":"INFO","msg":"stream: closed","id":"bzc9z48s"}
data/wandb/run-20260304_035219-bzc9z48s/logs/debug.log ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-04 03:52:19,563 INFO MainThread:2895203 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-03-04 03:52:19,563 INFO MainThread:2895203 [wandb_setup.py:_flush():81] Configure stats pid to 2895203
3
+ 2026-03-04 03:52:19,563 INFO MainThread:2895203 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-03-04 03:52:19,563 INFO MainThread:2895203 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /root/code/na/Idea-filtering-training/data/wandb/run-20260304_035219-bzc9z48s/logs/debug.log
5
+ 2026-03-04 03:52:19,563 INFO MainThread:2895203 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /root/code/na/Idea-filtering-training/data/wandb/run-20260304_035219-bzc9z48s/logs/debug-internal.log
6
+ 2026-03-04 03:52:19,563 INFO MainThread:2895203 [wandb_init.py:init():844] calling init triggers
7
+ 2026-03-04 03:52:19,563 INFO MainThread:2895203 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'_wandb': {}}
9
+ 2026-03-04 03:52:19,563 INFO MainThread:2895203 [wandb_init.py:init():892] starting backend
10
+ 2026-03-04 03:52:19,788 INFO MainThread:2895203 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-03-04 03:52:19,791 INFO MainThread:2895203 [wandb_init.py:init():903] backend started and connected
12
+ 2026-03-04 03:52:19,793 INFO MainThread:2895203 [wandb_init.py:init():973] updated telemetry
13
+ 2026-03-04 03:52:19,802 INFO MainThread:2895203 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-03-04 03:52:20,304 INFO MainThread:2895203 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-03-04 03:52:20,444 INFO MainThread:2895203 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-03-04 03:52:20,444 INFO MainThread:2895203 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-03-04 03:52:20,444 INFO MainThread:2895203 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-03-04 03:52:20,444 INFO MainThread:2895203 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-03-04 03:52:20,457 INFO MainThread:2895203 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-03-04 03:52:20,464 INFO MainThread:2895203 [wandb_run.py:_config_callback():1403] config_cb None None {'peft_config': {'default': {'task_type': 'CAUSAL_LM', 'peft_type': 'LORA', 'auto_mapping': None, 'peft_version': '0.18.1', 'base_model_name_or_path': '/root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e', 'revision': None, 'inference_mode': False, 'r': 32, 'target_modules': ['o_proj', 'q_proj', 'down_proj', 'v_proj', 'gate_proj', 'k_proj', 'up_proj'], 'exclude_modules': None, 'lora_alpha': 64, 'lora_dropout': 0.05, 'fan_in_fan_out': False, 'bias': 'none', 'use_rslora': False, 'modules_to_save': [], 'init_lora_weights': True, 'layers_to_transform': None, 'layers_pattern': None, 'rank_pattern': {}, 'alpha_pattern': {}, 'megatron_config': None, 'megatron_core': 'megatron.core', 'trainable_token_indices': None, 'loftq_config': {}, 'eva_config': None, 'corda_config': None, 'use_dora': False, 'alora_invocation_tokens': None, 'use_qalora': False, 'qalora_group_size': 16, 'layer_replication': None, 'runtime_config': {'ephemeral_gpu_offload': False}, 'lora_bias': False, 'target_parameters': None, 'arrow_config': None, 'ensure_weight_tying': False, 'lora_dtype': None, 'lorap_lr_ratio': None, 'lorap_emb_lr': 1e-06}}, 'vocab_size': 151936, 'max_position_embeddings': 40960, 'hidden_size': 2048, 'intermediate_size': 6144, 'num_hidden_layers': 28, 'num_attention_heads': 16, 'use_sliding_window': False, 'sliding_window': None, 'max_window_layers': 28, 'num_key_value_heads': 8, 'head_dim': 128, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-06, 'use_cache': False, 'attention_bias': False, 'attention_dropout': 0.0, 'layer_types': ['full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention'], 'pad_token_id': 151643, 'bos_token_id': None, 'eos_token_id': 151645, 'tie_word_embeddings': True, 'rope_parameters': {'rope_theta': 1000000, 'rope_type': 'default'}, 'return_dict': True, 'output_hidden_states': False, 'dtype': 'bfloat16', 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'architectures': ['Qwen3ForCausalLM'], 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'problem_type': None, '_name_or_path': '/root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e', 'transformers_version': '5.2.0', 'model_type': 'qwen3', 'output_attentions': False, 'output_dir': '/root/code/na/Idea-filtering-training/train/output/v0-20260304-035120', 'per_device_train_batch_size': 1, 'num_train_epochs': 3.0, 'max_steps': -1, 'learning_rate': 0.0002, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': None, 'warmup_steps': 0.05, 'optim': 'adamw_torch_fused', 'optim_args': None, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.95, 'adam_epsilon': 1e-08, 'optim_target_modules': None, 'gradient_accumulation_steps': 16, 'average_tokens_across_devices': True, 'max_grad_norm': 1.0, 'label_smoothing_factor': 0.0, 'bf16': True, 'fp16': False, 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'use_liger_kernel': False, 'liger_kernel_config': None, 'neftune_noise_alpha': None, 'torch_empty_cache_steps': None, 'auto_find_batch_size': False, 'logging_strategy': 'steps', 'logging_steps': 5, 'logging_first_step': True, 'log_on_each_node': True, 'logging_nan_inf_filter': True, 'include_num_input_tokens_seen': 'no', 'log_level': 'passive', 'log_level_replica': 'warning', 'disable_tqdm': False, 'report_to': ['wandb'], 'run_name': '/root/code/na/Idea-filtering-training/train/output/v0-20260304-035120', 'project': 'huggingface', 'trackio_space_id': 'trackio', 'eval_strategy': 'no', 'eval_steps': 500.0, 'eval_delay': 0, 'per_device_eval_batch_size': 1, 'prediction_loss_only': False, 'eval_on_start': False, 'eval_do_concat_batches': True, 'eval_use_gather_object': False, 'eval_accumulation_steps': None, 'include_for_metrics': [], 'batch_eval_metrics': False, 'save_only_model': False, 'save_strategy': 'steps', 'save_steps': 1000, 'save_on_each_node': False, 'save_total_limit': None, 'enable_jit_checkpoint': False, 'push_to_hub': False, 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_always_push': False, 'hub_revision': None, 'load_best_model_at_end': False, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'restore_callback_states_from_checkpoint': False, 'full_determinism': False, 'seed': 42, 'data_seed': 42, 'use_cpu': False, 'accelerator_config': {'split_batches': False, 'dispatch_batches': False, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'parallelism_config': None, 'dataloader_drop_last': False, 'dataloader_num_workers': 1, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'dataloader_prefetch_factor': 2, 'remove_unused_columns': False, 'label_names': None, 'train_sampling_strategy': 'random', 'length_column_name': 'length', 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'ddp_backend': None, 'ddp_timeout': 18000000, 'fsdp': [], 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'deepspeed': None, 'debug': [], 'skip_memory_metrics': True, 'do_train': False, 'do_eval': False, 'do_predict': False, 'resume_from_checkpoint': None, 'warmup_ratio': 0.05, 'logging_dir': '/root/code/na/Idea-filtering-training/train/output/v0-20260304-035120/runs', 'local_rank': -1, 'sortish_sampler': False, 'predict_with_generate': False, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None, 'tuner_backend': 'peft', 'vit_gradient_checkpointing': True, 'router_aux_loss_coef': 0.0, 'enable_dft_loss': False, 'enable_channel_loss': False, 'check_model': True, 'acc_strategy': 'token', 'train_dataloader_shuffle': True, 'group_by_length': False, 'max_epochs': 3, 'aligner_lr': None, 'vit_lr': None, 'use_logits_to_keep': None, 'ds3_gather_for_generation': True, 'resume_only_model': False, 'optimizer': None, 'loss_type': None, 'eval_metric': None, 'callbacks': [], 'early_stop_interval': None, 'eval_use_evalscope': False, 'eval_dataset': [], 'eval_dataset_args': None, 'eval_limit': None, 'eval_generation_config': None, 'extra_eval_args': None, 'tuner_type': 'lora', 'use_galore': False, 'galore_target_modules': None, 'galore_rank': 128, 'galore_update_proj_gap': 50, 'galore_scale': 1.0, 'galore_proj_type': 'std', 'galore_optim_per_parameter': False, 'galore_with_embedding': False, 'galore_quantization': False, 'galore_proj_quant': False, 'galore_proj_bits': 4, 'galore_proj_group_size': 256, 'galore_cos_threshold': 0.4, 'galore_gamma_proj': 2, 'galore_queue_size': 5, 'lisa_activated_layers': 0, 'lisa_step_interval': 20, 'use_flash_ckpt': False}
21
+ 2026-03-04 03:52:20,476 INFO MainThread:2895203 [wandb_config.py:__setitem__():155] [no run ID] config set model/num_parameters = 2066605056 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7ffdd71ffd40>>
22
+ 2026-03-04 03:52:20,476 INFO MainThread:2895203 [wandb_run.py:_config_callback():1403] config_cb model/num_parameters 2066605056 None
23
+ 2026-03-05 12:54:58,719 INFO wandb-AsyncioManager-main:2895203 [service_client.py:_forward_responses():134] Reached EOF.
24
+ 2026-03-05 12:54:58,720 INFO wandb-AsyncioManager-main:2895203 [mailbox.py:close():155] Closing mailbox, abandoning 1 handles.
data/wandb/run-20260304_035219-bzc9z48s/run-bzc9z48s.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfbdbb63608a901207189fc07dd128b30e789efd6d2ad661822b553b1d5bb4e0
3
+ size 9286321
data/wandb/run-20260306_073928-dpxzw953/files/output.log ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Train: 0%| | 0/4374 [00:00<?, ?it/s][INFO:swift] use_logits_to_keep: True
2
+ Train: 3%|██▍ | 150/4374 [1:05:30<30:16:07, 25.80s/it]
3
+ {'loss': '2.883', 'grad_norm': '0.759', 'learning_rate': '9.1e-07', 'token_acc': '0.5381', 'epoch': '0.0006863', 'global_step/max_steps': '1/4374', 'elapsed_time': '27s', 'remaining_time': '1d 8h 53m 59s', 'memory(GiB)': '8.7', 'train_speed(s/it)': '27.08'}
4
+ {'loss': '2.962', 'grad_norm': '0.7302', 'learning_rate': '4.57e-06', 'token_acc': '0.534', 'epoch': '0.003431', 'global_step/max_steps': '5/4374', 'elapsed_time': '2m 13s', 'remaining_time': '1d 8h 16m 8s', 'memory(GiB)': '9.95', 'train_speed(s/it)': '26.59'}
5
+ {'loss': '2.918', 'grad_norm': '0.6915', 'learning_rate': '9.13e-06', 'token_acc': '0.5313', 'epoch': '0.006863', 'global_step/max_steps': '10/4374', 'elapsed_time': '4m 25s', 'remaining_time': '1d 8h 4m 23s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.46'}
6
+ {'loss': '2.736', 'grad_norm': '0.4563', 'learning_rate': '1.37e-05', 'token_acc': '0.5312', 'epoch': '0.01029', 'global_step/max_steps': '15/4374', 'elapsed_time': '6m 35s', 'remaining_time': '1d 7h 54m 24s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.35'}
7
+ {'loss': '2.398', 'grad_norm': '0.2684', 'learning_rate': '1.826e-05', 'token_acc': '0.5442', 'epoch': '0.01373', 'global_step/max_steps': '20/4374', 'elapsed_time': '8m 46s', 'remaining_time': '1d 7h 50m 4s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.32'}
8
+ {'loss': '2.196', 'grad_norm': '0.149', 'learning_rate': '2.283e-05', 'token_acc': '0.5421', 'epoch': '0.01716', 'global_step/max_steps': '25/4374', 'elapsed_time': '10m 59s', 'remaining_time': '1d 7h 50m 15s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.35'}
9
+ {'loss': '2.046', 'grad_norm': '0.07151', 'learning_rate': '2.74e-05', 'token_acc': '0.5469', 'epoch': '0.02059', 'global_step/max_steps': '30/4374', 'elapsed_time': '13m 9s', 'remaining_time': '1d 7h 44m 51s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.31'}
10
+ {'loss': '1.945', 'grad_norm': '0.05515', 'learning_rate': '3.196e-05', 'token_acc': '0.545', 'epoch': '0.02402', 'global_step/max_steps': '35/4374', 'elapsed_time': '15m 22s', 'remaining_time': '1d 7h 44m 34s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.34'}
11
+ {'loss': '1.924', 'grad_norm': '0.05004', 'learning_rate': '3.653e-05', 'token_acc': '0.5445', 'epoch': '0.02745', 'global_step/max_steps': '40/4374', 'elapsed_time': '17m 33s', 'remaining_time': '1d 7h 41m 42s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.33'}
12
+ {'loss': '1.857', 'grad_norm': '0.05194', 'learning_rate': '4.11e-05', 'token_acc': '0.5542', 'epoch': '0.03088', 'global_step/max_steps': '45/4374', 'elapsed_time': '19m 44s', 'remaining_time': '1d 7h 37m 36s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.3'}
13
+ {'loss': '1.752', 'grad_norm': '0.04649', 'learning_rate': '4.566e-05', 'token_acc': '0.5766', 'epoch': '0.03431', 'global_step/max_steps': '50/4374', 'elapsed_time': '21m 56s', 'remaining_time': '1d 7h 36m 37s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.32'}
14
+ {'loss': '1.753', 'grad_norm': '0.0284', 'learning_rate': '5.023e-05', 'token_acc': '0.5762', 'epoch': '0.03774', 'global_step/max_steps': '55/4374', 'elapsed_time': '24m 8s', 'remaining_time': '1d 7h 35m 17s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.33'}
15
+ {'loss': '1.729', 'grad_norm': '0.03265', 'learning_rate': '5.479e-05', 'token_acc': '0.5771', 'epoch': '0.04118', 'global_step/max_steps': '60/4374', 'elapsed_time': '26m 20s', 'remaining_time': '1d 7h 33m 20s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.33'}
16
+ {'loss': '1.723', 'grad_norm': '0.02901', 'learning_rate': '5.936e-05', 'token_acc': '0.583', 'epoch': '0.04461', 'global_step/max_steps': '65/4374', 'elapsed_time': '28m 34s', 'remaining_time': '1d 7h 33m 53s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.37'}
17
+ {'loss': '1.613', 'grad_norm': '0.02471', 'learning_rate': '6.393e-05', 'token_acc': '0.5987', 'epoch': '0.04804', 'global_step/max_steps': '70/4374', 'elapsed_time': '30m 43s', 'remaining_time': '1d 7h 28m 57s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.33'}
18
+ {'loss': '1.641', 'grad_norm': '0.02978', 'learning_rate': '6.849e-05', 'token_acc': '0.5909', 'epoch': '0.05147', 'global_step/max_steps': '75/4374', 'elapsed_time': '32m 51s', 'remaining_time': '1d 7h 22m 57s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.28'}
19
+ {'loss': '1.621', 'grad_norm': '0.02749', 'learning_rate': '7.306e-05', 'token_acc': '0.5946', 'epoch': '0.0549', 'global_step/max_steps': '80/4374', 'elapsed_time': '35m 1s', 'remaining_time': '1d 7h 19m 11s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.26'}
20
+ {'loss': '1.587', 'grad_norm': '0.03279', 'learning_rate': '7.763e-05', 'token_acc': '0.5998', 'epoch': '0.05833', 'global_step/max_steps': '85/4374', 'elapsed_time': '37m 9s', 'remaining_time': '1d 7h 14m 44s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.23'}
21
+ {'loss': '1.565', 'grad_norm': '0.02734', 'learning_rate': '8.219e-05', 'token_acc': '0.6056', 'epoch': '0.06176', 'global_step/max_steps': '90/4374', 'elapsed_time': '39m 19s', 'remaining_time': '1d 7h 11m 14s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.21'}
22
+ {'loss': '1.518', 'grad_norm': '0.02591', 'learning_rate': '8.676e-05', 'token_acc': '0.6154', 'epoch': '0.06519', 'global_step/max_steps': '95/4374', 'elapsed_time': '41m 30s', 'remaining_time': '1d 7h 8m 56s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.21'}
23
+ {'loss': '1.506', 'grad_norm': '0.02655', 'learning_rate': '9.132e-05', 'token_acc': '0.6157', 'epoch': '0.06863', 'global_step/max_steps': '100/4374', 'elapsed_time': '43m 38s', 'remaining_time': '1d 7h 4m 58s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.18'}
24
+ {'loss': '1.522', 'grad_norm': '0.02904', 'learning_rate': '9.589e-05', 'token_acc': '0.6135', 'epoch': '0.07206', 'global_step/max_steps': '105/4374', 'elapsed_time': '45m 47s', 'remaining_time': '1d 7h 1m 26s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.16'}
25
+ {'loss': '1.469', 'grad_norm': '0.03006', 'learning_rate': '0.0001005', 'token_acc': '0.6211', 'epoch': '0.07549', 'global_step/max_steps': '110/4374', 'elapsed_time': '48m 3s', 'remaining_time': '1d 7h 2m 29s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.21'}
26
+ {'loss': '1.459', 'grad_norm': '0.03309', 'learning_rate': '0.000105', 'token_acc': '0.6205', 'epoch': '0.07892', 'global_step/max_steps': '115/4374', 'elapsed_time': '50m 15s', 'remaining_time': '1d 7h 0m 58s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.22'}
27
+ {'loss': '1.413', 'grad_norm': '0.02772', 'learning_rate': '0.0001096', 'token_acc': '0.6308', 'epoch': '0.08235', 'global_step/max_steps': '120/4374', 'elapsed_time': '52m 26s', 'remaining_time': '1d 6h 58m 30s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.21'}
28
+ {'loss': '1.428', 'grad_norm': '0.03154', 'learning_rate': '0.0001142', 'token_acc': '0.6297', 'epoch': '0.08578', 'global_step/max_steps': '125/4374', 'elapsed_time': '54m 38s', 'remaining_time': '1d 6h 57m 19s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.23'}
29
+ {'loss': '1.41', 'grad_norm': '0.03049', 'learning_rate': '0.0001187', 'token_acc': '0.6343', 'epoch': '0.08921', 'global_step/max_steps': '130/4374', 'elapsed_time': '56m 50s', 'remaining_time': '1d 6h 55m 20s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.23'}
30
+ {'loss': '1.426', 'grad_norm': '0.03384', 'learning_rate': '0.0001233', 'token_acc': '0.631', 'epoch': '0.09264', 'global_step/max_steps': '135/4374', 'elapsed_time': '58m 60s', 'remaining_time': '1d 6h 52m 27s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.22'}
31
+ {'loss': '1.328', 'grad_norm': '0.03483', 'learning_rate': '0.0001278', 'token_acc': '0.6469', 'epoch': '0.09608', 'global_step/max_steps': '140/4374', 'elapsed_time': '1h 1m 8s', 'remaining_time': '1d 6h 48m 55s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.2'}
32
+ {'loss': '1.445', 'grad_norm': '0.03589', 'learning_rate': '0.0001324', 'token_acc': '0.6279', 'epoch': '0.09951', 'global_step/max_steps': '145/4374', 'elapsed_time': '1h 3m 20s', 'remaining_time': '1d 6h 47m 14s', 'memory(GiB)': '11.16', 'train_speed(s/it)': '26.21'}
data/wandb/run-20260306_073928-dpxzw953/files/requirements.txt ADDED
@@ -0,0 +1,241 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ psutil==7.2.2
2
+ colorama==0.4.6
3
+ trl==0.28.0
4
+ pyarrow==23.0.1
5
+ msgpack==1.1.2
6
+ astor==0.8.1
7
+ cffi==2.0.0
8
+ aiosignal==1.4.0
9
+ sse-starlette==3.3.2
10
+ setproctitle==1.3.7
11
+ ms_swift==4.0.0
12
+ mistral_common==1.9.1
13
+ aliyun-python-sdk-kms==2.16.5
14
+ nvidia-cublas-cu12==12.8.4.1
15
+ markdown-it-py==4.0.0
16
+ cachetools==7.0.3
17
+ nltk==3.9.3
18
+ cuda-pathfinder==1.4.0
19
+ sentencepiece==0.2.1
20
+ pycryptodome==3.23.0
21
+ hf-xet==1.3.2
22
+ multiprocess==0.70.16
23
+ wandb==0.25.0
24
+ fastar==0.8.0
25
+ uv==0.10.8
26
+ interegular==0.3.3
27
+ rich==14.3.3
28
+ modelscope==1.34.0
29
+ websockets==15.0.1
30
+ binpacking==2.0.1
31
+ seaborn==0.13.2
32
+ uvloop==0.22.1
33
+ mpmath==1.3.0
34
+ certifi==2026.2.25
35
+ vllm==0.16.0
36
+ supervisor==4.3.0
37
+ flashinfer-python==0.6.3
38
+ cbor2==5.8.0
39
+ msgspec==0.20.0
40
+ threadpoolctl==3.6.0
41
+ watchfiles==1.1.1
42
+ python-multipart==0.0.22
43
+ peft==0.18.1
44
+ httpx==0.28.1
45
+ pillow==11.3.0
46
+ wheel==0.46.3
47
+ MarkupSafe==3.0.3
48
+ ruff==0.15.4
49
+ frozenlist==1.8.0
50
+ xxhash==3.6.0
51
+ psutil==7.2.2
52
+ pytz==2026.1.post1
53
+ platformdirs==4.9.2
54
+ joblib==1.5.3
55
+ transformers==5.3.0
56
+ scikit-learn==1.8.0
57
+ nvidia-cuda-nvrtc-cu12==12.8.93
58
+ nvidia-cutlass-dsl-libs-base==4.4.1
59
+ openai-harmony==0.0.8
60
+ safehttpx==0.1.7
61
+ gdown==5.2.1
62
+ diskcache==5.6.3
63
+ datasets==3.6.0
64
+ tqdm==4.67.3
65
+ openai==2.24.0
66
+ nvidia-nccl-cu12==2.27.5
67
+ sympy==1.14.0
68
+ tabulate==0.10.0
69
+ Werkzeug==3.1.6
70
+ sentry-sdk==2.54.0
71
+ numba==0.61.2
72
+ attrdict==2.0.1
73
+ referencing==0.37.0
74
+ cuda-python==13.1.1
75
+ jsonschema-specifications==2025.9.1
76
+ nvidia-nvtx-cu12==12.8.90
77
+ email-validator==2.3.0
78
+ PyYAML==6.0.3
79
+ requests==2.32.5
80
+ cloudpickle==3.1.2
81
+ xgrammar==0.1.29
82
+ importlib_metadata==8.7.1
83
+ prometheus_client==0.24.1
84
+ h11==0.16.0
85
+ json_repair==0.58.3
86
+ typer-slim==0.24.0
87
+ pandas==2.3.3
88
+ absl-py==2.4.0
89
+ tomlkit==0.13.3
90
+ nvidia-cuda-cupti-cu12==12.8.90
91
+ gitdb==4.0.12
92
+ tiktoken==0.12.0
93
+ soupsieve==2.8.3
94
+ cpm-kernels==1.0.11
95
+ tensorboard-data-server==0.7.2
96
+ sniffio==1.3.1
97
+ sortedcontainers==2.4.0
98
+ pycountry==26.2.16
99
+ triton==3.5.1
100
+ nvidia-curand-cu12==10.3.9.90
101
+ beautifulsoup4==4.14.3
102
+ aiohttp==3.13.3
103
+ nvidia-cutlass-dsl==4.4.1
104
+ transformers-stream-generator==0.0.5
105
+ gradio==5.50.0
106
+ lark==1.2.2
107
+ pycparser==3.0
108
+ yarl==1.23.0
109
+ outlines_core==0.2.11
110
+ anthropic==0.84.0
111
+ jsonschema==4.26.0
112
+ Pygments==2.19.2
113
+ nvidia-nvshmem-cu12==3.3.20
114
+ ffmpy==1.0.0
115
+ brotli==1.2.0
116
+ idna==3.11
117
+ Jinja2==3.1.6
118
+ pip==26.0.1
119
+ docstring_parser==0.17.0
120
+ jieba==0.42.1
121
+ typing_extensions==4.15.0
122
+ torch==2.9.1
123
+ attrs==25.4.0
124
+ filelock==3.25.0
125
+ fonttools==4.61.1
126
+ accelerate==1.12.0
127
+ httpcore==1.0.9
128
+ ijson==3.5.0
129
+ model-hosting-container-standards==0.1.13
130
+ PyJWT==2.11.0
131
+ charset-normalizer==3.4.4
132
+ regex==2026.2.28
133
+ simplejson==3.20.2
134
+ tensorboard==2.20.0
135
+ python-json-logger==4.0.0
136
+ python-dateutil==2.9.0.post0
137
+ nvidia-cudnn-frontend==1.18.0
138
+ tzdata==2025.3
139
+ typer==0.24.1
140
+ pydantic==2.12.3
141
+ jiter==0.13.0
142
+ annotated-doc==0.0.4
143
+ aliyun-python-sdk-core==2.16.0
144
+ oss2==2.19.1
145
+ fastapi==0.135.1
146
+ shellingham==1.5.4
147
+ addict==2.4.0
148
+ omegaconf==2.3.0
149
+ distro==1.9.0
150
+ depyf==0.20.0
151
+ numpy==2.2.6
152
+ rignore==0.7.6
153
+ torchaudio==2.9.1
154
+ ray==2.54.0
155
+ blake3==1.0.8
156
+ fsspec==2025.3.0
157
+ cycler==0.12.1
158
+ rouge==1.0.1
159
+ pydantic-settings==2.13.1
160
+ grpcio==1.78.0
161
+ kiwisolver==1.4.9
162
+ partial-json-parser==0.2.1.1.post7
163
+ Markdown==3.10.2
164
+ httptools==0.7.1
165
+ pydantic-extra-types==2.11.0
166
+ rich-toolkit==0.19.7
167
+ mcp==1.26.0
168
+ pydub==0.25.1
169
+ protobuf==6.33.5
170
+ crcmod==1.7
171
+ typing-inspection==0.4.2
172
+ zstandard==0.25.0
173
+ gradio_client==1.14.0
174
+ tokenizers==0.22.2
175
+ packaging==25.0
176
+ zipp==3.23.0
177
+ py-cpuinfo==9.0.0
178
+ mdurl==0.1.2
179
+ contourpy==1.3.3
180
+ pyzmq==27.1.0
181
+ aiohappyeyeballs==2.6.1
182
+ ninja==1.13.0
183
+ nvidia-cuda-runtime-cu12==12.8.90
184
+ semantic-version==2.10.0
185
+ dacite==1.9.2
186
+ antlr4-python3-runtime==4.9.3
187
+ python-dotenv==1.2.2
188
+ compressed-tensors==0.13.0
189
+ gguf==0.18.0
190
+ anyio==4.12.1
191
+ aiofiles==24.1.0
192
+ nvidia-cusparse-cu12==12.5.8.93
193
+ networkx==3.6.1
194
+ cryptography==46.0.5
195
+ uvicorn==0.41.0
196
+ nvidia-cusolver-cu12==11.7.3.90
197
+ fastapi-cloud-cli==0.14.0
198
+ fastapi-cli==0.0.24
199
+ pydantic_core==2.41.4
200
+ llvmlite==0.44.0
201
+ nvidia-cufile-cu12==1.13.1.3
202
+ rpds-py==0.30.0
203
+ pyparsing==3.3.2
204
+ nvidia-cusparselt-cu12==0.7.1
205
+ annotated-types==0.7.0
206
+ pybase64==1.4.3
207
+ propcache==0.4.1
208
+ prometheus-fastapi-instrumentator==7.1.0
209
+ click==8.3.1
210
+ starlette==0.52.1
211
+ PySocks==1.7.1
212
+ orjson==3.11.7
213
+ dill==0.3.8
214
+ httpx-sse==0.4.3
215
+ nvidia-ml-py==13.590.48
216
+ loguru==0.7.3
217
+ groovy==0.1.2
218
+ cupy-cuda12x==14.0.1
219
+ safetensors==0.7.0
220
+ GitPython==3.1.46
221
+ lm-format-enforcer==0.11.3
222
+ grpcio-reflection==1.78.0
223
+ smmap==5.0.2
224
+ opencv-python-headless==4.13.0.92
225
+ nvidia-nvjitlink-cu12==12.8.93
226
+ einops==0.8.2
227
+ nvidia-cufft-cu12==11.3.3.83
228
+ multidict==6.7.1
229
+ dnspython==2.8.0
230
+ matplotlib==3.10.8
231
+ six==1.17.0
232
+ jmespath==0.10.0
233
+ apache-tvm-ffi==0.1.9
234
+ llguidance==1.3.0
235
+ nvidia-cudnn-cu12==9.10.2.21
236
+ urllib3==2.6.3
237
+ setuptools==80.10.2
238
+ cuda-bindings==13.1.1
239
+ scipy==1.17.1
240
+ huggingface_hub==1.5.0
241
+ torchvision==0.24.1
data/wandb/run-20260306_073928-dpxzw953/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-87-generic-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.12.12",
4
+ "startedAt": "2026-03-06T07:39:28.778950Z",
5
+ "args": [
6
+ "--model",
7
+ "Qwen/Qwen3-1.7B",
8
+ "--use_hf",
9
+ "true",
10
+ "--dataset",
11
+ "/root/code/na/Idea-filtering-training/data/thinking_dataset_v2_swift.jsonl",
12
+ "--train_type",
13
+ "lora",
14
+ "--lora_rank",
15
+ "32",
16
+ "--lora_alpha",
17
+ "64",
18
+ "--lora_dropout",
19
+ "0.05",
20
+ "--target_modules",
21
+ "q_proj",
22
+ "k_proj",
23
+ "v_proj,o_proj",
24
+ "gate_proj",
25
+ "up_proj",
26
+ "down_proj",
27
+ "--per_device_train_batch_size",
28
+ "1",
29
+ "--gradient_accumulation_steps",
30
+ "16",
31
+ "--max_length",
32
+ "2600",
33
+ "--learning_rate",
34
+ "2e-4",
35
+ "--warmup_ratio",
36
+ "0.05",
37
+ "--max_epochs",
38
+ "3",
39
+ "--attn_impl",
40
+ "eager",
41
+ "--output_dir",
42
+ "/root/code/na/Idea-filtering-training/train/output",
43
+ "--save_strategy",
44
+ "steps",
45
+ "--save_steps",
46
+ "1000",
47
+ "--eval_steps",
48
+ "500",
49
+ "--dataloader_num_workers",
50
+ "1",
51
+ "--dataset_num_proc",
52
+ "16",
53
+ "--load_from_cache_file",
54
+ "true",
55
+ "--model_author",
56
+ "swift",
57
+ "--model_name",
58
+ "swift-robot",
59
+ "--loss_scale",
60
+ "ignore_empty_think",
61
+ "--report_to",
62
+ "wandb"
63
+ ],
64
+ "program": "/root/miniconda3/envs/RM/lib/python3.12/site-packages/swift/cli/sft.py",
65
+ "git": {
66
+ "remote": "https://tnnanh1010:@github.com/Drylab-AI/Idea-filtering-training.git",
67
+ "commit": "0dbe0bbbe696b2e49ffefa83123579a73fab61c6"
68
+ },
69
+ "email": "nhatanh10102005@gmail.com",
70
+ "root": "/root/code/na/Idea-filtering-training/data",
71
+ "host": "ubuntu",
72
+ "executable": "/root/miniconda3/envs/RM/bin/python3",
73
+ "cpu_count": 49,
74
+ "cpu_count_logical": 49,
75
+ "gpu": "NVIDIA GeForce RTX 3060",
76
+ "gpu_count": 1,
77
+ "disk": {
78
+ "/": {
79
+ "total": "960312868864",
80
+ "used": "872757264384"
81
+ }
82
+ },
83
+ "memory": {
84
+ "total": "628090474496"
85
+ },
86
+ "gpu_nvidia": [
87
+ {
88
+ "name": "NVIDIA GeForce RTX 3060",
89
+ "memoryTotal": "12884901888",
90
+ "cudaCores": 3584,
91
+ "architecture": "Ampere",
92
+ "uuid": "GPU-f370a18a-d749-7bf1-c827-82726afbb6cf"
93
+ }
94
+ ],
95
+ "cudaVersion": "13.0",
96
+ "writerId": "tjqljpihi24rr4l4o8xbv3328g1gs1n3"
97
+ }
data/wandb/run-20260306_073928-dpxzw953/logs/debug-core.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-06T07:39:28.88021523Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpg_k4o6rr/port-3272902.txt","pid":3272902,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2026-03-06T07:39:28.882449709Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":3272902}
3
+ {"time":"2026-03-06T07:39:28.882308612Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3272902-3286179-1404695856/socket","Net":"unix"}}
4
+ {"time":"2026-03-06T07:39:29.020770654Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2026-03-06T07:39:29.036363468Z","level":"INFO","msg":"handleInformInit: received","streamId":"dpxzw953","id":"1(@)"}
6
+ {"time":"2026-03-06T07:39:29.255684067Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"dpxzw953","id":"1(@)"}
7
+ {"time":"2026-03-06T07:39:35.284387334Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"cetk0o1tw0pu"}
8
+ {"time":"2026-03-06T08:45:03.956181644Z","level":"INFO","msg":"server: parent process exited, terminating service process"}
data/wandb/run-20260306_073928-dpxzw953/logs/debug-internal.log ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {"time":"2026-03-06T07:39:29.036940297Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-03-06T07:39:29.254845747Z","level":"INFO","msg":"stream: created new stream","id":"dpxzw953"}
3
+ {"time":"2026-03-06T07:39:29.255262411Z","level":"INFO","msg":"handler: started","stream_id":"dpxzw953"}
4
+ {"time":"2026-03-06T07:39:29.255639332Z","level":"INFO","msg":"stream: started","id":"dpxzw953"}
5
+ {"time":"2026-03-06T07:39:29.255751228Z","level":"INFO","msg":"writer: started","stream_id":"dpxzw953"}
6
+ {"time":"2026-03-06T07:39:29.255761468Z","level":"INFO","msg":"sender: started","stream_id":"dpxzw953"}
data/wandb/run-20260306_073928-dpxzw953/logs/debug.log ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-06 07:39:28,783 INFO MainThread:3272902 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-03-06 07:39:28,783 INFO MainThread:3272902 [wandb_setup.py:_flush():81] Configure stats pid to 3272902
3
+ 2026-03-06 07:39:28,784 INFO MainThread:3272902 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-03-06 07:39:28,784 INFO MainThread:3272902 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /root/code/na/Idea-filtering-training/data/wandb/run-20260306_073928-dpxzw953/logs/debug.log
5
+ 2026-03-06 07:39:28,784 INFO MainThread:3272902 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /root/code/na/Idea-filtering-training/data/wandb/run-20260306_073928-dpxzw953/logs/debug-internal.log
6
+ 2026-03-06 07:39:28,784 INFO MainThread:3272902 [wandb_init.py:init():844] calling init triggers
7
+ 2026-03-06 07:39:28,784 INFO MainThread:3272902 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'_wandb': {}}
9
+ 2026-03-06 07:39:28,785 INFO MainThread:3272902 [wandb_init.py:init():892] starting backend
10
+ 2026-03-06 07:39:29,020 INFO MainThread:3272902 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-03-06 07:39:29,026 INFO MainThread:3272902 [wandb_init.py:init():903] backend started and connected
12
+ 2026-03-06 07:39:29,030 INFO MainThread:3272902 [wandb_init.py:init():973] updated telemetry
13
+ 2026-03-06 07:39:29,039 INFO MainThread:3272902 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-03-06 07:39:30,085 INFO MainThread:3272902 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-03-06 07:39:30,275 INFO MainThread:3272902 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-03-06 07:39:30,275 INFO MainThread:3272902 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-03-06 07:39:30,275 INFO MainThread:3272902 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-03-06 07:39:30,275 INFO MainThread:3272902 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-03-06 07:39:30,280 INFO MainThread:3272902 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-03-06 07:39:30,282 INFO MainThread:3272902 [wandb_run.py:_config_callback():1403] config_cb None None {'peft_config': {'default': {'task_type': 'CAUSAL_LM', 'peft_type': 'LORA', 'auto_mapping': None, 'peft_version': '0.18.1', 'base_model_name_or_path': '/root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e', 'revision': None, 'inference_mode': False, 'r': 32, 'target_modules': ['v_proj,o_proj', 'gate_proj', 'up_proj', 'q_proj', 'k_proj', 'down_proj'], 'exclude_modules': None, 'lora_alpha': 64, 'lora_dropout': 0.05, 'fan_in_fan_out': False, 'bias': 'none', 'use_rslora': False, 'modules_to_save': [], 'init_lora_weights': True, 'layers_to_transform': None, 'layers_pattern': None, 'rank_pattern': {}, 'alpha_pattern': {}, 'megatron_config': None, 'megatron_core': 'megatron.core', 'trainable_token_indices': None, 'loftq_config': {}, 'eva_config': None, 'corda_config': None, 'use_dora': False, 'alora_invocation_tokens': None, 'use_qalora': False, 'qalora_group_size': 16, 'layer_replication': None, 'runtime_config': {'ephemeral_gpu_offload': False}, 'lora_bias': False, 'target_parameters': None, 'arrow_config': None, 'ensure_weight_tying': False, 'lora_dtype': None, 'lorap_lr_ratio': None, 'lorap_emb_lr': 1e-06}}, 'vocab_size': 151936, 'max_position_embeddings': 40960, 'hidden_size': 2048, 'intermediate_size': 6144, 'num_hidden_layers': 28, 'num_attention_heads': 16, 'use_sliding_window': False, 'sliding_window': None, 'max_window_layers': 28, 'num_key_value_heads': 8, 'head_dim': 128, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-06, 'use_cache': False, 'attention_bias': False, 'attention_dropout': 0.0, 'layer_types': ['full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention'], 'pad_token_id': 151643, 'bos_token_id': None, 'eos_token_id': 151645, 'tie_word_embeddings': True, 'rope_parameters': {'rope_theta': 1000000, 'rope_type': 'default'}, 'return_dict': True, 'output_hidden_states': False, 'dtype': 'bfloat16', 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'architectures': ['Qwen3ForCausalLM'], 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'problem_type': None, '_name_or_path': '/root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e', 'transformers_version': '5.3.0', 'model_type': 'qwen3', 'output_attentions': False, 'output_dir': '/root/code/na/Idea-filtering-training/train/output/v1-20260306-073834', 'per_device_train_batch_size': 1, 'num_train_epochs': 3.0, 'max_steps': -1, 'learning_rate': 0.0002, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': None, 'warmup_steps': 0.05, 'optim': 'adamw_torch_fused', 'optim_args': None, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.95, 'adam_epsilon': 1e-08, 'optim_target_modules': None, 'gradient_accumulation_steps': 16, 'average_tokens_across_devices': True, 'max_grad_norm': 1.0, 'label_smoothing_factor': 0.0, 'bf16': True, 'fp16': False, 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'use_liger_kernel': False, 'liger_kernel_config': None, 'neftune_noise_alpha': None, 'torch_empty_cache_steps': None, 'auto_find_batch_size': False, 'logging_strategy': 'steps', 'logging_steps': 5, 'logging_first_step': True, 'log_on_each_node': True, 'logging_nan_inf_filter': True, 'include_num_input_tokens_seen': 'no', 'log_level': 'passive', 'log_level_replica': 'warning', 'disable_tqdm': False, 'report_to': ['wandb'], 'run_name': '/root/code/na/Idea-filtering-training/train/output/v1-20260306-073834', 'project': 'huggingface', 'trackio_space_id': 'trackio', 'eval_strategy': 'no', 'eval_steps': 500.0, 'eval_delay': 0, 'per_device_eval_batch_size': 1, 'prediction_loss_only': False, 'eval_on_start': False, 'eval_do_concat_batches': True, 'eval_use_gather_object': False, 'eval_accumulation_steps': None, 'include_for_metrics': [], 'batch_eval_metrics': False, 'save_only_model': False, 'save_strategy': 'steps', 'save_steps': 1000, 'save_on_each_node': False, 'save_total_limit': None, 'enable_jit_checkpoint': False, 'push_to_hub': False, 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_always_push': False, 'hub_revision': None, 'load_best_model_at_end': False, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'restore_callback_states_from_checkpoint': False, 'full_determinism': False, 'seed': 42, 'data_seed': 42, 'use_cpu': False, 'accelerator_config': {'split_batches': False, 'dispatch_batches': False, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'parallelism_config': None, 'dataloader_drop_last': False, 'dataloader_num_workers': 1, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'dataloader_prefetch_factor': 2, 'remove_unused_columns': False, 'label_names': None, 'train_sampling_strategy': 'random', 'length_column_name': 'length', 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'ddp_backend': None, 'ddp_timeout': 18000000, 'fsdp': [], 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'deepspeed': None, 'debug': [], 'skip_memory_metrics': True, 'do_train': False, 'do_eval': False, 'do_predict': False, 'resume_from_checkpoint': None, 'warmup_ratio': 0.05, 'logging_dir': '/root/code/na/Idea-filtering-training/train/output/v1-20260306-073834/runs', 'local_rank': -1, 'sortish_sampler': False, 'predict_with_generate': False, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None, 'tuner_backend': 'peft', 'vit_gradient_checkpointing': True, 'router_aux_loss_coef': 0.0, 'enable_dft_loss': False, 'enable_channel_loss': False, 'check_model': True, 'acc_strategy': 'token', 'train_dataloader_shuffle': True, 'group_by_length': False, 'max_epochs': 3, 'aligner_lr': None, 'vit_lr': None, 'use_logits_to_keep': None, 'ds3_gather_for_generation': True, 'resume_only_model': False, 'optimizer': None, 'loss_type': None, 'eval_metric': None, 'callbacks': [], 'early_stop_interval': None, 'eval_use_evalscope': False, 'eval_dataset': [], 'eval_dataset_args': None, 'eval_limit': None, 'eval_generation_config': None, 'extra_eval_args': None, 'tuner_type': 'lora', 'use_galore': False, 'galore_target_modules': None, 'galore_rank': 128, 'galore_update_proj_gap': 50, 'galore_scale': 1.0, 'galore_proj_type': 'std', 'galore_optim_per_parameter': False, 'galore_with_embedding': False, 'galore_quantization': False, 'galore_proj_quant': False, 'galore_proj_bits': 4, 'galore_proj_group_size': 256, 'galore_cos_threshold': 0.4, 'galore_gamma_proj': 2, 'galore_queue_size': 5, 'lisa_activated_layers': 0, 'lisa_step_interval': 20, 'use_flash_ckpt': False}
21
+ 2026-03-06 07:39:30,290 INFO MainThread:3272902 [wandb_config.py:__setitem__():155] [no run ID] config set model/num_parameters = 2060182528 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7ffe8032fd10>>
22
+ 2026-03-06 07:39:30,290 INFO MainThread:3272902 [wandb_run.py:_config_callback():1403] config_cb model/num_parameters 2060182528 None
data/wandb/run-20260306_073928-dpxzw953/run-dpxzw953.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2472aae39445ec308f7f20624eb857ffabd422e57c2005b739a860f3cf49d028
3
+ size 327680
data/wandb/run-20260306_105746-8846dbl6/files/output.log ADDED
The diff for this file is too large to render. See raw diff
 
data/wandb/run-20260306_105746-8846dbl6/files/requirements.txt ADDED
@@ -0,0 +1,243 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ psutil==7.2.2
2
+ colorama==0.4.6
3
+ trl==0.28.0
4
+ pyarrow==23.0.1
5
+ msgpack==1.1.2
6
+ astor==0.8.1
7
+ cffi==2.0.0
8
+ qwen-vl-utils==0.0.14
9
+ aiosignal==1.4.0
10
+ sse-starlette==3.3.2
11
+ setproctitle==1.3.7
12
+ ms_swift==4.0.0
13
+ mistral_common==1.9.1
14
+ aliyun-python-sdk-kms==2.16.5
15
+ nvidia-cublas-cu12==12.8.4.1
16
+ markdown-it-py==4.0.0
17
+ cachetools==7.0.3
18
+ nltk==3.9.3
19
+ cuda-pathfinder==1.4.0
20
+ sentencepiece==0.2.1
21
+ pycryptodome==3.23.0
22
+ hf-xet==1.3.2
23
+ multiprocess==0.70.16
24
+ wandb==0.25.0
25
+ fastar==0.8.0
26
+ uv==0.10.8
27
+ interegular==0.3.3
28
+ rich==14.3.3
29
+ modelscope==1.34.0
30
+ websockets==15.0.1
31
+ binpacking==2.0.1
32
+ seaborn==0.13.2
33
+ uvloop==0.22.1
34
+ mpmath==1.3.0
35
+ certifi==2026.2.25
36
+ vllm==0.16.0
37
+ supervisor==4.3.0
38
+ flashinfer-python==0.6.3
39
+ cbor2==5.8.0
40
+ msgspec==0.20.0
41
+ threadpoolctl==3.6.0
42
+ watchfiles==1.1.1
43
+ python-multipart==0.0.22
44
+ peft==0.18.1
45
+ httpx==0.28.1
46
+ pillow==11.3.0
47
+ wheel==0.46.3
48
+ MarkupSafe==3.0.3
49
+ ruff==0.15.4
50
+ frozenlist==1.8.0
51
+ xxhash==3.6.0
52
+ psutil==7.2.2
53
+ pytz==2026.1.post1
54
+ platformdirs==4.9.2
55
+ joblib==1.5.3
56
+ transformers==5.3.0
57
+ scikit-learn==1.8.0
58
+ nvidia-cuda-nvrtc-cu12==12.8.93
59
+ nvidia-cutlass-dsl-libs-base==4.4.1
60
+ openai-harmony==0.0.8
61
+ safehttpx==0.1.7
62
+ gdown==5.2.1
63
+ diskcache==5.6.3
64
+ datasets==3.6.0
65
+ tqdm==4.67.3
66
+ openai==2.24.0
67
+ nvidia-nccl-cu12==2.27.5
68
+ sympy==1.14.0
69
+ tabulate==0.10.0
70
+ Werkzeug==3.1.6
71
+ sentry-sdk==2.54.0
72
+ numba==0.61.2
73
+ attrdict==2.0.1
74
+ referencing==0.37.0
75
+ cuda-python==13.1.1
76
+ jsonschema-specifications==2025.9.1
77
+ nvidia-nvtx-cu12==12.8.90
78
+ email-validator==2.3.0
79
+ PyYAML==6.0.3
80
+ requests==2.32.5
81
+ cloudpickle==3.1.2
82
+ xgrammar==0.1.29
83
+ importlib_metadata==8.7.1
84
+ prometheus_client==0.24.1
85
+ h11==0.16.0
86
+ json_repair==0.58.3
87
+ typer-slim==0.24.0
88
+ pandas==2.3.3
89
+ absl-py==2.4.0
90
+ tomlkit==0.13.3
91
+ nvidia-cuda-cupti-cu12==12.8.90
92
+ gitdb==4.0.12
93
+ tiktoken==0.12.0
94
+ soupsieve==2.8.3
95
+ cpm-kernels==1.0.11
96
+ tensorboard-data-server==0.7.2
97
+ sniffio==1.3.1
98
+ sortedcontainers==2.4.0
99
+ pycountry==26.2.16
100
+ triton==3.5.1
101
+ nvidia-curand-cu12==10.3.9.90
102
+ beautifulsoup4==4.14.3
103
+ aiohttp==3.13.3
104
+ nvidia-cutlass-dsl==4.4.1
105
+ transformers-stream-generator==0.0.5
106
+ gradio==5.50.0
107
+ lark==1.2.2
108
+ pycparser==3.0
109
+ yarl==1.23.0
110
+ outlines_core==0.2.11
111
+ anthropic==0.84.0
112
+ jsonschema==4.26.0
113
+ Pygments==2.19.2
114
+ nvidia-nvshmem-cu12==3.3.20
115
+ ffmpy==1.0.0
116
+ brotli==1.2.0
117
+ idna==3.11
118
+ Jinja2==3.1.6
119
+ pip==26.0.1
120
+ docstring_parser==0.17.0
121
+ jieba==0.42.1
122
+ typing_extensions==4.15.0
123
+ torch==2.9.1
124
+ attrs==25.4.0
125
+ filelock==3.25.0
126
+ fonttools==4.61.1
127
+ accelerate==1.12.0
128
+ httpcore==1.0.9
129
+ ijson==3.5.0
130
+ model-hosting-container-standards==0.1.13
131
+ PyJWT==2.11.0
132
+ charset-normalizer==3.4.4
133
+ regex==2026.2.28
134
+ simplejson==3.20.2
135
+ tensorboard==2.20.0
136
+ python-json-logger==4.0.0
137
+ python-dateutil==2.9.0.post0
138
+ nvidia-cudnn-frontend==1.18.0
139
+ tzdata==2025.3
140
+ typer==0.24.1
141
+ pydantic==2.12.3
142
+ jiter==0.13.0
143
+ annotated-doc==0.0.4
144
+ aliyun-python-sdk-core==2.16.0
145
+ oss2==2.19.1
146
+ fastapi==0.135.1
147
+ shellingham==1.5.4
148
+ addict==2.4.0
149
+ omegaconf==2.3.0
150
+ distro==1.9.0
151
+ depyf==0.20.0
152
+ numpy==2.2.6
153
+ rignore==0.7.6
154
+ torchaudio==2.9.1
155
+ ray==2.54.0
156
+ blake3==1.0.8
157
+ fsspec==2025.3.0
158
+ cycler==0.12.1
159
+ rouge==1.0.1
160
+ pydantic-settings==2.13.1
161
+ grpcio==1.78.0
162
+ kiwisolver==1.4.9
163
+ partial-json-parser==0.2.1.1.post7
164
+ Markdown==3.10.2
165
+ httptools==0.7.1
166
+ pydantic-extra-types==2.11.0
167
+ rich-toolkit==0.19.7
168
+ mcp==1.26.0
169
+ pydub==0.25.1
170
+ protobuf==6.33.5
171
+ crcmod==1.7
172
+ typing-inspection==0.4.2
173
+ zstandard==0.25.0
174
+ gradio_client==1.14.0
175
+ tokenizers==0.22.2
176
+ packaging==25.0
177
+ zipp==3.23.0
178
+ py-cpuinfo==9.0.0
179
+ mdurl==0.1.2
180
+ contourpy==1.3.3
181
+ pyzmq==27.1.0
182
+ aiohappyeyeballs==2.6.1
183
+ ninja==1.13.0
184
+ nvidia-cuda-runtime-cu12==12.8.90
185
+ semantic-version==2.10.0
186
+ dacite==1.9.2
187
+ antlr4-python3-runtime==4.9.3
188
+ python-dotenv==1.2.2
189
+ compressed-tensors==0.13.0
190
+ gguf==0.18.0
191
+ anyio==4.12.1
192
+ aiofiles==24.1.0
193
+ nvidia-cusparse-cu12==12.5.8.93
194
+ networkx==3.6.1
195
+ cryptography==46.0.5
196
+ uvicorn==0.41.0
197
+ nvidia-cusolver-cu12==11.7.3.90
198
+ fastapi-cloud-cli==0.14.0
199
+ fastapi-cli==0.0.24
200
+ pydantic_core==2.41.4
201
+ llvmlite==0.44.0
202
+ nvidia-cufile-cu12==1.13.1.3
203
+ rpds-py==0.30.0
204
+ pyparsing==3.3.2
205
+ nvidia-cusparselt-cu12==0.7.1
206
+ annotated-types==0.7.0
207
+ pybase64==1.4.3
208
+ propcache==0.4.1
209
+ prometheus-fastapi-instrumentator==7.1.0
210
+ click==8.3.1
211
+ starlette==0.52.1
212
+ PySocks==1.7.1
213
+ orjson==3.11.7
214
+ dill==0.3.8
215
+ httpx-sse==0.4.3
216
+ nvidia-ml-py==13.590.48
217
+ loguru==0.7.3
218
+ groovy==0.1.2
219
+ cupy-cuda12x==14.0.1
220
+ safetensors==0.7.0
221
+ GitPython==3.1.46
222
+ lm-format-enforcer==0.11.3
223
+ grpcio-reflection==1.78.0
224
+ smmap==5.0.2
225
+ opencv-python-headless==4.13.0.92
226
+ nvidia-nvjitlink-cu12==12.8.93
227
+ einops==0.8.2
228
+ nvidia-cufft-cu12==11.3.3.83
229
+ multidict==6.7.1
230
+ dnspython==2.8.0
231
+ matplotlib==3.10.8
232
+ six==1.17.0
233
+ jmespath==0.10.0
234
+ apache-tvm-ffi==0.1.9
235
+ llguidance==1.3.0
236
+ nvidia-cudnn-cu12==9.10.2.21
237
+ urllib3==2.6.3
238
+ setuptools==80.10.2
239
+ cuda-bindings==13.1.1
240
+ av==16.1.0
241
+ scipy==1.17.1
242
+ huggingface_hub==1.5.0
243
+ torchvision==0.24.1
data/wandb/run-20260306_105746-8846dbl6/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-87-generic-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.12.12",
4
+ "startedAt": "2026-03-06T10:57:46.231329Z",
5
+ "args": [
6
+ "--model",
7
+ "Qwen/Qwen3-1.7B",
8
+ "--use_hf",
9
+ "true",
10
+ "--dataset",
11
+ "/root/code/na/Idea-filtering-training/data/thinking_dataset_v2_swift.jsonl",
12
+ "--train_type",
13
+ "lora",
14
+ "--lora_rank",
15
+ "32",
16
+ "--lora_alpha",
17
+ "64",
18
+ "--lora_dropout",
19
+ "0.05",
20
+ "--target_modules",
21
+ "q_proj",
22
+ "k_proj",
23
+ "v_proj,o_proj",
24
+ "gate_proj",
25
+ "up_proj",
26
+ "down_proj",
27
+ "--per_device_train_batch_size",
28
+ "1",
29
+ "--gradient_accumulation_steps",
30
+ "16",
31
+ "--max_length",
32
+ "2600",
33
+ "--learning_rate",
34
+ "2e-4",
35
+ "--warmup_ratio",
36
+ "0.05",
37
+ "--max_epochs",
38
+ "3",
39
+ "--attn_impl",
40
+ "eager",
41
+ "--output_dir",
42
+ "/root/code/na/Idea-filtering-training/train/output",
43
+ "--save_strategy",
44
+ "steps",
45
+ "--save_steps",
46
+ "1000",
47
+ "--eval_steps",
48
+ "500",
49
+ "--dataloader_num_workers",
50
+ "1",
51
+ "--dataset_num_proc",
52
+ "16",
53
+ "--load_from_cache_file",
54
+ "true",
55
+ "--model_author",
56
+ "swift",
57
+ "--model_name",
58
+ "swift-robot",
59
+ "--loss_scale",
60
+ "ignore_empty_think",
61
+ "--report_to",
62
+ "wandb"
63
+ ],
64
+ "program": "/root/miniconda3/envs/RM/lib/python3.12/site-packages/swift/cli/sft.py",
65
+ "git": {
66
+ "remote": "https://tnnanh1010:@github.com/Drylab-AI/Idea-filtering-training.git",
67
+ "commit": "0dbe0bbbe696b2e49ffefa83123579a73fab61c6"
68
+ },
69
+ "email": "nhatanh10102005@gmail.com",
70
+ "root": "/root/code/na/Idea-filtering-training/data",
71
+ "host": "ubuntu",
72
+ "executable": "/root/miniconda3/envs/RM/bin/python3",
73
+ "cpu_count": 49,
74
+ "cpu_count_logical": 49,
75
+ "gpu": "NVIDIA GeForce RTX 3060",
76
+ "gpu_count": 1,
77
+ "disk": {
78
+ "/": {
79
+ "total": "960312868864",
80
+ "used": "892185260032"
81
+ }
82
+ },
83
+ "memory": {
84
+ "total": "628090474496"
85
+ },
86
+ "gpu_nvidia": [
87
+ {
88
+ "name": "NVIDIA GeForce RTX 3060",
89
+ "memoryTotal": "12884901888",
90
+ "cudaCores": 3584,
91
+ "architecture": "Ampere",
92
+ "uuid": "GPU-f370a18a-d749-7bf1-c827-82726afbb6cf"
93
+ }
94
+ ],
95
+ "cudaVersion": "13.0",
96
+ "writerId": "pnelgqh2vvvgt4d15fvpxk9xawojic2y"
97
+ }
data/wandb/run-20260306_105746-8846dbl6/logs/debug-core.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-06T10:57:46.314363437Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpg5b71rb2/port-1654770.txt","pid":1654770,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2026-03-06T10:57:46.316418659Z","level":"INFO","msg":"server: will exit if parent process dies","ppid":1654770}
3
+ {"time":"2026-03-06T10:57:46.316317152Z","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-1654770-1660166-284857756/socket","Net":"unix"}}
4
+ {"time":"2026-03-06T10:57:46.467869627Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2026-03-06T10:57:46.485944099Z","level":"INFO","msg":"handleInformInit: received","streamId":"8846dbl6","id":"1(@)"}
6
+ {"time":"2026-03-06T10:57:46.706399787Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"8846dbl6","id":"1(@)"}
7
+ {"time":"2026-03-06T10:57:52.301492638Z","level":"INFO","msg":"connection: cancelling request","id":"1(@)","requestId":"sijs3l0vpag5"}
8
+ {"time":"2026-03-07T02:55:05.849650733Z","level":"INFO","msg":"server: parent process exited, terminating service process"}
data/wandb/run-20260306_105746-8846dbl6/logs/debug-internal.log ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-03-06T10:57:46.486394548Z","level":"INFO","msg":"stream: starting","core version":"0.25.0"}
2
+ {"time":"2026-03-06T10:57:46.705840771Z","level":"INFO","msg":"stream: created new stream","id":"8846dbl6"}
3
+ {"time":"2026-03-06T10:57:46.706184448Z","level":"INFO","msg":"handler: started","stream_id":"8846dbl6"}
4
+ {"time":"2026-03-06T10:57:46.706376796Z","level":"INFO","msg":"stream: started","id":"8846dbl6"}
5
+ {"time":"2026-03-06T10:57:46.70647031Z","level":"INFO","msg":"writer: started","stream_id":"8846dbl6"}
6
+ {"time":"2026-03-06T10:57:46.706495415Z","level":"INFO","msg":"sender: started","stream_id":"8846dbl6"}
7
+ {"time":"2026-03-06T19:23:17.154701842Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/nhatanh10102005-vinuniversity/ms-swift/8846dbl6/file_stream","body":"\n<html><head>\n<meta http-equiv=\"content-type\" content=\"text/html;charset=utf-8\">\n<title>502 Server Error</title>\n</head>\n<body text=#000000 bgcolor=#ffffff>\n<h1>Error: Server Error</h1>\n<h2>The server encountered a temporary error and could not complete your request.<p>Please try again in 30 seconds.</h2>\n<h2></h2>\n</body></html>\n"}
8
+ {"time":"2026-03-07T01:23:02.15976294Z","level":"INFO","msg":"api: retrying HTTP error","status":502,"url":"https://api.wandb.ai/files/nhatanh10102005-vinuniversity/ms-swift/8846dbl6/file_stream","body":"\n<html><head>\n<meta http-equiv=\"content-type\" content=\"text/html;charset=utf-8\">\n<title>502 Server Error</title>\n</head>\n<body text=#000000 bgcolor=#ffffff>\n<h1>Error: Server Error</h1>\n<h2>The server encountered a temporary error and could not complete your request.<p>Please try again in 30 seconds.</h2>\n<h2></h2>\n</body></html>\n"}
9
+ {"time":"2026-03-07T02:55:05.824285344Z","level":"INFO","msg":"flowcontrol: backed up, offloading to disk","recordNumber":15580}
10
+ {"time":"2026-03-07T02:55:05.824494952Z","level":"INFO","msg":"flowcontrol: unblocked","totalOffloaded":1}
11
+ {"time":"2026-03-07T02:55:05.825190016Z","level":"INFO","msg":"flowcontrol: backed up, offloading to disk","recordNumber":15613}
12
+ {"time":"2026-03-07T02:55:05.840550203Z","level":"INFO","msg":"flowcontrol: unblocked","totalOffloaded":904}
data/wandb/run-20260306_105746-8846dbl6/logs/debug.log ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-03-06 10:57:46,235 INFO MainThread:1654770 [wandb_setup.py:_flush():81] Current SDK version is 0.25.0
2
+ 2026-03-06 10:57:46,236 INFO MainThread:1654770 [wandb_setup.py:_flush():81] Configure stats pid to 1654770
3
+ 2026-03-06 10:57:46,236 INFO MainThread:1654770 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-03-06 10:57:46,236 INFO MainThread:1654770 [wandb_init.py:setup_run_log_directory():717] Logging user logs to /root/code/na/Idea-filtering-training/data/wandb/run-20260306_105746-8846dbl6/logs/debug.log
5
+ 2026-03-06 10:57:46,236 INFO MainThread:1654770 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to /root/code/na/Idea-filtering-training/data/wandb/run-20260306_105746-8846dbl6/logs/debug-internal.log
6
+ 2026-03-06 10:57:46,236 INFO MainThread:1654770 [wandb_init.py:init():844] calling init triggers
7
+ 2026-03-06 10:57:46,237 INFO MainThread:1654770 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'_wandb': {}}
9
+ 2026-03-06 10:57:46,237 INFO MainThread:1654770 [wandb_init.py:init():892] starting backend
10
+ 2026-03-06 10:57:46,468 INFO MainThread:1654770 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-03-06 10:57:46,475 INFO MainThread:1654770 [wandb_init.py:init():903] backend started and connected
12
+ 2026-03-06 10:57:46,480 INFO MainThread:1654770 [wandb_init.py:init():973] updated telemetry
13
+ 2026-03-06 10:57:46,491 INFO MainThread:1654770 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-03-06 10:57:47,094 INFO MainThread:1654770 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-03-06 10:57:47,290 INFO MainThread:1654770 [wandb_run.py:_console_start():2524] atexit reg
16
+ 2026-03-06 10:57:47,290 INFO MainThread:1654770 [wandb_run.py:_redirect():2373] redirect: wrap_raw
17
+ 2026-03-06 10:57:47,290 INFO MainThread:1654770 [wandb_run.py:_redirect():2442] Wrapping output streams.
18
+ 2026-03-06 10:57:47,290 INFO MainThread:1654770 [wandb_run.py:_redirect():2465] Redirects installed.
19
+ 2026-03-06 10:57:47,295 INFO MainThread:1654770 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-03-06 10:57:47,296 INFO MainThread:1654770 [wandb_run.py:_config_callback():1403] config_cb None None {'peft_config': {'default': {'task_type': 'CAUSAL_LM', 'peft_type': 'LORA', 'auto_mapping': None, 'peft_version': '0.18.1', 'base_model_name_or_path': '/root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e', 'revision': None, 'inference_mode': False, 'r': 32, 'target_modules': ['gate_proj', 'up_proj', 'v_proj,o_proj', 'down_proj', 'k_proj', 'q_proj'], 'exclude_modules': None, 'lora_alpha': 64, 'lora_dropout': 0.05, 'fan_in_fan_out': False, 'bias': 'none', 'use_rslora': False, 'modules_to_save': [], 'init_lora_weights': True, 'layers_to_transform': None, 'layers_pattern': None, 'rank_pattern': {}, 'alpha_pattern': {}, 'megatron_config': None, 'megatron_core': 'megatron.core', 'trainable_token_indices': None, 'loftq_config': {}, 'eva_config': None, 'corda_config': None, 'use_dora': False, 'alora_invocation_tokens': None, 'use_qalora': False, 'qalora_group_size': 16, 'layer_replication': None, 'runtime_config': {'ephemeral_gpu_offload': False}, 'lora_bias': False, 'target_parameters': None, 'arrow_config': None, 'ensure_weight_tying': False, 'lora_dtype': None, 'lorap_lr_ratio': None, 'lorap_emb_lr': 1e-06}}, 'vocab_size': 151936, 'max_position_embeddings': 40960, 'hidden_size': 2048, 'intermediate_size': 6144, 'num_hidden_layers': 28, 'num_attention_heads': 16, 'use_sliding_window': False, 'sliding_window': None, 'max_window_layers': 28, 'num_key_value_heads': 8, 'head_dim': 128, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-06, 'use_cache': False, 'attention_bias': False, 'attention_dropout': 0.0, 'layer_types': ['full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention', 'full_attention'], 'pad_token_id': 151643, 'bos_token_id': None, 'eos_token_id': 151645, 'tie_word_embeddings': True, 'rope_parameters': {'rope_theta': 1000000, 'rope_type': 'default'}, 'return_dict': True, 'output_hidden_states': False, 'dtype': 'bfloat16', 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'architectures': ['Qwen3ForCausalLM'], 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'problem_type': None, '_name_or_path': '/root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e', 'transformers_version': '5.3.0', 'model_type': 'qwen3', 'output_attentions': False, 'output_dir': '/root/code/na/Idea-filtering-training/train/output/v5-20260306-105729', 'per_device_train_batch_size': 1, 'num_train_epochs': 3.0, 'max_steps': -1, 'learning_rate': 0.0002, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': None, 'warmup_steps': 0.05, 'optim': 'adamw_torch_fused', 'optim_args': None, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.95, 'adam_epsilon': 1e-08, 'optim_target_modules': None, 'gradient_accumulation_steps': 16, 'average_tokens_across_devices': True, 'max_grad_norm': 1.0, 'label_smoothing_factor': 0.0, 'bf16': True, 'fp16': False, 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'use_liger_kernel': False, 'liger_kernel_config': None, 'neftune_noise_alpha': None, 'torch_empty_cache_steps': None, 'auto_find_batch_size': False, 'logging_strategy': 'steps', 'logging_steps': 5, 'logging_first_step': True, 'log_on_each_node': True, 'logging_nan_inf_filter': True, 'include_num_input_tokens_seen': 'no', 'log_level': 'passive', 'log_level_replica': 'warning', 'disable_tqdm': False, 'report_to': ['wandb'], 'run_name': '/root/code/na/Idea-filtering-training/train/output/v5-20260306-105729', 'project': 'huggingface', 'trackio_space_id': 'trackio', 'eval_strategy': 'no', 'eval_steps': 500.0, 'eval_delay': 0, 'per_device_eval_batch_size': 1, 'prediction_loss_only': False, 'eval_on_start': False, 'eval_do_concat_batches': True, 'eval_use_gather_object': False, 'eval_accumulation_steps': None, 'include_for_metrics': [], 'batch_eval_metrics': False, 'save_only_model': False, 'save_strategy': 'steps', 'save_steps': 1000, 'save_on_each_node': False, 'save_total_limit': None, 'enable_jit_checkpoint': False, 'push_to_hub': False, 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_always_push': False, 'hub_revision': None, 'load_best_model_at_end': False, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'restore_callback_states_from_checkpoint': False, 'full_determinism': False, 'seed': 42, 'data_seed': 42, 'use_cpu': False, 'accelerator_config': {'split_batches': False, 'dispatch_batches': False, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'parallelism_config': None, 'dataloader_drop_last': False, 'dataloader_num_workers': 1, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'dataloader_prefetch_factor': 2, 'remove_unused_columns': False, 'label_names': None, 'train_sampling_strategy': 'random', 'length_column_name': 'length', 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'ddp_backend': None, 'ddp_timeout': 18000000, 'fsdp': [], 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'deepspeed': None, 'debug': [], 'skip_memory_metrics': True, 'do_train': False, 'do_eval': False, 'do_predict': False, 'resume_from_checkpoint': None, 'warmup_ratio': 0.05, 'logging_dir': '/root/code/na/Idea-filtering-training/train/output/v5-20260306-105729/runs', 'local_rank': -1, 'sortish_sampler': False, 'predict_with_generate': False, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None, 'tuner_backend': 'peft', 'vit_gradient_checkpointing': True, 'router_aux_loss_coef': 0.0, 'enable_dft_loss': False, 'enable_channel_loss': False, 'check_model': True, 'acc_strategy': 'token', 'train_dataloader_shuffle': True, 'group_by_length': False, 'max_epochs': 3, 'aligner_lr': None, 'vit_lr': None, 'use_logits_to_keep': None, 'ds3_gather_for_generation': True, 'resume_only_model': False, 'optimizer': None, 'loss_type': None, 'eval_metric': None, 'callbacks': [], 'early_stop_interval': None, 'eval_use_evalscope': False, 'eval_dataset': [], 'eval_dataset_args': None, 'eval_limit': None, 'eval_generation_config': None, 'extra_eval_args': None, 'tuner_type': 'lora', 'use_galore': False, 'galore_target_modules': None, 'galore_rank': 128, 'galore_update_proj_gap': 50, 'galore_scale': 1.0, 'galore_proj_type': 'std', 'galore_optim_per_parameter': False, 'galore_with_embedding': False, 'galore_quantization': False, 'galore_proj_quant': False, 'galore_proj_bits': 4, 'galore_proj_group_size': 256, 'galore_cos_threshold': 0.4, 'galore_gamma_proj': 2, 'galore_queue_size': 5, 'lisa_activated_layers': 0, 'lisa_step_interval': 20, 'use_flash_ckpt': False}
21
+ 2026-03-06 10:57:47,308 INFO MainThread:1654770 [wandb_config.py:__setitem__():155] [no run ID] config set model/num_parameters = 2060182528 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7ffc51dc5100>>
22
+ 2026-03-06 10:57:47,308 INFO MainThread:1654770 [wandb_run.py:_config_callback():1403] config_cb model/num_parameters 2060182528 None
data/wandb/run-20260306_105746-8846dbl6/run-8846dbl6.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e0d044c0d3dfab768ea82c3b508f38854050786dec53711998fe38a11b2fb26
3
+ size 5136760
output/Qwen3-1.7B/v0-20260304-031440/args.json ADDED
@@ -0,0 +1,346 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_dir": "/root/code/na/Idea-filtering-training/output/Qwen3-1.7B/v0-20260304-031440",
3
+ "per_device_train_batch_size": 1,
4
+ "num_train_epochs": 3.0,
5
+ "max_steps": -1,
6
+ "learning_rate": 0.0001,
7
+ "lr_scheduler_type": "cosine",
8
+ "lr_scheduler_kwargs": null,
9
+ "warmup_steps": 0,
10
+ "optim": "adamw_torch_fused",
11
+ "optim_args": null,
12
+ "weight_decay": 0.1,
13
+ "adam_beta1": 0.9,
14
+ "adam_beta2": 0.95,
15
+ "adam_epsilon": 1e-08,
16
+ "optim_target_modules": null,
17
+ "gradient_accumulation_steps": null,
18
+ "average_tokens_across_devices": true,
19
+ "max_grad_norm": 1.0,
20
+ "label_smoothing_factor": 0.0,
21
+ "bf16": true,
22
+ "fp16": false,
23
+ "bf16_full_eval": false,
24
+ "fp16_full_eval": false,
25
+ "tf32": null,
26
+ "gradient_checkpointing": true,
27
+ "gradient_checkpointing_kwargs": null,
28
+ "torch_compile": false,
29
+ "torch_compile_backend": null,
30
+ "torch_compile_mode": null,
31
+ "use_liger_kernel": false,
32
+ "liger_kernel_config": null,
33
+ "use_cache": false,
34
+ "neftune_noise_alpha": null,
35
+ "torch_empty_cache_steps": null,
36
+ "auto_find_batch_size": false,
37
+ "logging_strategy": "steps",
38
+ "logging_steps": 5,
39
+ "logging_first_step": true,
40
+ "log_on_each_node": true,
41
+ "logging_nan_inf_filter": true,
42
+ "include_num_input_tokens_seen": false,
43
+ "log_level": "passive",
44
+ "log_level_replica": "warning",
45
+ "disable_tqdm": null,
46
+ "report_to": [
47
+ "tensorboard"
48
+ ],
49
+ "run_name": "/root/code/na/Idea-filtering-training/output/Qwen3-1.7B/v0-20260304-031440",
50
+ "project": "huggingface",
51
+ "trackio_space_id": "trackio",
52
+ "eval_strategy": "no",
53
+ "eval_steps": 500,
54
+ "eval_delay": 0,
55
+ "per_device_eval_batch_size": 1,
56
+ "prediction_loss_only": false,
57
+ "eval_on_start": false,
58
+ "eval_do_concat_batches": true,
59
+ "eval_use_gather_object": false,
60
+ "eval_accumulation_steps": null,
61
+ "include_for_metrics": [],
62
+ "batch_eval_metrics": false,
63
+ "save_only_model": false,
64
+ "save_strategy": "steps",
65
+ "save_steps": 500,
66
+ "save_on_each_node": false,
67
+ "save_total_limit": null,
68
+ "enable_jit_checkpoint": false,
69
+ "push_to_hub": false,
70
+ "hub_token": null,
71
+ "hub_private_repo": null,
72
+ "hub_model_id": null,
73
+ "hub_strategy": "every_save",
74
+ "hub_always_push": false,
75
+ "hub_revision": null,
76
+ "load_best_model_at_end": false,
77
+ "metric_for_best_model": "loss",
78
+ "greater_is_better": false,
79
+ "ignore_data_skip": false,
80
+ "restore_callback_states_from_checkpoint": false,
81
+ "full_determinism": false,
82
+ "seed": 42,
83
+ "data_seed": 42,
84
+ "use_cpu": false,
85
+ "accelerator_config": {
86
+ "dispatch_batches": false
87
+ },
88
+ "parallelism_config": null,
89
+ "dataloader_drop_last": false,
90
+ "dataloader_num_workers": null,
91
+ "dataloader_pin_memory": true,
92
+ "dataloader_persistent_workers": false,
93
+ "dataloader_prefetch_factor": null,
94
+ "remove_unused_columns": true,
95
+ "label_names": null,
96
+ "train_sampling_strategy": "random",
97
+ "length_column_name": "length",
98
+ "ddp_find_unused_parameters": null,
99
+ "ddp_bucket_cap_mb": null,
100
+ "ddp_broadcast_buffers": null,
101
+ "ddp_backend": null,
102
+ "ddp_timeout": 18000000,
103
+ "fsdp": [],
104
+ "fsdp_config": null,
105
+ "deepspeed": null,
106
+ "debug": null,
107
+ "skip_memory_metrics": true,
108
+ "do_train": false,
109
+ "do_eval": false,
110
+ "do_predict": false,
111
+ "resume_from_checkpoint": null,
112
+ "warmup_ratio": null,
113
+ "logging_dir": "/root/code/na/Idea-filtering-training/output/Qwen3-1.7B/v0-20260304-031440/runs",
114
+ "local_rank": -1,
115
+ "sortish_sampler": false,
116
+ "predict_with_generate": false,
117
+ "generation_max_length": null,
118
+ "generation_num_beams": null,
119
+ "generation_config": null,
120
+ "tuner_backend": "peft",
121
+ "vit_gradient_checkpointing": null,
122
+ "router_aux_loss_coef": 0.0,
123
+ "enable_dft_loss": false,
124
+ "enable_channel_loss": false,
125
+ "check_model": true,
126
+ "acc_strategy": "token",
127
+ "train_dataloader_shuffle": true,
128
+ "group_by_length": false,
129
+ "max_epochs": null,
130
+ "aligner_lr": null,
131
+ "vit_lr": null,
132
+ "use_logits_to_keep": null,
133
+ "ds3_gather_for_generation": true,
134
+ "resume_only_model": false,
135
+ "optimizer": null,
136
+ "loss_type": null,
137
+ "eval_metric": null,
138
+ "callbacks": [],
139
+ "early_stop_interval": null,
140
+ "eval_use_evalscope": false,
141
+ "eval_dataset": [],
142
+ "eval_dataset_args": null,
143
+ "eval_limit": null,
144
+ "eval_generation_config": null,
145
+ "extra_eval_args": null,
146
+ "tuner_type": "lora",
147
+ "use_galore": false,
148
+ "galore_target_modules": null,
149
+ "galore_rank": 128,
150
+ "galore_update_proj_gap": 50,
151
+ "galore_scale": 1.0,
152
+ "galore_proj_type": "std",
153
+ "galore_optim_per_parameter": false,
154
+ "galore_with_embedding": false,
155
+ "galore_quantization": false,
156
+ "galore_proj_quant": false,
157
+ "galore_proj_bits": 4,
158
+ "galore_proj_group_size": 256,
159
+ "galore_cos_threshold": 0.4,
160
+ "galore_gamma_proj": 2,
161
+ "galore_queue_size": 5,
162
+ "lisa_activated_layers": 0,
163
+ "lisa_step_interval": 20,
164
+ "use_flash_ckpt": false,
165
+ "use_ray": false,
166
+ "ray_exp_name": null,
167
+ "device_groups": null,
168
+ "model": "Qwen/Qwen3-1.7B",
169
+ "model_type": "qwen3",
170
+ "model_revision": null,
171
+ "task_type": "causal_lm",
172
+ "torch_dtype": "bfloat16",
173
+ "attn_impl": null,
174
+ "experts_impl": null,
175
+ "new_special_tokens": [],
176
+ "num_labels": null,
177
+ "problem_type": null,
178
+ "rope_scaling": null,
179
+ "device_map": null,
180
+ "max_memory": {},
181
+ "max_model_len": null,
182
+ "local_repo_path": null,
183
+ "init_strategy": null,
184
+ "template": "qwen3",
185
+ "system": null,
186
+ "max_length": 40960,
187
+ "truncation_strategy": "delete",
188
+ "max_pixels": null,
189
+ "agent_template": null,
190
+ "norm_bbox": null,
191
+ "use_chat_template": true,
192
+ "padding_side": "right",
193
+ "padding_free": false,
194
+ "loss_scale": "default",
195
+ "sequence_parallel_size": 1,
196
+ "template_backend": "swift",
197
+ "response_prefix": null,
198
+ "enable_thinking": null,
199
+ "add_non_thinking_prefix": true,
200
+ "dataset": [
201
+ "/root/code/na/Idea-filtering-training/data/thinking_dataset_merged.jsonl"
202
+ ],
203
+ "val_dataset": [],
204
+ "cached_dataset": [],
205
+ "cached_val_dataset": [],
206
+ "split_dataset_ratio": 0.0,
207
+ "dataset_num_proc": 1,
208
+ "load_from_cache_file": false,
209
+ "dataset_shuffle": true,
210
+ "val_dataset_shuffle": false,
211
+ "streaming": false,
212
+ "interleave_prob": null,
213
+ "stopping_strategy": "first_exhausted",
214
+ "shuffle_buffer_size": 1000,
215
+ "download_mode": "reuse_dataset_if_exists",
216
+ "columns": {},
217
+ "strict": false,
218
+ "model_name": null,
219
+ "model_author": null,
220
+ "custom_dataset_info": [],
221
+ "quant_method": null,
222
+ "quant_bits": null,
223
+ "hqq_axis": null,
224
+ "bnb_4bit_compute_dtype": "bfloat16",
225
+ "bnb_4bit_quant_type": "nf4",
226
+ "bnb_4bit_use_double_quant": true,
227
+ "bnb_4bit_quant_storage": null,
228
+ "max_new_tokens": 64,
229
+ "temperature": 0.0,
230
+ "top_k": null,
231
+ "top_p": null,
232
+ "repetition_penalty": null,
233
+ "num_beams": 1,
234
+ "stream": false,
235
+ "stop_words": [],
236
+ "logprobs": false,
237
+ "top_logprobs": null,
238
+ "structured_outputs_regex": null,
239
+ "train_type": "lora",
240
+ "adapters": [],
241
+ "external_plugins": [],
242
+ "custom_register_path": [],
243
+ "model_kwargs": {},
244
+ "load_args": false,
245
+ "load_data_args": false,
246
+ "packing": false,
247
+ "packing_length": null,
248
+ "packing_num_proc": 1,
249
+ "lazy_tokenize": false,
250
+ "use_hf": true,
251
+ "ignore_args_error": false,
252
+ "use_swift_lora": false,
253
+ "freeze_parameters": [],
254
+ "freeze_parameters_regex": null,
255
+ "freeze_parameters_ratio": 0.0,
256
+ "trainable_parameters": [],
257
+ "trainable_parameters_regex": null,
258
+ "freeze_llm": false,
259
+ "freeze_vit": true,
260
+ "freeze_aligner": true,
261
+ "target_modules": [
262
+ "q_proj",
263
+ "k_proj",
264
+ "v_proj",
265
+ "o_proj",
266
+ "gate_proj",
267
+ "up_proj",
268
+ "down_proj"
269
+ ],
270
+ "target_regex": null,
271
+ "target_parameters": null,
272
+ "modules_to_save": [],
273
+ "lora_rank": 32,
274
+ "lora_alpha": 64,
275
+ "lora_dropout": 0.05,
276
+ "lora_bias": "none",
277
+ "lora_dtype": null,
278
+ "lorap_lr_ratio": null,
279
+ "use_rslora": false,
280
+ "use_dora": false,
281
+ "lora_ga_batch_size": 2,
282
+ "lora_ga_iters": 2,
283
+ "lora_ga_max_length": 1024,
284
+ "lora_ga_direction": "ArB2r",
285
+ "lora_ga_scale": "stable",
286
+ "lora_ga_stable_gamma": 16,
287
+ "init_weights": true,
288
+ "fourier_n_frequency": 2000,
289
+ "fourier_scaling": 300.0,
290
+ "boft_block_size": 4,
291
+ "boft_block_num": 0,
292
+ "boft_n_butterfly_factor": 1,
293
+ "boft_dropout": 0.0,
294
+ "vera_rank": 256,
295
+ "vera_projection_prng_key": 0,
296
+ "vera_dropout": 0.0,
297
+ "vera_d_initial": 0.1,
298
+ "adapter_act": "gelu",
299
+ "adapter_length": 128,
300
+ "adalora_target_r": 8,
301
+ "adalora_init_r": 12,
302
+ "adalora_tinit": 0,
303
+ "adalora_tfinal": 0,
304
+ "adalora_deltaT": 1,
305
+ "adalora_beta1": 0.85,
306
+ "adalora_beta2": 0.85,
307
+ "adalora_orth_reg_weight": 0.5,
308
+ "llamapro_num_new_blocks": 4,
309
+ "llamapro_num_groups": null,
310
+ "reft_layer_key": null,
311
+ "reft_layers": null,
312
+ "reft_rank": 4,
313
+ "reft_intervention_type": "LoreftIntervention",
314
+ "reft_args": null,
315
+ "swanlab_token": null,
316
+ "swanlab_project": "ms-swift",
317
+ "swanlab_workspace": null,
318
+ "swanlab_exp_name": null,
319
+ "swanlab_notification_method": null,
320
+ "swanlab_webhook_url": null,
321
+ "swanlab_secret": null,
322
+ "swanlab_sender_email": null,
323
+ "swanlab_receiver_email": null,
324
+ "swanlab_smtp_server": null,
325
+ "swanlab_smtp_port": null,
326
+ "swanlab_email_language": "zh",
327
+ "swanlab_mode": "cloud",
328
+ "add_version": true,
329
+ "create_checkpoint_symlink": false,
330
+ "zero_hpz_partition_size": null,
331
+ "deepspeed_autotp_size": null,
332
+ "swift_version": "4.0.0",
333
+ "ckpt_dir": null,
334
+ "rank": -1,
335
+ "global_world_size": 1,
336
+ "local_world_size": 1,
337
+ "model_suffix": "Qwen3-1.7B",
338
+ "model_info": "ModelInfo(model_type='qwen3', model_dir='/root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e', torch_dtype=torch.bfloat16, max_model_len=40960, quant_method=None, quant_bits=None, rope_scaling={'rope_theta': 1000000, 'rope_type': 'default'}, is_moe_model=False, is_multimodal=False, config=None, task_type='causal_lm', num_labels=None)",
339
+ "model_meta": "ModelMeta(model_type='qwen3', model_groups=[ModelGroup(models=[Model(ms_model_id='Qwen/Qwen3-0.6B-Base', hf_model_id='Qwen/Qwen3-0.6B-Base', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-1.7B-Base', hf_model_id='Qwen/Qwen3-1.7B-Base', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-4B-Base', hf_model_id='Qwen/Qwen3-4B-Base', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-8B-Base', hf_model_id='Qwen/Qwen3-8B-Base', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-14B-Base', hf_model_id='Qwen/Qwen3-14B-Base', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-0.6B', hf_model_id='Qwen/Qwen3-0.6B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-1.7B', hf_model_id='Qwen/Qwen3-1.7B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-4B', hf_model_id='Qwen/Qwen3-4B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-8B', hf_model_id='Qwen/Qwen3-8B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-14B', hf_model_id='Qwen/Qwen3-14B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-32B', hf_model_id='Qwen/Qwen3-32B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-0.6B-FP8', hf_model_id='Qwen/Qwen3-0.6B-FP8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-1.7B-FP8', hf_model_id='Qwen/Qwen3-1.7B-FP8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-4B-FP8', hf_model_id='Qwen/Qwen3-4B-FP8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-8B-FP8', hf_model_id='Qwen/Qwen3-8B-FP8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-14B-FP8', hf_model_id='Qwen/Qwen3-14B-FP8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-32B-FP8', hf_model_id='Qwen/Qwen3-32B-FP8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-4B-AWQ', hf_model_id='Qwen/Qwen3-4B-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-8B-AWQ', hf_model_id='Qwen/Qwen3-8B-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-14B-AWQ', hf_model_id='Qwen/Qwen3-14B-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-32B-AWQ', hf_model_id='Qwen/Qwen3-32B-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='swift/Qwen3-32B-AWQ', hf_model_id=None, model_path=None, ms_revision=None, hf_revision=None)], template='qwen3', ignore_patterns=None, requires=None, tags=[]), ModelGroup(models=[Model(ms_model_id='deepseek-ai/DeepSeek-R1-0528-Qwen3-8B', hf_model_id='deepseek-ai/DeepSeek-R1-0528-Qwen3-8B', model_path=None, ms_revision=None, hf_revision=None)], template='deepseek_r1', ignore_patterns=None, requires=None, tags=[]), ModelGroup(models=[Model(ms_model_id='Qwen/Qwen3Guard-Gen-0.6B', hf_model_id='Qwen/Qwen3Guard-Gen-0.6B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3Guard-Gen-4B', hf_model_id='Qwen/Qwen3Guard-Gen-4B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3Guard-Gen-8B', hf_model_id='Qwen/Qwen3Guard-Gen-8B', model_path=None, ms_revision=None, hf_revision=None)], template='qwen3_guard', ignore_patterns=None, requires=None, tags=[]), ModelGroup(models=[Model(ms_model_id='Qwen/Qwen3-4B-Thinking-2507', hf_model_id='Qwen/Qwen3-4B-Thinking-2507', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-4B-Thinking-2507-FP8', hf_model_id='Qwen/Qwen3-4B-Thinking-2507-FP8', model_path=None, ms_revision=None, hf_revision=None)], template='qwen3_thinking', ignore_patterns=None, requires=None, tags=[]), ModelGroup(models=[Model(ms_model_id='Qwen/Qwen3-4B-Instruct-2507', hf_model_id='Qwen/Qwen3-4B-Instruct-2507', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen3-4B-Instruct-2507-FP8', hf_model_id='Qwen/Qwen3-4B-Instruct-2507-FP8', model_path=None, ms_revision=None, hf_revision=None)], template='qwen3_nothinking', ignore_patterns=None, requires=None, tags=[])], loader=<class 'swift.model.register.ModelLoader'>, template='qwen3', model_arch=ModelKeys(arch_name='llama', embedding='model.embed_tokens', module_list='model.layers', lm_head='lm_head', q_proj='model.layers.{}.self_attn.q_proj', k_proj='model.layers.{}.self_attn.k_proj', v_proj='model.layers.{}.self_attn.v_proj', o_proj='model.layers.{}.self_attn.o_proj', attention='model.layers.{}.self_attn', mlp='model.layers.{}.mlp', down_proj='model.layers.{}.mlp.down_proj', qkv_proj=None, qk_proj=None, qa_proj=None, qb_proj=None, kv_proj=None, kva_proj=None, kvb_proj=None), architectures=['Qwen3ForCausalLM'], additional_saved_files=[], torch_dtype=None, is_multimodal=False, is_reward=False, task_type=None, ignore_patterns=None, requires=['transformers>=4.51'], tags=[])",
340
+ "model_dir": "/root/.cache/huggingface/hub/models--Qwen--Qwen3-1.7B/snapshots/70d244cc86ccca08cf5af4e1e306ecf908b1ad5e",
341
+ "template_meta": "Qwen3MixedTemplateMeta(template_type='qwen3', prefix=[], prompt=['<|im_start|>user\\n{{QUERY}}<|im_end|>\\n<|im_start|>assistant\\n'], chat_sep=['<|im_end|>\\n'], suffix=['<|im_end|>\\n'], template_cls=<class 'swift.template.base.Template'>, system_prefix=['<|im_start|>system\\n{{SYSTEM}}<|im_end|>\\n'], default_system=None, auto_add_bos=False, stop_words=['<|endoftext|>'], agent_template='hermes', is_thinking=True, thinking_prefix='', non_thinking_prefix='<think>\\n\\n</think>\\n\\n', history_thinking_prefix='')",
342
+ "_val_dataset_exists": false,
343
+ "hub": "<class 'swift.hub.hub.HFHub'>",
344
+ "evaluation_strategy": "steps",
345
+ "training_args": "Seq2SeqTrainingArguments(output_dir='/root/code/na/Idea-filtering-training/output/Qwen3-1.7B/v0-20260304-031440', per_device_train_batch_size=1, num_train_epochs=3.0, max_steps=-1, learning_rate=0.0001, lr_scheduler_type=<SchedulerType.COSINE: 'cosine'>, lr_scheduler_kwargs=None, warmup_steps=0, optim=<OptimizerNames.ADAMW_TORCH_FUSED: 'adamw_torch_fused'>, optim_args=None, weight_decay=0.1, adam_beta1=0.9, adam_beta2=0.95, adam_epsilon=1e-08, optim_target_modules=None, gradient_accumulation_steps=16, average_tokens_across_devices=None, max_grad_norm=1.0, label_smoothing_factor=0.0, bf16=True, fp16=False, bf16_full_eval=False, fp16_full_eval=False, tf32=None, gradient_checkpointing=True, gradient_checkpointing_kwargs=None, torch_compile=False, torch_compile_backend=None, torch_compile_mode=None, use_liger_kernel=False, liger_kernel_config=None, use_cache=False, neftune_noise_alpha=None, torch_empty_cache_steps=None, auto_find_batch_size=False, logging_strategy=<IntervalStrategy.STEPS: 'steps'>, logging_steps=5, logging_first_step=True, log_on_each_node=True, logging_nan_inf_filter=True, include_num_input_tokens_seen=None, log_level='passive', log_level_replica='warning', disable_tqdm=False, report_to=['tensorboard'], run_name='/root/code/na/Idea-filtering-training/output/Qwen3-1.7B/v0-20260304-031440', project='huggingface', trackio_space_id='trackio', eval_strategy=<IntervalStrategy.NO: 'no'>, eval_steps=500, eval_delay=0, per_device_eval_batch_size=1, prediction_loss_only=False, eval_on_start=False, eval_do_concat_batches=True, eval_use_gather_object=False, eval_accumulation_steps=None, include_for_metrics=[], batch_eval_metrics=False, save_only_model=False, save_strategy=<SaveStrategy.STEPS: 'steps'>, save_steps=500, save_on_each_node=False, save_total_limit=None, enable_jit_checkpoint=False, push_to_hub=False, hub_token=None, hub_private_repo=None, hub_model_id=None, hub_strategy=<HubStrategy.EVERY_SAVE: 'every_save'>, hub_always_push=False, hub_revision=None, load_best_model_at_end=False, metric_for_best_model='loss', greater_is_better=False, ignore_data_skip=False, restore_callback_states_from_checkpoint=False, full_determinism=False, seed=42, data_seed=42, use_cpu=False, accelerator_config=AcceleratorConfig(split_batches=False, dispatch_batches=False, even_batches=True, use_seedable_sampler=True, non_blocking=False, gradient_accumulation_kwargs=None, use_configured_state=False), parallelism_config=None, dataloader_drop_last=False, dataloader_num_workers=1, dataloader_pin_memory=True, dataloader_persistent_workers=False, dataloader_prefetch_factor=2, remove_unused_columns=False, label_names=None, train_sampling_strategy='random', length_column_name='length', ddp_find_unused_parameters=None, ddp_bucket_cap_mb=None, ddp_broadcast_buffers=None, ddp_backend=None, ddp_timeout=18000000, fsdp=[], fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, deepspeed=None, debug=[], skip_memory_metrics=True, do_train=False, do_eval=False, do_predict=False, resume_from_checkpoint=None, warmup_ratio=None, logging_dir='/root/code/na/Idea-filtering-training/output/Qwen3-1.7B/v0-20260304-031440/runs', local_rank=-1, sortish_sampler=False, predict_with_generate=False, generation_max_length=None, generation_num_beams=None, generation_config=None, tuner_backend='peft', vit_gradient_checkpointing=True, router_aux_loss_coef=0.0, enable_dft_loss=False, enable_channel_loss=False, check_model=True, acc_strategy='token', train_dataloader_shuffle=True, group_by_length=False, max_epochs=None, aligner_lr=None, vit_lr=None, use_logits_to_keep=None, ds3_gather_for_generation=True, resume_only_model=False, optimizer=None, loss_type=None, eval_metric=None, callbacks=[], early_stop_interval=None, eval_use_evalscope=False, eval_dataset=[], eval_dataset_args=None, eval_limit=None, eval_generation_config=None, extra_eval_args=None, tuner_type='lora', use_galore=False, galore_target_modules=None, galore_rank=128, galore_update_proj_gap=50, galore_scale=1.0, galore_proj_type='std', galore_optim_per_parameter=False, galore_with_embedding=False, galore_quantization=False, galore_proj_quant=False, galore_proj_bits=4, galore_proj_group_size=256, galore_cos_threshold=0.4, galore_gamma_proj=2, galore_queue_size=5, lisa_activated_layers=0, lisa_step_interval=20, use_flash_ckpt=False)"
346
+ }
output/Qwen3-1.7B/v0-20260304-031440/images/train_epoch.png ADDED
output/Qwen3-1.7B/v0-20260304-031440/images/train_grad_norm.png ADDED
output/Qwen3-1.7B/v0-20260304-031440/images/train_learning_rate.png ADDED
output/Qwen3-1.7B/v0-20260304-031440/images/train_loss.png ADDED
output/Qwen3-1.7B/v0-20260304-031440/images/train_token_acc.png ADDED
output/Qwen3-1.7B/v0-20260304-031440/logging.jsonl ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"loss": 2.80828404, "grad_norm": 0.70461369, "learning_rate": 0.0001, "token_acc": 0.54860752, "epoch": 0.00039497, "global_step/max_steps": "1/7596", "elapsed_time": "22s", "remaining_time": "1d 22h 44m 52s", "memory(GiB)": 9.3, "train_speed(s/it)": 22.158197}
2
+ {"loss": 2.10956383, "grad_norm": 0.06330702, "learning_rate": 0.0001, "token_acc": 0.54739539, "epoch": 0.00197487, "global_step/max_steps": "5/7596", "elapsed_time": "1m 45s", "remaining_time": "1d 20h 26m 11s", "memory(GiB)": 11.09, "train_speed(s/it)": 21.073719}
3
+ {"loss": 1.88895645, "grad_norm": 0.06611262, "learning_rate": 0.0001, "token_acc": 0.54985593, "epoch": 0.00394974, "global_step/max_steps": "10/7596", "elapsed_time": "3m 30s", "remaining_time": "1d 20h 20m 37s", "memory(GiB)": 11.09, "train_speed(s/it)": 21.043592}
4
+ {"loss": 1.81560631, "grad_norm": 0.03588019, "learning_rate": 0.0001, "token_acc": 0.55962598, "epoch": 0.00592461, "global_step/max_steps": "15/7596", "elapsed_time": "5m 17s", "remaining_time": "1d 20h 30m 47s", "memory(GiB)": 11.09, "train_speed(s/it)": 21.137915}
5
+ {"loss": 1.68152924, "grad_norm": 0.03120046, "learning_rate": 0.0001, "token_acc": 0.58999685, "epoch": 0.00789948, "global_step/max_steps": "20/7596", "elapsed_time": "7m 5s", "remaining_time": "1d 20h 43m 44s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.254456}
6
+ {"loss": 1.73896103, "grad_norm": 0.02689376, "learning_rate": 0.0001, "token_acc": 0.58061652, "epoch": 0.00987435, "global_step/max_steps": "25/7596", "elapsed_time": "8m 48s", "remaining_time": "1d 20h 26m 37s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.132875}
7
+ {"loss": 1.63750725, "grad_norm": 0.02659717, "learning_rate": 0.0001, "token_acc": 0.59855935, "epoch": 0.01184922, "global_step/max_steps": "30/7596", "elapsed_time": "10m 35s", "remaining_time": "1d 20h 27m 52s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.156749}
8
+ {"loss": 1.62336521, "grad_norm": 0.02456974, "learning_rate": 9.999e-05, "token_acc": 0.59320185, "epoch": 0.01382409, "global_step/max_steps": "35/7596", "elapsed_time": "12m 21s", "remaining_time": "1d 20h 27m 45s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.169781}
9
+ {"loss": 1.56308775, "grad_norm": 0.02603255, "learning_rate": 9.999e-05, "token_acc": 0.60313748, "epoch": 0.01579896, "global_step/max_steps": "40/7596", "elapsed_time": "14m 8s", "remaining_time": "1d 20h 29m 47s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.200038}
10
+ {"loss": 1.60969524, "grad_norm": 0.02644657, "learning_rate": 9.999e-05, "token_acc": 0.59429054, "epoch": 0.01777383, "global_step/max_steps": "45/7596", "elapsed_time": "15m 50s", "remaining_time": "1d 20h 16m 44s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.110333}
11
+ {"loss": 1.54298725, "grad_norm": 0.02468126, "learning_rate": 9.999e-05, "token_acc": 0.61193504, "epoch": 0.0197487, "global_step/max_steps": "50/7596", "elapsed_time": "17m 37s", "remaining_time": "1d 20h 18m 46s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.140509}
12
+ {"loss": 1.50155048, "grad_norm": 0.02367931, "learning_rate": 9.999e-05, "token_acc": 0.61832011, "epoch": 0.02172357, "global_step/max_steps": "55/7596", "elapsed_time": "19m 22s", "remaining_time": "1d 20h 15m 47s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.130747}
13
+ {"loss": 1.52671709, "grad_norm": 0.02718145, "learning_rate": 9.998e-05, "token_acc": 0.60960499, "epoch": 0.02369844, "global_step/max_steps": "60/7596", "elapsed_time": "21m 8s", "remaining_time": "1d 20h 13m 59s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.130449}
14
+ {"loss": 1.54125071, "grad_norm": 0.02729321, "learning_rate": 9.998e-05, "token_acc": 0.61034281, "epoch": 0.02567331, "global_step/max_steps": "65/7596", "elapsed_time": "22m 51s", "remaining_time": "1d 20h 7m 16s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.090954}
15
+ {"loss": 1.500739, "grad_norm": 0.02714855, "learning_rate": 9.998e-05, "token_acc": 0.61303845, "epoch": 0.02764818, "global_step/max_steps": "70/7596", "elapsed_time": "24m 34s", "remaining_time": "1d 20h 2m 6s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.063844}
16
+ {"loss": 1.51486931, "grad_norm": 0.02550467, "learning_rate": 9.998e-05, "token_acc": 0.61092864, "epoch": 0.02962305, "global_step/max_steps": "75/7596", "elapsed_time": "26m 19s", "remaining_time": "1d 19h 59m 50s", "memory(GiB)": 11.15, "train_speed(s/it)": 21.059681}
17
+ {"train_dataset": "1665.517589±354.737407, min=757.000000, max=38083.000000, size=40509", "model_parameter_info": "PeftModelForCausalLM: 2066.6051M Params (34.8652M Trainable [1.6871%]), 0.0001M Buffers.", "last_model_checkpoint": null, "best_model_checkpoint": null, "best_metric": null, "global_step": 78, "log_history": [{"loss": 2.808284044265747, "grad_norm": 0.7046136856079102, "learning_rate": 9.999999572368355e-05, "token_acc": 0.5486075163060358, "epoch": 0.0003949739564047496, "step": 1}, {"loss": 2.1095638275146484, "grad_norm": 0.063307024538517, "learning_rate": 9.999989309212533e-05, "token_acc": 0.547395388556789, "epoch": 0.0019748697820237476, "step": 5}, {"loss": 1.8889564514160155, "grad_norm": 0.06611262261867523, "learning_rate": 9.999957236895852e-05, "token_acc": 0.5498559329433336, "epoch": 0.003949739564047495, "step": 10}, {"loss": 1.8156063079833984, "grad_norm": 0.0358801893889904, "learning_rate": 9.999903783187103e-05, "token_acc": 0.5596259794991111, "epoch": 0.005924609346071244, "step": 15}, {"loss": 1.6815292358398437, "grad_norm": 0.031200457364320755, "learning_rate": 9.999828948314876e-05, "token_acc": 0.5899968477461385, "epoch": 0.00789947912809499, "step": 20}, {"loss": 1.7389610290527344, "grad_norm": 0.02689375728368759, "learning_rate": 9.999732732599186e-05, "token_acc": 0.5806165194426737, "epoch": 0.009874348910118739, "step": 25}, {"loss": 1.6375072479248047, "grad_norm": 0.026597166433930397, "learning_rate": 9.999615136451483e-05, "token_acc": 0.5985593512767425, "epoch": 0.011849218692142487, "step": 30}, {"loss": 1.6233652114868165, "grad_norm": 0.024569744244217873, "learning_rate": 9.999476160374644e-05, "token_acc": 0.5932018540398073, "epoch": 0.013824088474166234, "step": 35}, {"loss": 1.5630877494812012, "grad_norm": 0.026032552123069763, "learning_rate": 9.999315804962974e-05, "token_acc": 0.603137482149879, "epoch": 0.01579895825618998, "step": 40}, {"loss": 1.6096952438354493, "grad_norm": 0.026446569710969925, "learning_rate": 9.999134070902207e-05, "token_acc": 0.5942905357896218, "epoch": 0.01777382803821373, "step": 45}, {"loss": 1.5429872512817382, "grad_norm": 0.02468126453459263, "learning_rate": 9.99893095896949e-05, "token_acc": 0.6119350428766726, "epoch": 0.019748697820237478, "step": 50}, {"loss": 1.5015504837036133, "grad_norm": 0.023679306730628014, "learning_rate": 9.998706470033396e-05, "token_acc": 0.6183201072739662, "epoch": 0.021723567602261226, "step": 55}, {"loss": 1.5267170906066894, "grad_norm": 0.02718144841492176, "learning_rate": 9.998460605053911e-05, "token_acc": 0.6096049924254415, "epoch": 0.023698437384284975, "step": 60}, {"loss": 1.5412507057189941, "grad_norm": 0.027293212711811066, "learning_rate": 9.99819336508243e-05, "token_acc": 0.6103428082346052, "epoch": 0.025673307166308723, "step": 65}, {"loss": 1.5007390022277831, "grad_norm": 0.027148548513650894, "learning_rate": 9.997904751261756e-05, "token_acc": 0.6130384536946818, "epoch": 0.027648176948332468, "step": 70}, {"loss": 1.5148693084716798, "grad_norm": 0.025504665449261665, "learning_rate": 9.997594764826092e-05, "token_acc": 0.6109286421308378, "epoch": 0.029623046730356217, "step": 75}], "memory": 11.15234375}
output/Qwen3-1.7B/v0-20260304-031440/runs/Mar04_03-19-12_ubuntu/events.out.tfevents.1772594352.ubuntu.2490446.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04588eed7d3811342a53b44cd39e2a3c6c1789117591fe16de77bed141c4d4b0
3
+ size 10843
output/proposals_RM_gpt_oss_120b_ensemble/polaris/polaris-adme-fang-hclint-1_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_120b_ensemble/polaris/polaris-adme-fang-hppb-1_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_120b_ensemble/polaris/polaris-adme-fang-solu-1_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_120b_ensemble/polaris/tdcommons-bbb-martins_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_120b_ensemble/polaris/tdcommons-cyp2d6-substrate-carbonmangels_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_120b_ensemble/polaris/tdcommons-herg_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_120b_ensemble/polaris/tdcommons-lipophilicity-astrazeneca_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_20b_ensemble/polaris/polaris-adme-fang-hclint-1_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_20b_ensemble/polaris/polaris-adme-fang-hppb-1_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_20b_ensemble/polaris/polaris-adme-fang-solu-1_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_20b_ensemble/polaris/tdcommons-bbb-martins_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_20b_ensemble/polaris/tdcommons-cyp2d6-substrate-carbonmangels_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_20b_ensemble/polaris/tdcommons-herg_scored_16.json ADDED
The diff for this file is too large to render. See raw diff
 
output/proposals_RM_gpt_oss_20b_ensemble/polaris/tdcommons-lipophilicity-astrazeneca_scored_16.json ADDED
The diff for this file is too large to render. See raw diff