checkpoint step 400 epoch 2
Browse files
0402_1110_checkpoint_step_400_epoch_2/config.json
ADDED
|
@@ -0,0 +1,101 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 2,
|
| 3 |
+
"global_step": 400,
|
| 4 |
+
"loss": 0.7921013832092285,
|
| 5 |
+
"training_config": {
|
| 6 |
+
"llm_backbone": "dasheng",
|
| 7 |
+
"qwen3_name": "Qwen/Qwen3-Embedding-0.6B",
|
| 8 |
+
"dasheng_name": "mispeech/midashenglm-7b-0804-fp32",
|
| 9 |
+
"dasheng_path": "/workspace/cache/huggingface/dasheng_lm",
|
| 10 |
+
"trainable_modules": [
|
| 11 |
+
"backbone",
|
| 12 |
+
"dasheng",
|
| 13 |
+
"dasheng_down",
|
| 14 |
+
"dasheng_proj",
|
| 15 |
+
"siglip_head"
|
| 16 |
+
],
|
| 17 |
+
"use_lora": true,
|
| 18 |
+
"lora_r": 16,
|
| 19 |
+
"lora_alpha": 32,
|
| 20 |
+
"lora_dropout": 0.05,
|
| 21 |
+
"lora_target_modules": [
|
| 22 |
+
"q_proj",
|
| 23 |
+
"k_proj",
|
| 24 |
+
"v_proj"
|
| 25 |
+
],
|
| 26 |
+
"use_dasheng_lora": false,
|
| 27 |
+
"dasheng_lora_r": 8,
|
| 28 |
+
"dasheng_lora_alpha": 16,
|
| 29 |
+
"dasheng_lora_dropout": 0.1,
|
| 30 |
+
"dasheng_lora_target_modules": null,
|
| 31 |
+
"train_layer_ratio": 1.0,
|
| 32 |
+
"train_layer_strategy": "last_n",
|
| 33 |
+
"output_dim": null,
|
| 34 |
+
"output_identity": false,
|
| 35 |
+
"use_logit_scale": true,
|
| 36 |
+
"loss_type": "infonce",
|
| 37 |
+
"use_checkpointing": true,
|
| 38 |
+
"checkpoint_reentrant": false,
|
| 39 |
+
"gather_negatives": true,
|
| 40 |
+
"use_loss_mask": true,
|
| 41 |
+
"duplicate_doc_threshold": 0.999,
|
| 42 |
+
"duplicate_query_threshold": 0.999,
|
| 43 |
+
"hard_negative_margin": 0.1,
|
| 44 |
+
"add_speaker_mask": false,
|
| 45 |
+
"s3_base_path": "https://d2j287p0ytux1o.cloudfront.net",
|
| 46 |
+
"dataset_config": "/workspace/SpeechRAG_exp/dataset_configs/setting_countingqa_embedding.json",
|
| 47 |
+
"aws_profile": "test_user",
|
| 48 |
+
"cache_dir": "/workspace/cache/huggingface",
|
| 49 |
+
"enable_audio_cache": false,
|
| 50 |
+
"audio_cache_dir": null,
|
| 51 |
+
"target_sr": 16000,
|
| 52 |
+
"mono": true,
|
| 53 |
+
"max_query_audio_length": 45.0,
|
| 54 |
+
"max_doc_audio_length": 45.0,
|
| 55 |
+
"max_query_text_length": 1000,
|
| 56 |
+
"max_doc_text_length": 1000,
|
| 57 |
+
"eval_max_query_audio_length": null,
|
| 58 |
+
"eval_max_doc_audio_length": null,
|
| 59 |
+
"eval_max_query_text_length": null,
|
| 60 |
+
"eval_max_doc_text_length": null,
|
| 61 |
+
"batch_size": 64,
|
| 62 |
+
"num_epochs": 10,
|
| 63 |
+
"learning_rate": 0.0001,
|
| 64 |
+
"gradient_accumulation_steps": 2,
|
| 65 |
+
"use_grad_cache": true,
|
| 66 |
+
"gc_query_chunk_size": 8,
|
| 67 |
+
"gc_doc_chunk_size": 8,
|
| 68 |
+
"gc_no_sync_except_last": true,
|
| 69 |
+
"ddp_find_unused_parameters": false,
|
| 70 |
+
"weight_decay": 0.001,
|
| 71 |
+
"optimizer_bits": "default",
|
| 72 |
+
"num_workers": 16,
|
| 73 |
+
"train_batch_task_mode": "mixed_task",
|
| 74 |
+
"task_batch_ratio": "{\"semantic\": 0.5, \"cross\": 0.5}",
|
| 75 |
+
"save_dir": "checkpoints",
|
| 76 |
+
"save_steps": 100,
|
| 77 |
+
"keep_checkpoints": 1,
|
| 78 |
+
"upload_steps": 100,
|
| 79 |
+
"upload_repo_id": "jdosjcd/embedding_checkpoint",
|
| 80 |
+
"log_dir": "logs",
|
| 81 |
+
"log_steps": 5,
|
| 82 |
+
"mixed_precision": "bf16",
|
| 83 |
+
"use_deepspeed": false,
|
| 84 |
+
"deepspeed_config": null,
|
| 85 |
+
"use_fsdp": false,
|
| 86 |
+
"fsdp_config": null,
|
| 87 |
+
"use_ema": false,
|
| 88 |
+
"ema_decay": 0.9999,
|
| 89 |
+
"ema_update_after": 0,
|
| 90 |
+
"ema_update_every": 1,
|
| 91 |
+
"scheduler_type": "warmup_cosine_decay",
|
| 92 |
+
"warmup_steps": 500,
|
| 93 |
+
"warmup_ratio": 0.1,
|
| 94 |
+
"min_lr": 1e-05,
|
| 95 |
+
"eval_steps": 100,
|
| 96 |
+
"eval_batch_size": 32,
|
| 97 |
+
"resume_from": null,
|
| 98 |
+
"test_mode": false,
|
| 99 |
+
"mock_dataset_size": 100
|
| 100 |
+
}
|
| 101 |
+
}
|
0402_1110_checkpoint_step_400_epoch_2/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:24bc7ddc8525299a99ae89bf07d0927c34d7f30e08ac118479f2eaa569f70bd4
|
| 3 |
+
size 31026342367
|
0402_1110_checkpoint_step_400_epoch_2/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:17b5c0837294a6b6a8300b97f1f35d378806e626c26814625c07eb968b5c5142
|
| 3 |
+
size 5487302675
|
0402_1110_checkpoint_step_400_epoch_2/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d1f3bf0f2f1b687feac25ff8333d29d687617a0c27d714d34c98017b69075dda
|
| 3 |
+
size 1401
|