File size: 1,944 Bytes
c94c8c9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
# Experiment general info
name: "OV_SQA3D"
rng_seed: 42
num_gpu: 1
mode: "train"
note: ""
# Choose keywords to feature your saving directory
naming_keywords: ["dataloader.batchsize", "task", "note", "time"]
base_dir: "results"
exp_dir: ""
save_frequency: 100

resume: False

debug:
  flag: False
  hard_debug: False
  debug_size: 20

logger:
  name: "wandb"
  entity: "yem"

# dataset details
data:
  train: ['ScanNetSQA3D']
  val: ['ScanNetSQA3D']
  test: ['ScanNetSQA3D']
  ScanNetSQA3D:
    train:
      use_unanswer: True
    val:
      use_unanswer: True
    test:
      use_unanswer: True
  scan_family_base: "../PointMapVerse/existing_datasets/ScanNet"

# task details: 'pretrain', 'scanrefer', 'referit3d', 'scanqa', 'default'
task: 'SQA3D'
data_wrapper:
  train: 'ScanFamilyDatasetWrapperQA'
  val: 'ScanFamilyDatasetWrapperQA'
  test: 'ScanFamilyDatasetWrapperQA'

# Training details
trainer: "DefaultTrainer"
ckpt_path: ""
pretrain_ckpt_path: "/home/m50048399/transfered/ye_project/Project2/results/sceneverse_scannet_exp1_b64_Pretrain_all_scannet_training_run1/2026-01-19-23:46:36.901933/ckpt/ckpt_20.pth"

# dataloader details
dataloader:
  batchsize: 128
  num_workers: 2
  balance_dataset: False
  filter_empty_annotations: False

solver:
  gradient_accumulation_steps: 1
  epochs_per_save: 100
  epochs_per_eval: 1
  lr: 1e-4
  grad_norm: 5.0
  epochs: 100
  optim:
    name: "AdamW"
    args:
      betas: [0.9, 0.98]
  sched:
    name: "warmup_cosine"
    args:
      warmup_steps: 5000

eval:
  name: "SQA3DEval"
  save: False

# Model details
model:
  name: OpenVocab
  vision:
    name: 'fg-clip-base'
    lr: 1e-4
  heads:
    head_list: ["qa_head"]
    qa_head:
      name: "QAHeadV1"
      args:
        hidden_size: 768
        mlp_size: 256
        glimpse: 1
        flat_out_size: 512
        num_answers: 706
  loss_type: "ListLoss"
  loss_list: [
      "answer_loss"
  ]
  vis_loss_list: [
      "answer_loss"
  ]