anicolson commited on
Commit
0ce506a
·
verified ·
1 Parent(s): 672b974

Upload CXRMate2ForConditionalGeneration

Browse files
Files changed (2) hide show
  1. config.json +115 -0
  2. generation_config.json +1 -2
config.json ADDED
@@ -0,0 +1,115 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "CXRMate2ForConditionalGeneration"
4
+ ],
5
+ "auto_map": {
6
+ "AutoConfig": "configuration_cxrmate2.CXRMate2Config",
7
+ "AutoModelForCausalLM": "modelling_cxrmate2.CXRMate2ForConditionalGeneration"
8
+ },
9
+ "bos_token_id": 128000,
10
+ "dtype": "float32",
11
+ "findings_token_type_id": 128029,
12
+ "generate_both_sections_token_id": null,
13
+ "generate_findings_token_id": 128003,
14
+ "generate_impression_token_id": 128005,
15
+ "image_token_index": 128002,
16
+ "impression_token_type_id": 128030,
17
+ "missing_time_delta_token_id": 128012,
18
+ "model_type": "cxrmate-2",
19
+ "multimodal_projector_bias": false,
20
+ "num_q_adapter_layers": 2,
21
+ "num_q_adapter_positions": 1497,
22
+ "num_q_adapter_queries": 128,
23
+ "num_token_types": null,
24
+ "permute_encoder_last_hidden_state": true,
25
+ "projector_hidden_act": null,
26
+ "sep_token_id": 128072,
27
+ "text_config": {
28
+ "_name_or_path": "meta-llama/Llama-3.2-3B",
29
+ "architectures": [
30
+ "LlamaForCausalLM"
31
+ ],
32
+ "attention_bias": false,
33
+ "attention_dropout": 0.0,
34
+ "bos_token_id": 128000,
35
+ "dtype": "float32",
36
+ "eos_token_id": 128001,
37
+ "head_dim": 128,
38
+ "hidden_act": "silu",
39
+ "hidden_size": 3072,
40
+ "initializer_range": 0.02,
41
+ "intermediate_size": 8192,
42
+ "max_position_embeddings": 131072,
43
+ "mlp_bias": false,
44
+ "model_type": "llama",
45
+ "num_attention_heads": 24,
46
+ "num_hidden_layers": 28,
47
+ "num_key_value_heads": 8,
48
+ "pretraining_tp": 1,
49
+ "rms_norm_eps": 1e-05,
50
+ "rope_scaling": {
51
+ "factor": 32.0,
52
+ "high_freq_factor": 4.0,
53
+ "low_freq_factor": 1.0,
54
+ "original_max_position_embeddings": 8192,
55
+ "rope_type": "llama3"
56
+ },
57
+ "rope_theta": 500000.0,
58
+ "tie_word_embeddings": true,
59
+ "use_cache": true,
60
+ "vocab_size": 128256
61
+ },
62
+ "time_delta_encoder_intermediate_size": 2048,
63
+ "time_delta_monotonic_inversion": true,
64
+ "transformers_version": "4.57.1",
65
+ "vision_config": {
66
+ "_name_or_path": "microsoft/rad-dino-maira-2",
67
+ "apply_layernorm": true,
68
+ "architectures": [
69
+ "Dinov2Model"
70
+ ],
71
+ "attention_probs_dropout_prob": 0.0,
72
+ "drop_path_rate": 0.0,
73
+ "dtype": "float32",
74
+ "hidden_act": "gelu",
75
+ "hidden_dropout_prob": 0.0,
76
+ "hidden_size": 768,
77
+ "image_size": 518,
78
+ "initializer_range": 0.02,
79
+ "layer_norm_eps": 1e-06,
80
+ "layerscale_value": 1.0,
81
+ "mlp_ratio": 4,
82
+ "model_type": "dinov2",
83
+ "num_attention_heads": 12,
84
+ "num_channels": 3,
85
+ "num_hidden_layers": 12,
86
+ "out_features": [
87
+ "stage12"
88
+ ],
89
+ "out_indices": [
90
+ 12
91
+ ],
92
+ "patch_size": 14,
93
+ "qkv_bias": true,
94
+ "reshape_hidden_states": true,
95
+ "stage_names": [
96
+ "stem",
97
+ "stage1",
98
+ "stage2",
99
+ "stage3",
100
+ "stage4",
101
+ "stage5",
102
+ "stage6",
103
+ "stage7",
104
+ "stage8",
105
+ "stage9",
106
+ "stage10",
107
+ "stage11",
108
+ "stage12"
109
+ ],
110
+ "use_mask_token": true,
111
+ "use_swiglu_ffn": false
112
+ },
113
+ "vision_feature_layer": -1,
114
+ "vision_feature_select_strategy": "full"
115
+ }
generation_config.json CHANGED
@@ -1,7 +1,6 @@
1
  {
 
2
  "bos_token_id": 128000,
3
  "eos_token_id": 128001,
4
- "max_new_tokens": 320,
5
- "pad_token_id": 128073,
6
  "transformers_version": "4.57.1"
7
  }
 
1
  {
2
+ "_from_model_config": true,
3
  "bos_token_id": 128000,
4
  "eos_token_id": 128001,
 
 
5
  "transformers_version": "4.57.1"
6
  }