Update llava-med-gpt-medrag-schain/config.json
#12
by
nghiemkythu
- opened
llava-med-gpt-medrag-schain/config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "
|
| 3 |
"architectures": [
|
| 4 |
"LlavaLlamaForCausalLM"
|
| 5 |
],
|
|
@@ -15,7 +15,7 @@
|
|
| 15 |
"mm_projector_type": "mlp2x_gelu",
|
| 16 |
"mm_use_im_start_end": true,
|
| 17 |
"mm_vision_select_layer": -2,
|
| 18 |
-
"mm_vision_tower": "/
|
| 19 |
"model_type": "llava",
|
| 20 |
"more_mlp": false,
|
| 21 |
"num_attention_heads": 32,
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "./weights_full/checkpoint_llava_med_instruct_60k_inline_mention_version_1-5",
|
| 3 |
"architectures": [
|
| 4 |
"LlavaLlamaForCausalLM"
|
| 5 |
],
|
|
|
|
| 15 |
"mm_projector_type": "mlp2x_gelu",
|
| 16 |
"mm_use_im_start_end": true,
|
| 17 |
"mm_vision_select_layer": -2,
|
| 18 |
+
"mm_vision_tower": "openai/clip-vit-large-patch14",
|
| 19 |
"model_type": "llava",
|
| 20 |
"more_mlp": false,
|
| 21 |
"num_attention_heads": 32,
|