Fix: remove pyc binary files and update gitignore
Browse files- .gitignore +1 -0
- demo_brainvlm_folder_files.py → app.py +1 -1
- checkpoint_0.pth +3 -0
- checkpoint_1.pth +3 -0
- eval_configs/minigpt4_eval.yaml +1 -1
- minigpt4/__pycache__/__init__.cpython-310.pyc +0 -0
- minigpt4/__pycache__/__init__.cpython-312.pyc +0 -0
- minigpt4/__pycache__/__init__.cpython-38.pyc +0 -0
- minigpt4/configs/models/minigpt4_vicuna0.yaml +1 -1
- minigpt4/conversation/__pycache__/conversation.cpython-39.pyc +0 -0
- requirements.txt +32 -0
.gitignore
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
**/__pycache__/
|
demo_brainvlm_folder_files.py → app.py
RENAMED
|
@@ -955,7 +955,7 @@ import copy
|
|
| 955 |
|
| 956 |
# cfg_model_2 = cfg.model_cfg
|
| 957 |
# cfg_model_2['ckpt'] = '/home/ynwang/brain_llm/MiniGPT-4/eval_configs/minigpt4_eval_report.yaml'
|
| 958 |
-
model_config.ckpt = '
|
| 959 |
model_2 = model_cls.from_config(model_config).to('cuda:{}'.format(args.gpu_id))
|
| 960 |
|
| 961 |
# model_2 = copy.deepcopy(model.to('cpu')).to('cuda:{}'.format(args.gpu_id+1))
|
|
|
|
| 955 |
|
| 956 |
# cfg_model_2 = cfg.model_cfg
|
| 957 |
# cfg_model_2['ckpt'] = '/home/ynwang/brain_llm/MiniGPT-4/eval_configs/minigpt4_eval_report.yaml'
|
| 958 |
+
model_config.ckpt = './checkpoint_1.pth'
|
| 959 |
model_2 = model_cls.from_config(model_config).to('cuda:{}'.format(args.gpu_id))
|
| 960 |
|
| 961 |
# model_2 = copy.deepcopy(model.to('cpu')).to('cuda:{}'.format(args.gpu_id+1))
|
checkpoint_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b686015efa2e881ea64245e67ee5f2e55868765389b3e4a38ca3e4dbf73d2e98
|
| 3 |
+
size 313018367
|
checkpoint_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2f5da752c50173f4a198ecf7448561cf07c0e0b0b68bc038ea4f08a0c00fc053
|
| 3 |
+
size 313018687
|
eval_configs/minigpt4_eval.yaml
CHANGED
|
@@ -11,7 +11,7 @@ model:
|
|
| 11 |
# ckpt: /home/ynwang/MiniGPT-4/minigpt4/ckpt/lora_128/checkpoint_7.pth
|
| 12 |
# ckpt: "/disk1/Data/Medical-Ours/Brain_Tumor_data/ckpt/brain_llm/output/minigpt4_stage2_finetune/20250521154/checkpoint_5.pth"
|
| 13 |
# ckpt: "/home/ynwang/brain_llm/MiniGPT-4/minigpt4/ckpt/20250611193/checkpoint_1.pth"
|
| 14 |
-
ckpt: "
|
| 15 |
datasets:
|
| 16 |
# tcia_classification:
|
| 17 |
# batch_size: 1
|
|
|
|
| 11 |
# ckpt: /home/ynwang/MiniGPT-4/minigpt4/ckpt/lora_128/checkpoint_7.pth
|
| 12 |
# ckpt: "/disk1/Data/Medical-Ours/Brain_Tumor_data/ckpt/brain_llm/output/minigpt4_stage2_finetune/20250521154/checkpoint_5.pth"
|
| 13 |
# ckpt: "/home/ynwang/brain_llm/MiniGPT-4/minigpt4/ckpt/20250611193/checkpoint_1.pth"
|
| 14 |
+
ckpt: "./checkpoint_0.pth"
|
| 15 |
datasets:
|
| 16 |
# tcia_classification:
|
| 17 |
# batch_size: 1
|
minigpt4/__pycache__/__init__.cpython-310.pyc
DELETED
|
Binary file (1.02 kB)
|
|
|
minigpt4/__pycache__/__init__.cpython-312.pyc
DELETED
|
Binary file (1.66 kB)
|
|
|
minigpt4/__pycache__/__init__.cpython-38.pyc
DELETED
|
Binary file (1.02 kB)
|
|
|
minigpt4/configs/models/minigpt4_vicuna0.yaml
CHANGED
|
@@ -16,7 +16,7 @@ model:
|
|
| 16 |
prompt: ""
|
| 17 |
|
| 18 |
# llama_model: "/home/haoran/yinong/mnigpt-4/ckpt/vicuna-7b/"
|
| 19 |
-
llama_model: "
|
| 20 |
|
| 21 |
preprocess:
|
| 22 |
vis_processor:
|
|
|
|
| 16 |
prompt: ""
|
| 17 |
|
| 18 |
# llama_model: "/home/haoran/yinong/mnigpt-4/ckpt/vicuna-7b/"
|
| 19 |
+
llama_model: "meta-llama/Llama-3.1-8B-Instruct"
|
| 20 |
|
| 21 |
preprocess:
|
| 22 |
vis_processor:
|
minigpt4/conversation/__pycache__/conversation.cpython-39.pyc
CHANGED
|
Binary files a/minigpt4/conversation/__pycache__/conversation.cpython-39.pyc and b/minigpt4/conversation/__pycache__/conversation.cpython-39.pyc differ
|
|
|
requirements.txt
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
--extra-index-url https://download.pytorch.org/whl/cu118
|
| 2 |
+
torch==2.0.0+cu118
|
| 3 |
+
torchaudio==2.0.1+cu118
|
| 4 |
+
torchvision==0.15.1+cu118
|
| 5 |
+
huggingface-hub==0.29.2
|
| 6 |
+
matplotlib==3.7.0
|
| 7 |
+
psutil==5.9.4
|
| 8 |
+
iopath
|
| 9 |
+
pyyaml==6.0
|
| 10 |
+
regex==2022.10.31
|
| 11 |
+
tokenizers==0.20.3
|
| 12 |
+
tqdm==4.64.1
|
| 13 |
+
transformers==4.45.2
|
| 14 |
+
timm==0.6.13
|
| 15 |
+
webdataset==0.2.48
|
| 16 |
+
omegaconf==2.3.0
|
| 17 |
+
opencv-python==4.7.0.72
|
| 18 |
+
decord==0.6.0
|
| 19 |
+
peft==0.2.0
|
| 20 |
+
sentence-transformers
|
| 21 |
+
gradio==4.44.1
|
| 22 |
+
accelerate==1.4.0
|
| 23 |
+
bitsandbytes==0.37.0
|
| 24 |
+
scikit-image
|
| 25 |
+
visual-genome
|
| 26 |
+
wandb
|
| 27 |
+
nibabel==5.3.2
|
| 28 |
+
open_clip_torch==2.31.0
|
| 29 |
+
einops==0.8.1
|
| 30 |
+
monai==1.4.0
|
| 31 |
+
pydicom==2.4.4
|
| 32 |
+
torchio==0.20.19
|