revert to origin
Browse files
config/infer.yaml
CHANGED
|
@@ -2,6 +2,7 @@ device: 'cuda'
|
|
| 2 |
|
| 3 |
vicuna:
|
| 4 |
model_path: '/home/user/app/vicuna-7b'
|
|
|
|
| 5 |
device: 'cuda'
|
| 6 |
num_gpus: 'auto'
|
| 7 |
max_gpu_memory: '24Gib'
|
|
@@ -11,6 +12,7 @@ vicuna:
|
|
| 11 |
max_new_tokens: 512
|
| 12 |
debug: False
|
| 13 |
output_path: '/home/user/app/vicuna_out.json'
|
|
|
|
| 14 |
|
| 15 |
vid2seq:
|
| 16 |
enable: False
|
|
|
|
| 2 |
|
| 3 |
vicuna:
|
| 4 |
model_path: '/home/user/app/vicuna-7b'
|
| 5 |
+
# model_path: '/mnt/petrelfs/wangyiqin/vid_cap/ChatVID_huggingface/vicuna-7b'
|
| 6 |
device: 'cuda'
|
| 7 |
num_gpus: 'auto'
|
| 8 |
max_gpu_memory: '24Gib'
|
|
|
|
| 12 |
max_new_tokens: 512
|
| 13 |
debug: False
|
| 14 |
output_path: '/home/user/app/vicuna_out.json'
|
| 15 |
+
# output_path: '/mnt/petrelfs/wangyiqin/vid_cap/ChatVID_huggingface/vicuna_out.json'
|
| 16 |
|
| 17 |
vid2seq:
|
| 18 |
enable: False
|
model/fastchat/serve/compression.py
CHANGED
|
@@ -28,7 +28,7 @@ class CLinear(nn.Module):
|
|
| 28 |
def __init__(self, weight, bias, device):
|
| 29 |
super().__init__()
|
| 30 |
|
| 31 |
-
self.weight = compress(weight.
|
| 32 |
self.bias = bias
|
| 33 |
|
| 34 |
def forward(self, input: Tensor) -> Tensor:
|
|
|
|
| 28 |
def __init__(self, weight, bias, device):
|
| 29 |
super().__init__()
|
| 30 |
|
| 31 |
+
self.weight = compress(weight.data.to(device), default_compression_config)
|
| 32 |
self.bias = bias
|
| 33 |
|
| 34 |
def forward(self, input: Tensor) -> Tensor:
|
model/vision/ImageCaptioner.py
CHANGED
|
@@ -15,6 +15,11 @@ class ImageCaptioner:
|
|
| 15 |
self.model = Blip2ForConditionalGeneration.from_pretrained(
|
| 16 |
"/home/user/app/pretrained_models/blip2-opt-2.7b",
|
| 17 |
torch_dtype=self.data_type, device_map="auto")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
|
| 19 |
def __call__(self, imgs):
|
| 20 |
inputs = self.processor(
|
|
|
|
| 15 |
self.model = Blip2ForConditionalGeneration.from_pretrained(
|
| 16 |
"/home/user/app/pretrained_models/blip2-opt-2.7b",
|
| 17 |
torch_dtype=self.data_type, device_map="auto")
|
| 18 |
+
# self.processor = Blip2Processor.from_pretrained(
|
| 19 |
+
# "/mnt/petrelfs/wangyiqin/vid_cap/ChatVID_huggingface/pretrained_models/blip2-opt-2.7b")
|
| 20 |
+
# self.model = Blip2ForConditionalGeneration.from_pretrained(
|
| 21 |
+
# "/mnt/petrelfs/wangyiqin/vid_cap/ChatVID_huggingface/pretrained_models/blip2-opt-2.7b",
|
| 22 |
+
# torch_dtype=self.data_type, device_map="auto")
|
| 23 |
|
| 24 |
def __call__(self, imgs):
|
| 25 |
inputs = self.processor(
|