Update app.py
Browse files
app.py
CHANGED
|
@@ -165,7 +165,8 @@ def get_image_tensors_batch_cpu(images, image_processor, model=None):
|
|
| 165 |
image_tensors.append(image_tensor)
|
| 166 |
|
| 167 |
if len(image_tensors) == 1:
|
| 168 |
-
|
|
|
|
| 169 |
else:
|
| 170 |
return torch.stack(image_tensors, dim=0)
|
| 171 |
|
|
@@ -204,15 +205,15 @@ ccd_eval = ccd_eval_cpu_wrapper
|
|
| 204 |
# Global Configuration
|
| 205 |
# =========================================
|
| 206 |
MODEL_CATALOGUE = {
|
| 207 |
-
"MAIRA-2
|
| 208 |
-
"Libra-v1.0-3B": "X-iZhang/libra-v1.0-3b",
|
| 209 |
"Libra-v1.0-7B": "X-iZhang/libra-v1.0-7b",
|
| 210 |
"LLaVA-Med-v1.5": "X-iZhang/libra-llava-med-v1.5-mistral-7b",
|
| 211 |
"LLaVA-Rad": "X-iZhang/libra-llava-rad",
|
| 212 |
"Med-CXRGen-F": "X-iZhang/Med-CXRGen-F",
|
| 213 |
"Med-CXRGen-I": "X-iZhang/Med-CXRGen-I"
|
| 214 |
}
|
| 215 |
-
DEFAULT_MODEL_NAME = "
|
| 216 |
_loaded_models = {}
|
| 217 |
|
| 218 |
|
|
|
|
| 165 |
image_tensors.append(image_tensor)
|
| 166 |
|
| 167 |
if len(image_tensors) == 1:
|
| 168 |
+
img = image_tensors[0].unsqueeze(0)
|
| 169 |
+
return torch.cat([img, img.clone()], dim=0)
|
| 170 |
else:
|
| 171 |
return torch.stack(image_tensors, dim=0)
|
| 172 |
|
|
|
|
| 205 |
# Global Configuration
|
| 206 |
# =========================================
|
| 207 |
MODEL_CATALOGUE = {
|
| 208 |
+
"MAIRA-2": "X-iZhang/libra-maira-2",
|
| 209 |
+
"Libra-v1.0-3B (⚡Recommended for CPU)": "X-iZhang/libra-v1.0-3b",
|
| 210 |
"Libra-v1.0-7B": "X-iZhang/libra-v1.0-7b",
|
| 211 |
"LLaVA-Med-v1.5": "X-iZhang/libra-llava-med-v1.5-mistral-7b",
|
| 212 |
"LLaVA-Rad": "X-iZhang/libra-llava-rad",
|
| 213 |
"Med-CXRGen-F": "X-iZhang/Med-CXRGen-F",
|
| 214 |
"Med-CXRGen-I": "X-iZhang/Med-CXRGen-I"
|
| 215 |
}
|
| 216 |
+
DEFAULT_MODEL_NAME = "Libra-v1.0-3B (⚡Recommended for CPU)"
|
| 217 |
_loaded_models = {}
|
| 218 |
|
| 219 |
|