Spaces:
Runtime error
Runtime error
Hritik
commited on
Commit
·
cfe5653
1
Parent(s):
6ab097e
load on cpu for memory requirements
Browse files
app.py
CHANGED
|
@@ -35,11 +35,14 @@ processor = MplugOwlProcessor(image_processor, tokenizer)
|
|
| 35 |
model = MplugOwlForConditionalGeneration.from_pretrained(
|
| 36 |
pretrained_ckpt,
|
| 37 |
torch_dtype=torch.bfloat16,
|
| 38 |
-
device_map={'':
|
|
|
|
| 39 |
)
|
| 40 |
|
| 41 |
-
for name, param in model.named_parameters():
|
| 42 |
-
|
|
|
|
|
|
|
| 43 |
peft_config = LoraConfig(
|
| 44 |
target_modules=r'.*language_model.*\.(q_proj|v_proj|k_proj|o_proj|gate_proj|down_proj|up_proj)',
|
| 45 |
inference_mode=True,
|
|
@@ -50,14 +53,14 @@ peft_config = LoraConfig(
|
|
| 50 |
model = get_peft_model(model, peft_config)
|
| 51 |
model.print_trainable_parameters()
|
| 52 |
with open(trained_ckpt, 'rb') as f:
|
| 53 |
-
ckpt = torch.load(f, map_location = torch.device(
|
| 54 |
model.load_state_dict(ckpt)
|
| 55 |
-
model = model.to(torch.bfloat16)
|
| 56 |
print('Model Loaded')
|
| 57 |
|
| 58 |
PROMPT = """The following is a conversation between a curious human and AI assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
|
| 59 |
Human: <|video|>
|
| 60 |
-
Human: Does this video entail the description: ""A
|
| 61 |
AI: """
|
| 62 |
|
| 63 |
valid_data = MultiModalDataset("examples/y5xuvHpDPZQ_000005_000015.mp4", PROMPT, tokenizer, processor, max_length = 256, loss_objective = 'sequential')
|
|
|
|
| 35 |
model = MplugOwlForConditionalGeneration.from_pretrained(
|
| 36 |
pretrained_ckpt,
|
| 37 |
torch_dtype=torch.bfloat16,
|
| 38 |
+
device_map={'': 'cpu'}
|
| 39 |
+
# device_map={'':0}
|
| 40 |
)
|
| 41 |
|
| 42 |
+
# for name, param in model.named_parameters():
|
| 43 |
+
# print(param.device)
|
| 44 |
+
# break
|
| 45 |
+
|
| 46 |
peft_config = LoraConfig(
|
| 47 |
target_modules=r'.*language_model.*\.(q_proj|v_proj|k_proj|o_proj|gate_proj|down_proj|up_proj)',
|
| 48 |
inference_mode=True,
|
|
|
|
| 53 |
model = get_peft_model(model, peft_config)
|
| 54 |
model.print_trainable_parameters()
|
| 55 |
with open(trained_ckpt, 'rb') as f:
|
| 56 |
+
ckpt = torch.load(f, map_location = torch.device("cpu"))
|
| 57 |
model.load_state_dict(ckpt)
|
| 58 |
+
model = model.to("cuda:0").to(torch.bfloat16)
|
| 59 |
print('Model Loaded')
|
| 60 |
|
| 61 |
PROMPT = """The following is a conversation between a curious human and AI assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
|
| 62 |
Human: <|video|>
|
| 63 |
+
Human: Does this video entail the description: ""A soccer team walking off the field while the audience claps.""?
|
| 64 |
AI: """
|
| 65 |
|
| 66 |
valid_data = MultiModalDataset("examples/y5xuvHpDPZQ_000005_000015.mp4", PROMPT, tokenizer, processor, max_length = 256, loss_objective = 'sequential')
|