Ishgan commited on
Commit
ba92118
Β·
verified Β·
1 Parent(s): 24665ce

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -0
app.py CHANGED
@@ -2,8 +2,13 @@ import requests
2
  from PIL import Image
3
 
4
  import torch
 
5
  from transformers import AutoProcessor, LlavaForConditionalGeneration
6
 
 
 
 
 
7
  model_id = "llava-hf/llava-1.5-7b-hf"
8
  model = LlavaForConditionalGeneration.from_pretrained(
9
  model_id,
@@ -11,6 +16,9 @@ model = LlavaForConditionalGeneration.from_pretrained(
11
  low_cpu_mem_usage=True,
12
  ).to(0)
13
 
 
 
 
14
  processor = AutoProcessor.from_pretrained(model_id)
15
 
16
  # Define a chat history and use `apply_chat_template` to get correctly formatted prompt
 
2
  from PIL import Image
3
 
4
  import torch
5
+
6
  from transformers import AutoProcessor, LlavaForConditionalGeneration
7
 
8
+ =
9
+
10
+
11
+
12
  model_id = "llava-hf/llava-1.5-7b-hf"
13
  model = LlavaForConditionalGeneration.from_pretrained(
14
  model_id,
 
16
  low_cpu_mem_usage=True,
17
  ).to(0)
18
 
19
+ device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
20
+ model = model.to(device)
21
+
22
  processor = AutoProcessor.from_pretrained(model_id)
23
 
24
  # Define a chat history and use `apply_chat_template` to get correctly formatted prompt