fthor commited on
Commit
bc91b52
·
1 Parent(s): 3ac1ccb

set temperature 0.3

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -19,7 +19,7 @@ model = LlavaForConditionalGeneration.from_pretrained(
19
  model_id,
20
  quantization_config=quantization_config,
21
  device_map="auto",
22
- use_flash_attention_2=True,
23
  low_cpu_mem_usage=True
24
  )
25
 
@@ -28,7 +28,7 @@ def text_to_image(image, prompt):
28
  prompt = f'USER: <image>\n{prompt}\nASSISTANT:'
29
 
30
  inputs = processor([prompt], images=[image], padding=True, return_tensors="pt").to(model.device)
31
- output = model.generate(**inputs, max_new_tokens=500)
32
  generated_text = processor.batch_decode(output, skip_special_tokens=True)
33
  text = generated_text.pop()
34
  text_output = text.split("ASSISTANT:")[-1]
 
19
  model_id,
20
  quantization_config=quantization_config,
21
  device_map="auto",
22
+ # use_flash_attention_2=True,
23
  low_cpu_mem_usage=True
24
  )
25
 
 
28
  prompt = f'USER: <image>\n{prompt}\nASSISTANT:'
29
 
30
  inputs = processor([prompt], images=[image], padding=True, return_tensors="pt").to(model.device)
31
+ output = model.generate(**inputs, max_new_tokens=500, temperature=0.3)
32
  generated_text = processor.batch_decode(output, skip_special_tokens=True)
33
  text = generated_text.pop()
34
  text_output = text.split("ASSISTANT:")[-1]