choco-conoz commited on
Commit
e117631
·
1 Parent(s): dad4e7a
Files changed (1) hide show
  1. src/streamlit_app.py +18 -17
src/streamlit_app.py CHANGED
@@ -12,29 +12,30 @@ from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
12
  # model_id = "meta-llama/Llama-3.2-1B"
13
  model_id = "choco-conoz/TwinLlama-3.1-8B"
14
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  processor = pipeline(
16
  "text-generation",
17
- model=model_id,
18
- model_kwargs={
19
- "torch_dtype": torch.float16,
20
- "quantization_config": {"load_in_4bit": True},
21
- "low_cpu_mem_usage": True,
22
- },
23
  )
 
24
  terminators = [
25
- processor.tokenizer.eos_token_id,
26
- processor.tokenizer.convert_tokens_to_ids(""),
27
  ]
28
 
29
- # tokenizer = AutoTokenizer.from_pretrained(model_id)
30
- # model = AutoModelForCausalLM.from_pretrained(model_id)
31
- # processor = pipeline(
32
- # "text-generation",
33
- # model=model,
34
- # tokenizer=tokenizer,
35
- # max_new_tokens=10
36
- # )
37
-
38
 
39
  def main():
40
  st.title('Text Generator')
 
12
  # model_id = "meta-llama/Llama-3.2-1B"
13
  model_id = "choco-conoz/TwinLlama-3.1-8B"
14
 
15
+ # processor = pipeline(
16
+ # "text-generation",
17
+ # model=model_id,
18
+ # model_kwargs={
19
+ # "torch_dtype": torch.float16,
20
+ # "quantization_config": {"load_in_4bit": True},
21
+ # "low_cpu_mem_usage": True,
22
+ # },
23
+ # )
24
+
25
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
26
+ model = AutoModelForCausalLM.from_pretrained(model_id)
27
  processor = pipeline(
28
  "text-generation",
29
+ model=model,
30
+ tokenizer=tokenizer,
31
+ max_new_tokens=10
 
 
 
32
  )
33
+
34
  terminators = [
35
+ tokenizer.eos_token_id,
36
+ tokenizer.convert_tokens_to_ids(""),
37
  ]
38
 
 
 
 
 
 
 
 
 
 
39
 
40
  def main():
41
  st.title('Text Generator')