choco-conoz commited on
Commit
a8d11ef
·
1 Parent(s): 478a68e

feat: change pipeline

Browse files
Files changed (1) hide show
  1. src/streamlit_app.py +26 -15
src/streamlit_app.py CHANGED
@@ -1,7 +1,7 @@
1
  import os
2
  import streamlit as st
3
  import torch
4
- from transformers import pipeline
5
  # from huggingface_hub import notebook_login
6
  # from unsloth import FastLanguageModel, is_bfloat16_supported
7
 
@@ -12,21 +12,29 @@ from transformers import pipeline
12
  # model_id = "meta-llama/Llama-3.2-1B"
13
  model_id = "choco-conoz/TwinLlama-3.1-8B"
14
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  processor = pipeline(
16
  "text-generation",
17
- model=model_id,
18
- model_kwargs={
19
- "torch_dtype": torch.float16,
20
- "quantization_config": {"load_in_4bit": True},
21
- "low_cpu_mem_usage": True,
22
- },
23
  )
24
 
25
- terminators = [
26
- processor.tokenizer.eos_token_id,
27
- processor.tokenizer.convert_tokens_to_ids(""),
28
- ]
29
-
30
 
31
  def main():
32
  st.title('Text Generator')
@@ -46,9 +54,12 @@ def main():
46
  print('user_prompt', user_prompt)
47
  prompt = processor.tokenizer.apply_chat_template(
48
  user_prompt, tokenize=False, add_generation_prompt=True)
49
- outputs = processor(prompt, max_new_tokens=4096, eos_token_id=terminators, do_sample=True,
50
- temperature=0.6, top_p=0.9
51
- )
 
 
 
52
  response = outputs[0]["generated_text"][len(prompt):]
53
  st.write(response)
54
 
 
1
  import os
2
  import streamlit as st
3
  import torch
4
+ from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
5
  # from huggingface_hub import notebook_login
6
  # from unsloth import FastLanguageModel, is_bfloat16_supported
7
 
 
12
  # model_id = "meta-llama/Llama-3.2-1B"
13
  model_id = "choco-conoz/TwinLlama-3.1-8B"
14
 
15
+ # processor = pipeline(
16
+ # "text-generation",
17
+ # model=model_id,
18
+ # model_kwargs={
19
+ # "torch_dtype": torch.float16,
20
+ # "quantization_config": {"load_in_4bit": True},
21
+ # "low_cpu_mem_usage": True,
22
+ # },
23
+ # )
24
+ # terminators = [
25
+ # processor.tokenizer.eos_token_id,
26
+ # processor.tokenizer.convert_tokens_to_ids(""),
27
+ # ]
28
+
29
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
30
+ model = AutoModelForCausalLM.from_pretrained(model_id)
31
  processor = pipeline(
32
  "text-generation",
33
+ model=model,
34
+ tokenizer=tokenizer,
35
+ max_new_tokens=10
 
 
 
36
  )
37
 
 
 
 
 
 
38
 
39
  def main():
40
  st.title('Text Generator')
 
54
  print('user_prompt', user_prompt)
55
  prompt = processor.tokenizer.apply_chat_template(
56
  user_prompt, tokenize=False, add_generation_prompt=True)
57
+ # prompt = user_prompt
58
+ outputs = processor(prompt)
59
+
60
+ # outputs = processor(prompt, max_new_tokens=4096, eos_token_id=terminators, do_sample=True,
61
+ # temperature=0.6, top_p=0.9
62
+ # )
63
  response = outputs[0]["generated_text"][len(prompt):]
64
  st.write(response)
65