EduuGomes commited on
Commit
e54949b
·
verified ·
1 Parent(s): 40ab165

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +23 -13
app.py CHANGED
@@ -1,16 +1,26 @@
1
- import torch
2
- from transformers import AutoModelForCausalLM, AutoTokenizer
3
-
4
- torch.set_default_device("cuda")
5
 
6
- model = AutoModelForCausalLM.from_pretrained("microsoft/phi-2", torch_dtype="auto", trust_remote_code=True)
7
- tokenizer = AutoTokenizer.from_pretrained("microsoft/phi-2", trust_remote_code=True)
8
 
9
- inputs = tokenizer('''def print_prime(n):
10
- """
11
- Print all primes between 1 and n
12
- """''', return_tensors="pt", return_attention_mask=False)
13
 
14
- outputs = model.generate(**inputs, max_length=200)
15
- text = tokenizer.batch_decode(outputs)[0]
16
- print(text)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Install transformers from source - only needed for versions <= v4.34
2
+ # pip install git+https://github.com/huggingface/transformers.git
3
+ # pip install accelerate
 
4
 
5
+ import torch
6
+ from transformers import pipeline
7
 
8
+ pipe = pipeline("text-generation", model="TinyLlama/TinyLlama-1.1B-Chat-v1.0", torch_dtype=torch.bfloat16, device_map="auto")
 
 
 
9
 
10
+ # We use the tokenizer's chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating
11
+ messages = [
12
+ {
13
+ "role": "system",
14
+ "content": "You are a friendly chatbot who always responds in the style of a pirate",
15
+ },
16
+ {"role": "user", "content": "How many helicopters can a human eat in one sitting?"},
17
+ ]
18
+ prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
19
+ outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
20
+ print(outputs[0]["generated_text"])
21
+ # <|system|>
22
+ # You are a friendly chatbot who always responds in the style of a pirate.</s>
23
+ # <|user|>
24
+ # How many helicopters can a human eat in one sitting?</s>
25
+ # <|assistant|>
26
+ # ...