cminja commited on
Commit
a1e2f92
·
verified ·
1 Parent(s): 6e98e2c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -3
README.md CHANGED
@@ -1,3 +1,32 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+
5
+
6
+ Usage:
7
+
8
+ import torch
9
+ from transformers import AutoModelForCausalLM, AutoTokenizer
10
+ device = "cuda" # or "cpu"
11
+ model_path = "cminja/granitte-8b-code-instruct"
12
+ tokenizer = AutoTokenizer.from_pretrained(model_path)
13
+ # drop device_map if running on CPU
14
+ model = AutoModelForCausalLM.from_pretrained(model_path, device_map=device)
15
+ model.eval()
16
+ # change input text as desired
17
+ chat = [
18
+ { "role": "user", "content": "Write a code to find the maximum value in a list of numbers." },
19
+ ]
20
+ chat = tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=True)
21
+ # tokenize the text
22
+ input_tokens = tokenizer(chat, return_tensors="pt")
23
+ # transfer tokenized inputs to the device
24
+ for i in input_tokens:
25
+ input_tokens[i] = input_tokens[i].to(device)
26
+ # generate output tokens
27
+ output = model.generate(**input_tokens, max_new_tokens=100)
28
+ # decode output tokens into text
29
+ output = tokenizer.batch_decode(output)
30
+ # loop over the batch to print, in this example the batch size is 1
31
+ for i in output:
32
+ print(i)