blackhole33 commited on
Commit
343fa25
·
verified ·
1 Parent(s): f8a6765

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -0
README.md CHANGED
@@ -45,3 +45,22 @@ outputs = model.generate(**inputs, max_new_tokens = 64, use_cache = True)
45
  tokenizer.batch_decode(outputs)
46
 
47
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  tokenizer.batch_decode(outputs)
46
 
47
  ```
48
+
49
+ # Streamly generate response:
50
+
51
+ ```
52
+ FastLanguageModel.for_inference(model) # Enable native 2x faster inference
53
+ inputs = tokenizer(
54
+ [
55
+ alpaca_prompt.format(
56
+ "Continue the fibonnaci sequence.", # instruction
57
+ "1, 1, 2, 3, 5, 8", # input
58
+ "", # output - leave this blank for generation!
59
+ )
60
+ ], return_tensors = "pt").to("cuda")
61
+
62
+ from transformers import TextStreamer
63
+ text_streamer = TextStreamer(tokenizer)
64
+ _ = model.generate(**inputs, streamer = text_streamer, max_new_tokens = 128)
65
+
66
+ ```