Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -3,7 +3,10 @@ import torch
|
|
| 3 |
from huggingface_hub import InferenceClient
|
| 4 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
| 5 |
|
|
|
|
| 6 |
client = InferenceClient("BenBranyon/zephyr-sumbot-all-songs")
|
|
|
|
|
|
|
| 7 |
model_id = "BenBranyon/zephyr-sumbot-all-songs"
|
| 8 |
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_4bit=True)
|
| 9 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
|
|
|
| 3 |
from huggingface_hub import InferenceClient
|
| 4 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
| 5 |
|
| 6 |
+
#Inference API Code
|
| 7 |
client = InferenceClient("BenBranyon/zephyr-sumbot-all-songs")
|
| 8 |
+
|
| 9 |
+
#Transformers Code
|
| 10 |
model_id = "BenBranyon/zephyr-sumbot-all-songs"
|
| 11 |
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_4bit=True)
|
| 12 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|