Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -8,8 +8,9 @@ from transformers import pipeline
|
|
| 8 |
#pipe = pipeline("text-generation", model="microsoft/Phi-3-mini-128k-instruct", trust_remote_code=True)
|
| 9 |
|
| 10 |
#client = InferenceClient("microsoft/Phi-3-mini-128k-instruct")
|
| 11 |
-
client = InferenceClient("HuggingFaceH4/zephyr-7b-beta")
|
| 12 |
#client = InferenceClient("google/gemma-1.1-7b-it")
|
|
|
|
| 13 |
|
| 14 |
|
| 15 |
@spaces.GPU
|
|
@@ -33,7 +34,7 @@ def respond(
|
|
| 33 |
|
| 34 |
response = ""
|
| 35 |
|
| 36 |
-
for message in
|
| 37 |
messages,
|
| 38 |
max_tokens=max_tokens,
|
| 39 |
stream=True,
|
|
|
|
| 8 |
#pipe = pipeline("text-generation", model="microsoft/Phi-3-mini-128k-instruct", trust_remote_code=True)
|
| 9 |
|
| 10 |
#client = InferenceClient("microsoft/Phi-3-mini-128k-instruct")
|
| 11 |
+
#client = InferenceClient("HuggingFaceH4/zephyr-7b-beta")
|
| 12 |
#client = InferenceClient("google/gemma-1.1-7b-it")
|
| 13 |
+
pipe = pipeline("text-generation", model="segolilylabs/Lily-Cybersecurity-7B-v0.2")
|
| 14 |
|
| 15 |
|
| 16 |
@spaces.GPU
|
|
|
|
| 34 |
|
| 35 |
response = ""
|
| 36 |
|
| 37 |
+
for message in pipe.chat_completion(
|
| 38 |
messages,
|
| 39 |
max_tokens=max_tokens,
|
| 40 |
stream=True,
|