Spaces:
Runtime error
Runtime error
ffreemt
commited on
Commit
·
0e7c90d
1
Parent(s):
1132eaf
Update block.load to update buff Textbox
Browse files
app.py
CHANGED
|
@@ -219,6 +219,12 @@ logger.info("start dl")
|
|
| 219 |
_ = """full url: https://huggingface.co/TheBloke/mpt-30B-chat-GGML/blob/main/mpt-30b-chat.ggmlv0.q4_1.bin"""
|
| 220 |
|
| 221 |
repo_id = "TheBloke/mpt-30B-chat-GGML"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 222 |
model_filename = "mpt-30b-chat.ggmlv0.q4_1.bin"
|
| 223 |
destination_folder = "models"
|
| 224 |
|
|
@@ -431,8 +437,8 @@ with gr.Blocks(
|
|
| 431 |
show_progress="full",
|
| 432 |
)
|
| 433 |
|
| 434 |
-
# update buff Textbox
|
| 435 |
-
block.
|
| 436 |
|
| 437 |
# concurrency_count=5, max_size=20
|
| 438 |
# max_size=36, concurrency_count=14
|
|
|
|
| 219 |
_ = """full url: https://huggingface.co/TheBloke/mpt-30B-chat-GGML/blob/main/mpt-30b-chat.ggmlv0.q4_1.bin"""
|
| 220 |
|
| 221 |
repo_id = "TheBloke/mpt-30B-chat-GGML"
|
| 222 |
+
|
| 223 |
+
# https://huggingface.co/TheBloke/mpt-30B-chat-GGML
|
| 224 |
+
_ = """
|
| 225 |
+
mpt-30b-chat.ggmlv0.q4_0.bin q4_0 4 16.85 GB 19.35 GB 4-bit.
|
| 226 |
+
mpt-30b-chat.ggmlv0.q4_1.bin q4_1 4 18.73 GB 21.23 GB 4-bit. Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models.
|
| 227 |
+
"""
|
| 228 |
model_filename = "mpt-30b-chat.ggmlv0.q4_1.bin"
|
| 229 |
destination_folder = "models"
|
| 230 |
|
|
|
|
| 437 |
show_progress="full",
|
| 438 |
)
|
| 439 |
|
| 440 |
+
# update buff Textbox, every: units in seconds)
|
| 441 |
+
block.run_forever(lambda: ns.response, None, [buff], every=1)
|
| 442 |
|
| 443 |
# concurrency_count=5, max_size=20
|
| 444 |
# max_size=36, concurrency_count=14
|