Commit ·
bae1d90
1
Parent(s): 566b50e
Update app.py
Browse files
app.py
CHANGED
|
@@ -14,7 +14,7 @@ DEFAULT_MAX_NEW_TOKENS = 1024
|
|
| 14 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
| 15 |
|
| 16 |
DESCRIPTION = """\
|
| 17 |
-
Llama. Protected.
|
| 18 |
"""
|
| 19 |
|
| 20 |
if not torch.cuda.is_available():
|
|
|
|
| 14 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
| 15 |
|
| 16 |
DESCRIPTION = """\
|
| 17 |
+
# Llama. Protected. With Protecto.
|
| 18 |
"""
|
| 19 |
|
| 20 |
if not torch.cuda.is_available():
|