Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -11,7 +11,7 @@ from peft import PeftModel
|
|
| 11 |
|
| 12 |
# Constants
|
| 13 |
MAX_MAX_NEW_TOKENS = 2048
|
| 14 |
-
DEFAULT_MAX_NEW_TOKENS =
|
| 15 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
| 16 |
|
| 17 |
# Description and License Texts
|
|
@@ -30,7 +30,7 @@ if not torch.cuda.is_available():
|
|
| 30 |
DESCRIPTION += "\n<p>Running on CPU 🥶 This demo does not work on CPU.</p>"
|
| 31 |
|
| 32 |
# Model and Tokenizer Configuration
|
| 33 |
-
model_id = "meta-llama/Llama-2-7b-
|
| 34 |
bnb_config = BitsAndBytesConfig(
|
| 35 |
load_in_4bit=True,
|
| 36 |
bnb_4bit_use_double_quant=False,
|
|
|
|
| 11 |
|
| 12 |
# Constants
|
| 13 |
MAX_MAX_NEW_TOKENS = 2048
|
| 14 |
+
DEFAULT_MAX_NEW_TOKENS = 950
|
| 15 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
| 16 |
|
| 17 |
# Description and License Texts
|
|
|
|
| 30 |
DESCRIPTION += "\n<p>Running on CPU 🥶 This demo does not work on CPU.</p>"
|
| 31 |
|
| 32 |
# Model and Tokenizer Configuration
|
| 33 |
+
model_id = "meta-llama/Llama-2-7b-hf"
|
| 34 |
bnb_config = BitsAndBytesConfig(
|
| 35 |
load_in_4bit=True,
|
| 36 |
bnb_4bit_use_double_quant=False,
|