Update app.py
Browse files
app.py
CHANGED
|
@@ -22,6 +22,7 @@ DEVICE = "cuda" if torch.cuda.is_available() else "cpu"
|
|
| 22 |
|
| 23 |
_tok = None
|
| 24 |
_model = None
|
|
|
|
| 25 |
|
| 26 |
def _lazy_load() -> Tuple[AutoTokenizer, AutoModelForCausalLM]:
|
| 27 |
global _tok, _model
|
|
|
|
| 22 |
|
| 23 |
_tok = None
|
| 24 |
_model = None
|
| 25 |
+
_salamandra = None
|
| 26 |
|
| 27 |
def _lazy_load() -> Tuple[AutoTokenizer, AutoModelForCausalLM]:
|
| 28 |
global _tok, _model
|