Spaces:
No application file
No application file
Update app.py
Browse files
app.py
CHANGED
|
@@ -11,9 +11,12 @@ max_seq_length = 2048
|
|
| 11 |
tokenizer = AutoTokenizer.from_pretrained("ua-l/gemma-2-9b-legal-steps200-merged-16bit-uk")
|
| 12 |
model = AutoModelForCausalLM.from_pretrained(
|
| 13 |
"ua-l/gemma-2-9b-legal-steps200-merged-16bit-uk",
|
| 14 |
-
device_map='auto'
|
|
|
|
| 15 |
)
|
| 16 |
|
|
|
|
|
|
|
| 17 |
|
| 18 |
@spaces.GPU
|
| 19 |
def predict(question):
|
|
|
|
| 11 |
tokenizer = AutoTokenizer.from_pretrained("ua-l/gemma-2-9b-legal-steps200-merged-16bit-uk")
|
| 12 |
model = AutoModelForCausalLM.from_pretrained(
|
| 13 |
"ua-l/gemma-2-9b-legal-steps200-merged-16bit-uk",
|
| 14 |
+
device_map='auto',
|
| 15 |
+
torch_dtype='auto',
|
| 16 |
)
|
| 17 |
|
| 18 |
+
print('Model dtype:', model.dtype)
|
| 19 |
+
|
| 20 |
|
| 21 |
@spaces.GPU
|
| 22 |
def predict(question):
|