Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -7,17 +7,17 @@ urlretrieve("https://github.com/equ1/generative_python_transformer/tree/main/GPT
|
|
| 7 |
|
| 8 |
# inference function
|
| 9 |
def inference(inp):
|
| 10 |
-
tokenizer = AutoTokenizer.from_pretrained("GPT-python")
|
| 11 |
-
model = AutoModelWithLMHead.from_pretrained("GPT-python")
|
| 12 |
|
| 13 |
-
input_ids = tokenizer.encode(inp, return_tensors="pt")
|
| 14 |
beam_output = model.generate(input_ids,
|
| 15 |
max_length=512,
|
| 16 |
num_beams=10,
|
| 17 |
temperature=0.7,
|
| 18 |
no_repeat_ngram_size=5,
|
| 19 |
num_return_sequences=1,
|
| 20 |
-
)
|
| 21 |
|
| 22 |
output = []
|
| 23 |
for beam in beam_output:
|
|
|
|
| 7 |
|
| 8 |
# inference function
|
| 9 |
def inference(inp):
|
| 10 |
+
tokenizer = AutoTokenizer.from_pretrained("GPT-python").to("cuda")
|
| 11 |
+
model = AutoModelWithLMHead.from_pretrained("GPT-python").to("cuda")
|
| 12 |
|
| 13 |
+
input_ids = tokenizer.encode(inp, return_tensors="pt").to("cuda")
|
| 14 |
beam_output = model.generate(input_ids,
|
| 15 |
max_length=512,
|
| 16 |
num_beams=10,
|
| 17 |
temperature=0.7,
|
| 18 |
no_repeat_ngram_size=5,
|
| 19 |
num_return_sequences=1,
|
| 20 |
+
).to("cuda")
|
| 21 |
|
| 22 |
output = []
|
| 23 |
for beam in beam_output:
|