Update app.py
Browse files
app.py
CHANGED
|
@@ -10,9 +10,18 @@ gpt2_tkn=GPT2Tokenizer.from_pretrained('gpt2')
|
|
| 10 |
# response = gpt2_tensors
|
| 11 |
# return response
|
| 12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
def generate(starting_text):
|
| 14 |
tkn_ids = gpt2_tkn.encode(starting_text, return_tensors = 'pt')
|
| 15 |
-
gpt2_tensors = mdl.generate(tkn_ids)
|
| 16 |
response=""
|
| 17 |
#response = gpt2_tensors
|
| 18 |
for i, x in enumerate(gpt2_tensors):
|
|
|
|
| 10 |
# response = gpt2_tensors
|
| 11 |
# return response
|
| 12 |
|
| 13 |
+
# def generate(starting_text):
|
| 14 |
+
# tkn_ids = gpt2_tkn.encode(starting_text, return_tensors = 'pt')
|
| 15 |
+
# gpt2_tensors = mdl.generate(tkn_ids)
|
| 16 |
+
# response=""
|
| 17 |
+
# #response = gpt2_tensors
|
| 18 |
+
# for i, x in enumerate(gpt2_tensors):
|
| 19 |
+
# response=response+f"{i}: {gpt2_tkn.decode(x, skip_special_tokens=True)}"
|
| 20 |
+
# return response
|
| 21 |
+
|
| 22 |
def generate(starting_text):
|
| 23 |
tkn_ids = gpt2_tkn.encode(starting_text, return_tensors = 'pt')
|
| 24 |
+
gpt2_tensors = mdl.generate(tkn_ids,max_length=100)
|
| 25 |
response=""
|
| 26 |
#response = gpt2_tensors
|
| 27 |
for i, x in enumerate(gpt2_tensors):
|