Spaces:
Sleeping
Sleeping
Commit
·
cc6a446
1
Parent(s):
28433a3
Update app.py
Browse files
app.py
CHANGED
|
@@ -29,12 +29,17 @@ llm = ctransformers.AutoModelForCausalLM.from_pretrained('./llama-2-7b.ggmlv3.q4
|
|
| 29 |
print("Loaded model")
|
| 30 |
|
| 31 |
def complete(prompt, stop=["User", "Assistant"]):
|
|
|
|
| 32 |
tokens = llm.tokenize(prompt)
|
|
|
|
| 33 |
token_count = 0
|
| 34 |
output = ''
|
|
|
|
| 35 |
for token in llm.generate(tokens):
|
|
|
|
| 36 |
token_count += 1
|
| 37 |
result = llm.detokenize(token)
|
|
|
|
| 38 |
output += result
|
| 39 |
for word in stop:
|
| 40 |
if word in output:
|
|
|
|
| 29 |
print("Loaded model")
|
| 30 |
|
| 31 |
def complete(prompt, stop=["User", "Assistant"]):
|
| 32 |
+
print("0")
|
| 33 |
tokens = llm.tokenize(prompt)
|
| 34 |
+
print("1")
|
| 35 |
token_count = 0
|
| 36 |
output = ''
|
| 37 |
+
print("2")
|
| 38 |
for token in llm.generate(tokens):
|
| 39 |
+
print("tokens")
|
| 40 |
token_count += 1
|
| 41 |
result = llm.detokenize(token)
|
| 42 |
+
print("detokens")
|
| 43 |
output += result
|
| 44 |
for word in stop:
|
| 45 |
if word in output:
|