Update app.py
Browse files
app.py
CHANGED
|
@@ -109,11 +109,11 @@ def ra(user_question):
|
|
| 109 |
{"role": "system", "content": "You know things about League of Legends. You are a helpful AI that improves grammar and clarity. Do not give an explanation."},
|
| 110 |
{"role": "user", "content": "Please correct the following question with same structure: "+ user_question}
|
| 111 |
]
|
| 112 |
-
response = client.chat_completion(messages_q,
|
| 113 |
-
|
| 114 |
-
|
| 115 |
-
|
| 116 |
-
return response["choices"][0]["message"]["content"]
|
| 117 |
|
| 118 |
# chain = RunnablePassthrough() | RunnableLambda(ra) | prompt_template | client.chat_completion() | StrOutputParser() for notebook
|
| 119 |
|
|
|
|
| 109 |
{"role": "system", "content": "You know things about League of Legends. You are a helpful AI that improves grammar and clarity. Do not give an explanation."},
|
| 110 |
{"role": "user", "content": "Please correct the following question with same structure: "+ user_question}
|
| 111 |
]
|
| 112 |
+
#response = client.chat_completion(messages_q,
|
| 113 |
+
# max_tokens=20,
|
| 114 |
+
# stream=False,
|
| 115 |
+
# temperature=0.1)
|
| 116 |
+
return 0 #response["choices"][0]["message"]["content"]
|
| 117 |
|
| 118 |
# chain = RunnablePassthrough() | RunnableLambda(ra) | prompt_template | client.chat_completion() | StrOutputParser() for notebook
|
| 119 |
|