Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -48,7 +48,7 @@ def generate_feedback_llm(user_profile):
|
|
| 48 |
feedback = []
|
| 49 |
for i, (question, answer) in enumerate(zip(user_profile.get("questions", []), user_profile.get("user_answers", []))):
|
| 50 |
messages = [
|
| 51 |
-
{"role": "system", "content": f"You are a professional interviewer providing feedback for a candidate's response in a {user_profile['interview_type']} interview for a {user_profile['field']} role."},
|
| 52 |
{"role": "user", "content": f"Question: {question}\nAnswer: {answer}\nPlease give specific, constructive feedback."}
|
| 53 |
]
|
| 54 |
response = client.chat_completion(messages, max_tokens=150, stream=False)
|
|
|
|
| 48 |
feedback = []
|
| 49 |
for i, (question, answer) in enumerate(zip(user_profile.get("questions", []), user_profile.get("user_answers", []))):
|
| 50 |
messages = [
|
| 51 |
+
{"role": "system", "content": f"You are a professional interviewer providing feedback for a candidate's response in a {user_profile['interview_type']} interview for a {user_profile['field']} role. DO NOT include any candidate responses or dialogue. DO NOT include any 'Interviewer:', 'Candidate:' prefixes"},
|
| 52 |
{"role": "user", "content": f"Question: {question}\nAnswer: {answer}\nPlease give specific, constructive feedback."}
|
| 53 |
]
|
| 54 |
response = client.chat_completion(messages, max_tokens=150, stream=False)
|