Spaces:
Paused
Paused
| import os | |
| import logging | |
| from openai import OpenAI | |
| if __name__ == '__main__': | |
| from llm.demo_prompt import demo_prompt | |
| client = OpenAI( | |
| # This is the default and can be omitted | |
| api_key=os.environ.get("OPENAI_API_KEY"), | |
| ) | |
| chat_completion = client.chat.completions.create( | |
| messages=[ | |
| { | |
| "role": "user", | |
| "content": demo_prompt, | |
| } | |
| ], | |
| model="gpt-3.5-turbo-0125", | |
| ) | |
| print(chat_completion.choices[0].message.content) | |
| def openai_call(apikey, | |
| model="gpt-3.5-turbo", | |
| user_content="如何做西红柿炖牛腩?", | |
| system_content=None): | |
| client = OpenAI( | |
| # This is the default and can be omitted | |
| api_key=apikey, | |
| ) | |
| if system_content is not None and len(system_content.strip()): | |
| messages = [ | |
| {'role': 'system', 'content': system_content}, | |
| {'role': 'user', 'content': user_content} | |
| ] | |
| else: | |
| messages = [ | |
| {'role': 'user', 'content': user_content} | |
| ] | |
| chat_completion = client.chat.completions.create( | |
| messages=messages, | |
| model=model, | |
| ) | |
| logging.info("Openai model inference done.") | |
| return chat_completion.choices[0].message.content |