Spaces:
Runtime error
Runtime error
Commit
·
d6deec7
1
Parent(s):
e0d127e
update code format
Browse files
app.py
CHANGED
|
@@ -15,6 +15,9 @@ vicuna_model = AutoModelForCausalLM.from_pretrained("lmsys/vicuna-7b-v1.3")
|
|
| 15 |
llama_tokenizer = AutoTokenizer.from_pretrained("daryl149/llama-2-7b-chat-hf")
|
| 16 |
llama_model = AutoModelForCausalLM.from_pretrained("daryl149/llama-2-7b-chat-hf")
|
| 17 |
|
|
|
|
|
|
|
|
|
|
| 18 |
def linguistic_features_fn(message):
|
| 19 |
# Load a trained spaCy pipeline
|
| 20 |
nlp = spacy.load("en_core_web_sm")
|
|
@@ -37,14 +40,11 @@ def linguistic_features_fn(message):
|
|
| 37 |
|
| 38 |
return formatted_output
|
| 39 |
|
| 40 |
-
os.environ['OPENAI_API_KEY']
|
| 41 |
-
openai.api_key = os.environ['OPENAI_API_KEY']
|
| 42 |
-
|
| 43 |
def chat(user_prompt, model = 'gpt-3.5-turbo', temperature = 0, verbose = False):
|
| 44 |
''' Normal call of OpenAI API '''
|
| 45 |
response = openai.ChatCompletion.create(
|
| 46 |
temperature = temperature,
|
| 47 |
-
model=model,
|
| 48 |
messages=[
|
| 49 |
{"role": "user", "content": user_prompt}
|
| 50 |
])
|
|
|
|
| 15 |
llama_tokenizer = AutoTokenizer.from_pretrained("daryl149/llama-2-7b-chat-hf")
|
| 16 |
llama_model = AutoModelForCausalLM.from_pretrained("daryl149/llama-2-7b-chat-hf")
|
| 17 |
|
| 18 |
+
os.environ['OPENAI_API_KEY']
|
| 19 |
+
openai.api_key = os.environ['OPENAI_API_KEY']
|
| 20 |
+
|
| 21 |
def linguistic_features_fn(message):
|
| 22 |
# Load a trained spaCy pipeline
|
| 23 |
nlp = spacy.load("en_core_web_sm")
|
|
|
|
| 40 |
|
| 41 |
return formatted_output
|
| 42 |
|
|
|
|
|
|
|
|
|
|
| 43 |
def chat(user_prompt, model = 'gpt-3.5-turbo', temperature = 0, verbose = False):
|
| 44 |
''' Normal call of OpenAI API '''
|
| 45 |
response = openai.ChatCompletion.create(
|
| 46 |
temperature = temperature,
|
| 47 |
+
model = model,
|
| 48 |
messages=[
|
| 49 |
{"role": "user", "content": user_prompt}
|
| 50 |
])
|