Commit ·
6b7de2a
1
Parent(s): df5c935
Fixing version conflicts
Browse files- Messaging_system/LLM.py +5 -6
- requirements.txt +0 -0
Messaging_system/LLM.py
CHANGED
|
@@ -6,7 +6,6 @@ import json
|
|
| 6 |
import time
|
| 7 |
from openai import OpenAI
|
| 8 |
import openai
|
| 9 |
-
import ollama
|
| 10 |
import torch
|
| 11 |
import re
|
| 12 |
import anthropic
|
|
@@ -31,8 +30,8 @@ class LLM:
|
|
| 31 |
def get_response(self, prompt, instructions):
|
| 32 |
if self.model_type == "openai":
|
| 33 |
response = self.get_message_openai(prompt, instructions)
|
| 34 |
-
elif self.model_type == "ollama":
|
| 35 |
-
|
| 36 |
elif self.model_type == "inference":
|
| 37 |
response = self.get_message_inference(prompt, instructions)
|
| 38 |
elif self.model_type == "claude":
|
|
@@ -59,9 +58,9 @@ class LLM:
|
|
| 59 |
elif self.Core.model in self.Core.config_file["google_models"]:
|
| 60 |
self.model_type = "google"
|
| 61 |
|
| 62 |
-
elif self.Core.model in self.Core.config_file["ollama_models"]:
|
| 63 |
-
|
| 64 |
-
|
| 65 |
|
| 66 |
elif self.Core.model in self.Core.config_file["claude_models"]:
|
| 67 |
self.model_type = "claude"
|
|
|
|
| 6 |
import time
|
| 7 |
from openai import OpenAI
|
| 8 |
import openai
|
|
|
|
| 9 |
import torch
|
| 10 |
import re
|
| 11 |
import anthropic
|
|
|
|
| 30 |
def get_response(self, prompt, instructions):
|
| 31 |
if self.model_type == "openai":
|
| 32 |
response = self.get_message_openai(prompt, instructions)
|
| 33 |
+
# elif self.model_type == "ollama":
|
| 34 |
+
# response = self.get_message_ollama(prompt, instructions)
|
| 35 |
elif self.model_type == "inference":
|
| 36 |
response = self.get_message_inference(prompt, instructions)
|
| 37 |
elif self.model_type == "claude":
|
|
|
|
| 58 |
elif self.Core.model in self.Core.config_file["google_models"]:
|
| 59 |
self.model_type = "google"
|
| 60 |
|
| 61 |
+
# elif self.Core.model in self.Core.config_file["ollama_models"]:
|
| 62 |
+
# self.model_type = "ollama"
|
| 63 |
+
# self.client = ollama.Client()
|
| 64 |
|
| 65 |
elif self.Core.model in self.Core.config_file["claude_models"]:
|
| 66 |
self.model_type = "claude"
|
requirements.txt
CHANGED
|
Binary files a/requirements.txt and b/requirements.txt differ
|
|
|