sliitguy
updated for deployment
782bbd9
from langchain_groq import ChatGroq
import os
from dotenv import load_dotenv
class GroqLLM:
def __init__(self):
load_dotenv()
def get_llm(self):
try:
# Don't print API key - security risk
# print(os.getenv("GROQ_API_KEY"))
os.environ["GROQ_API_KEY"] = self.groq_api_key = os.getenv("GROQ_API_KEY")
# Using llama models which have better JSON support
llm=ChatGroq(
api_key=self.groq_api_key,
model="llama-3.3-70b-versatile", # Better JSON support
streaming=False,
temperature=0.1 # Lower temperature for more consistent structured output
)
return llm
except Exception as e:
raise ValueError(f"Error occurred with exception: {e}")
def get_moon(self):
try:
# Don't print API key - security risk
# print(os.getenv("GROQ_API_KEY"))
os.environ["GROQ_API_KEY"] = self.groq_api_key = os.getenv("GROQ_API_KEY")
llm=ChatGroq(
api_key=self.groq_api_key,
model="moonshotai/kimi-k2-instruct",
streaming=False
)
return llm
except Exception as e:
raise ValueError(f"Error occurred with exception: {e}")