Spaces:
Sleeping
Sleeping
File size: 1,466 Bytes
6c655a3 be3a5c4 3f2f8aa 6874dac 6c655a3 be3a5c4 6f57d05 5c271a3 be3a5c4 5c271a3 be3a5c4 6c655a3 11bd168 fbc17f4 3002e1b 8ce97f0 3002e1b a6a0614 8827263 0b2c9fd 6f57d05 6874dac 5c271a3 eb40d68 a6a0614 8ce97f0 a6a0614 5c271a3 eb40d68 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 |
import os
import numpy as np
from langchain_groq import ChatGroq
from langchain_openai import ChatOpenAI
from langchain_google_genai import ChatGoogleGenerativeAI
from langchain_anthropic import ChatAnthropic
from langchain_openai import OpenAIEmbeddings
from huggingface_hub import login
from dotenv import load_dotenv
import tiktoken
load_dotenv()
os.environ['HUGGINGFACEHUB_ACCESS_TOKEN']=os.getenv('HUGGINGFACEHUB_ACCESS_TOKEN')
login(os.environ['HUGGINGFACEHUB_ACCESS_TOKEN'])
embedding_model = OpenAIEmbeddings(model="text-embedding-3-small", dimensions=1536)
llm_anthropic = ChatAnthropic(model='claude-3-7-sonnet-latest', temperature=1)
llm_gemini = ChatGoogleGenerativeAI(model="gemini-1.5-flash")
llm_groq_openai = ChatGroq(model="openai/gpt-oss-120b",temperature=0.7)
llm_groq = ChatGroq(model="llama-3.3-70b-versatile",temperature=0)
llm_gpt_small = ChatOpenAI(model="gpt-3.5-turbo",temperature=0.3)
llm_gpt = ChatOpenAI(model="gpt-3.5-turbo",temperature=0.3)
llm_gpt_high = ChatOpenAI(model="gpt-5-nano",temperature=0.5)
encoding_model = tiktoken.encoding_for_model('gpt-4o-mini')
# encoding_model = 'encoding_model'
captioning_model = "meta-llama/llama-4-scout-17b-16e-instruct"
image_generation_model = "black-forest-labs/FLUX.1-schnell"
ideator_llm = llm_gpt_high
moderator_llm = llm_gpt
critic_llm = llm_gpt
simplifier_llm = llm_gpt
normalizer_llm = llm_gpt
validator_llm = llm_gpt_small
judge1_llm = llm_gpt
judge2_llm = llm_gpt
|