Spaces:
Sleeping
Sleeping
[NOTICKET] Update env key name
Browse files- .env.example +28 -28
- config/constant.py +9 -9
- externals/observability/langfuse.py +3 -3
- services/embed_model/embed_model.py +4 -4
- services/llms/LLM.py +15 -14
- services/uploader/azure_blob_service.py +2 -2
.env.example
CHANGED
|
@@ -8,15 +8,15 @@ causalogy--pg--host = ""
|
|
| 8 |
causalogy--pg--port = ""
|
| 9 |
|
| 10 |
# Local Postgresql
|
| 11 |
-
|
| 12 |
|
| 13 |
|
| 14 |
## -------------------- OBSERVABILITY -------------------- ##
|
| 15 |
|
| 16 |
# Langfuse (dev)
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
|
| 20 |
|
| 21 |
|
| 22 |
## -------------------- LLM CONFIGURATION -------------------- ##
|
|
@@ -25,22 +25,22 @@ langfuse--host = ""
|
|
| 25 |
GOOGLE_API_KEY= ""
|
| 26 |
|
| 27 |
# GPT-4O-MINI
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
|
| 33 |
# GPT-5-MINI
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
|
| 38 |
|
| 39 |
# GPT-4O-2
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
|
| 44 |
|
| 45 |
# Qdrant
|
| 46 |
ss--qdrant--api-key = ""
|
|
@@ -48,20 +48,20 @@ ss--qdrant--endpoint--url = ""
|
|
| 48 |
ss--qdrant--collection--name = ""
|
| 49 |
|
| 50 |
# Embedding Model
|
| 51 |
-
|
| 52 |
-
|
| 53 |
-
|
| 54 |
-
|
| 55 |
|
| 56 |
# AZURE BLOB
|
| 57 |
-
|
| 58 |
-
|
| 59 |
-
|
| 60 |
-
|
| 61 |
-
|
| 62 |
|
| 63 |
|
| 64 |
# JWT
|
| 65 |
-
|
| 66 |
-
|
| 67 |
-
|
|
|
|
| 8 |
causalogy--pg--port = ""
|
| 9 |
|
| 10 |
# Local Postgresql
|
| 11 |
+
ss__postgresql__url = ""
|
| 12 |
|
| 13 |
|
| 14 |
## -------------------- OBSERVABILITY -------------------- ##
|
| 15 |
|
| 16 |
# Langfuse (dev)
|
| 17 |
+
ss__langfuse__secret_key = ""
|
| 18 |
+
ss__langfuse__public_key = ""
|
| 19 |
+
langfuse__host = ""
|
| 20 |
|
| 21 |
|
| 22 |
## -------------------- LLM CONFIGURATION -------------------- ##
|
|
|
|
| 25 |
GOOGLE_API_KEY= ""
|
| 26 |
|
| 27 |
# GPT-4O-MINI
|
| 28 |
+
azureai__api_key__4omini = ""
|
| 29 |
+
azureai__endpoint__url__4omini = ""
|
| 30 |
+
azureai__deployment__name__4omini = ""
|
| 31 |
+
azureai__api__version__4omini = ""
|
| 32 |
|
| 33 |
# GPT-5-MINI
|
| 34 |
+
azureai__api_key__5mini = ""
|
| 35 |
+
azureai__endpoint__url__5mini = ""
|
| 36 |
+
azureai__deployment__name__5mini = ""
|
| 37 |
+
azureai__api__version__5mini = ""
|
| 38 |
|
| 39 |
# GPT-4O-2
|
| 40 |
+
azureai__api_key__4o_2 = ""
|
| 41 |
+
azureai__endpoint__url__4o_2 = ""
|
| 42 |
+
azureai__deployment__name__4o_2 = ""
|
| 43 |
+
azureai__api__version__4o_2 = ""
|
| 44 |
|
| 45 |
# Qdrant
|
| 46 |
ss--qdrant--api-key = ""
|
|
|
|
| 48 |
ss--qdrant--collection--name = ""
|
| 49 |
|
| 50 |
# Embedding Model
|
| 51 |
+
azureai__embedmodel__endpoint = ""
|
| 52 |
+
azureai__embedmodel__api_key = ""
|
| 53 |
+
azureai__embedmodel__api__version = ""
|
| 54 |
+
azureai__embedmodel__name = ""
|
| 55 |
|
| 56 |
# AZURE BLOB
|
| 57 |
+
azureai__search__sas = ""
|
| 58 |
+
azureai__container__endpoint = ""
|
| 59 |
+
azureai__container__name = ""
|
| 60 |
+
azureai__container__account__name = ""
|
| 61 |
+
azureai__max_file_size_mb = 5
|
| 62 |
|
| 63 |
|
| 64 |
# JWT
|
| 65 |
+
ss__jwt__secret_key = ""
|
| 66 |
+
ss__jwt__algorithm = ""
|
| 67 |
+
ss__jwt__access_token_expire_minutes = 60
|
config/constant.py
CHANGED
|
@@ -8,24 +8,24 @@ load_dotenv(EnvFilepath.ENVPATH)
|
|
| 8 |
|
| 9 |
@dataclass
|
| 10 |
class EnvPostgresConstants:
|
| 11 |
-
CONSTRING: str = os.environ.get("
|
| 12 |
|
| 13 |
@dataclass(frozen=True)
|
| 14 |
class SecurityConstants:
|
| 15 |
BCRYPT_MAX_BYTES: int = 72
|
| 16 |
-
JWT_SECRET_KEY: str = os.environ.get("
|
| 17 |
-
JWT_ALGORITHM: str = os.environ.get("
|
| 18 |
-
JWT_EXPIRE_MINUTES: int = int(os.environ.get("
|
| 19 |
|
| 20 |
@dataclass(frozen=True)
|
| 21 |
class AzureBlobConstants:
|
| 22 |
BLOB_PREFIX: str = "dev" # DOCS: change accordingly
|
| 23 |
-
ENDPOINT: str = os.environ.get("
|
| 24 |
-
CONTAINER_NAME: str = os.environ.get("
|
| 25 |
-
SAS_KEY: str = os.environ.get("
|
| 26 |
-
MAX_FILE_SIZE_MB: int = int(os.getenv("
|
| 27 |
CHUNK_SIZE: int = 4 * 1024 * 1024
|
| 28 |
-
ACCOUNT_NAME: str = os.environ.get("
|
| 29 |
|
| 30 |
@dataclass
|
| 31 |
class ProfileFieldTypes:
|
|
|
|
| 8 |
|
| 9 |
@dataclass
|
| 10 |
class EnvPostgresConstants:
|
| 11 |
+
CONSTRING: str = os.environ.get("ss__postgresql__url")
|
| 12 |
|
| 13 |
@dataclass(frozen=True)
|
| 14 |
class SecurityConstants:
|
| 15 |
BCRYPT_MAX_BYTES: int = 72
|
| 16 |
+
JWT_SECRET_KEY: str = os.environ.get("ss__jwt__secret_key")
|
| 17 |
+
JWT_ALGORITHM: str = os.environ.get("ss__jwt__algorithm")
|
| 18 |
+
JWT_EXPIRE_MINUTES: int = int(os.environ.get("ss__jwt__access_token_expire_minutes", 60))
|
| 19 |
|
| 20 |
@dataclass(frozen=True)
|
| 21 |
class AzureBlobConstants:
|
| 22 |
BLOB_PREFIX: str = "dev" # DOCS: change accordingly
|
| 23 |
+
ENDPOINT: str = os.environ.get("azureai__container__endpoint")
|
| 24 |
+
CONTAINER_NAME: str = os.environ.get("azureai__container__name")
|
| 25 |
+
SAS_KEY: str = os.environ.get("azureai__search__sas")
|
| 26 |
+
MAX_FILE_SIZE_MB: int = int(os.getenv("azureai__max_file_size_mb", "5"))
|
| 27 |
CHUNK_SIZE: int = 4 * 1024 * 1024
|
| 28 |
+
ACCOUNT_NAME: str = os.environ.get("azureai__container__account__name")
|
| 29 |
|
| 30 |
@dataclass
|
| 31 |
class ProfileFieldTypes:
|
externals/observability/langfuse.py
CHANGED
|
@@ -7,9 +7,9 @@ from langchain_core.callbacks.base import BaseCallbackHandler
|
|
| 7 |
|
| 8 |
|
| 9 |
langfuse = Langfuse(
|
| 10 |
-
secret_key=os.environ.get('
|
| 11 |
-
public_key=os.environ.get('
|
| 12 |
-
host=os.environ.get('
|
| 13 |
)
|
| 14 |
|
| 15 |
|
|
|
|
| 7 |
|
| 8 |
|
| 9 |
langfuse = Langfuse(
|
| 10 |
+
secret_key=os.environ.get('ss__langfuse__secret_key'),
|
| 11 |
+
public_key=os.environ.get('ss__langfuse__public_key'),
|
| 12 |
+
host=os.environ.get('langfuse__host'),
|
| 13 |
)
|
| 14 |
|
| 15 |
|
services/embed_model/embed_model.py
CHANGED
|
@@ -8,10 +8,10 @@ from langchain_openai import AzureOpenAIEmbeddings
|
|
| 8 |
def serve_embed_model():
|
| 9 |
try:
|
| 10 |
embed_model = AzureOpenAIEmbeddings(
|
| 11 |
-
model=os.environ.get('
|
| 12 |
-
azure_endpoint=os.environ.get('
|
| 13 |
-
api_key=os.environ.get('
|
| 14 |
-
api_version=os.environ.get('
|
| 15 |
)
|
| 16 |
# asyncio.run(embed_model.aembed_documents("Hello my name is EMA"))
|
| 17 |
print(f"✅ [INFO] Initialized embedding model succeeded")
|
|
|
|
| 8 |
def serve_embed_model():
|
| 9 |
try:
|
| 10 |
embed_model = AzureOpenAIEmbeddings(
|
| 11 |
+
model=os.environ.get('azureai__embedmodel__name'),
|
| 12 |
+
azure_endpoint=os.environ.get('azureai__embedmodel__endpoint'),
|
| 13 |
+
api_key=os.environ.get('azureai__embedmodel__api_key'),
|
| 14 |
+
api_version=os.environ.get('azureai__embedmodel__api__version')
|
| 15 |
)
|
| 16 |
# asyncio.run(embed_model.aembed_documents("Hello my name is EMA"))
|
| 17 |
print(f"✅ [INFO] Initialized embedding model succeeded")
|
services/llms/LLM.py
CHANGED
|
@@ -1,6 +1,7 @@
|
|
| 1 |
import os
|
| 2 |
from dotenv import load_dotenv
|
| 3 |
-
|
|
|
|
| 4 |
from langchain_openai import AzureChatOpenAI
|
| 5 |
# from langchain_google_genai import ChatGoogleGenerativeAI
|
| 6 |
# from langchain.callbacks import AsyncIteratorCallbackHandler
|
|
@@ -10,7 +11,7 @@ from langchain_openai import AzureChatOpenAI
|
|
| 10 |
# langfuse_handler = CallbackHandler(
|
| 11 |
# secret_key=os.environ.get('buma--langfuse--secret-key'),
|
| 12 |
# public_key=os.environ.get('buma--langfuse--public-key'),
|
| 13 |
-
# host=os.environ.get('buma--
|
| 14 |
# )
|
| 15 |
|
| 16 |
|
|
@@ -29,30 +30,30 @@ from langchain_openai import AzureChatOpenAI
|
|
| 29 |
# )
|
| 30 |
|
| 31 |
model_4o_2 = AzureChatOpenAI(
|
| 32 |
-
azure_endpoint=os.environ.get("
|
| 33 |
-
openai_api_version=os.environ.get("
|
| 34 |
-
deployment_name=os.environ.get("
|
| 35 |
-
openai_api_key=os.environ.get("
|
| 36 |
openai_api_type="azure",
|
| 37 |
max_retries=2,
|
| 38 |
disable_streaming=True
|
| 39 |
)
|
| 40 |
|
| 41 |
model_4omini = AzureChatOpenAI(
|
| 42 |
-
azure_endpoint=os.environ.get("
|
| 43 |
-
openai_api_version=os.environ.get("
|
| 44 |
-
deployment_name=os.environ.get("
|
| 45 |
-
openai_api_key=os.environ.get("
|
| 46 |
openai_api_type="azure",
|
| 47 |
max_retries=2,
|
| 48 |
disable_streaming=True
|
| 49 |
)
|
| 50 |
|
| 51 |
model_5mini = AzureChatOpenAI(
|
| 52 |
-
azure_endpoint=os.environ.get("
|
| 53 |
-
openai_api_version=os.environ.get("
|
| 54 |
-
deployment_name=os.environ.get("
|
| 55 |
-
openai_api_key=os.environ.get("
|
| 56 |
openai_api_type="azure",
|
| 57 |
max_retries=2,
|
| 58 |
disable_streaming=True
|
|
|
|
| 1 |
import os
|
| 2 |
from dotenv import load_dotenv
|
| 3 |
+
from config.env_constant import EnvFilepath
|
| 4 |
+
load_dotenv(EnvFilepath.ENVPATH)
|
| 5 |
from langchain_openai import AzureChatOpenAI
|
| 6 |
# from langchain_google_genai import ChatGoogleGenerativeAI
|
| 7 |
# from langchain.callbacks import AsyncIteratorCallbackHandler
|
|
|
|
| 11 |
# langfuse_handler = CallbackHandler(
|
| 12 |
# secret_key=os.environ.get('buma--langfuse--secret-key'),
|
| 13 |
# public_key=os.environ.get('buma--langfuse--public-key'),
|
| 14 |
+
# host=os.environ.get('buma--langfuse__host')
|
| 15 |
# )
|
| 16 |
|
| 17 |
|
|
|
|
| 30 |
# )
|
| 31 |
|
| 32 |
model_4o_2 = AzureChatOpenAI(
|
| 33 |
+
azure_endpoint=os.environ.get("azureai__endpoint__url__4o_2"), # or your deployment
|
| 34 |
+
openai_api_version=os.environ.get("azureai__api__version__4o_2"), # or your api version
|
| 35 |
+
deployment_name=os.environ.get("azureai__deployment__name__4o_2"),
|
| 36 |
+
openai_api_key=os.environ.get("azureai__api_key__4o_2"),
|
| 37 |
openai_api_type="azure",
|
| 38 |
max_retries=2,
|
| 39 |
disable_streaming=True
|
| 40 |
)
|
| 41 |
|
| 42 |
model_4omini = AzureChatOpenAI(
|
| 43 |
+
azure_endpoint=os.environ.get("azureai__endpoint__url__4omini"), # or your deployment
|
| 44 |
+
openai_api_version=os.environ.get("azureai__api__version__4omini"), # or your api version
|
| 45 |
+
deployment_name=os.environ.get("azureai__deployment__name__4omini"),
|
| 46 |
+
openai_api_key=os.environ.get("azureai__api_key__4omini"),
|
| 47 |
openai_api_type="azure",
|
| 48 |
max_retries=2,
|
| 49 |
disable_streaming=True
|
| 50 |
)
|
| 51 |
|
| 52 |
model_5mini = AzureChatOpenAI(
|
| 53 |
+
azure_endpoint=os.environ.get("azureai__endpoint__url__5mini"), # or your deployment
|
| 54 |
+
openai_api_version=os.environ.get("azureai__api__version__5mini"), # or your api version
|
| 55 |
+
deployment_name=os.environ.get("azureai__deployment__name__5mini"),
|
| 56 |
+
openai_api_key=os.environ.get("azureai__api_key__5mini"),
|
| 57 |
openai_api_type="azure",
|
| 58 |
max_retries=2,
|
| 59 |
disable_streaming=True
|
services/uploader/azure_blob_service.py
CHANGED
|
@@ -20,7 +20,7 @@ logger = get_logger("azure blob")
|
|
| 20 |
# CONNECTION_STRING = os.getenv("AZURE_STORAGE_CONNECTION_STRING")
|
| 21 |
# CONTAINER_NAME = "pdf-uploads" # The name of your Azure Blob Storage container
|
| 22 |
|
| 23 |
-
async def get_blob_service_client(url=os.environ.get('
|
| 24 |
try:
|
| 25 |
default_credential = DefaultAzureCredential()
|
| 26 |
blob_service_client = BlobServiceClient(url, credential=default_credential)
|
|
@@ -29,7 +29,7 @@ async def get_blob_service_client(url=os.environ.get('azureai--container--endpoi
|
|
| 29 |
logger.error(f'❌ Getting blob service client error, {E}')
|
| 30 |
|
| 31 |
|
| 32 |
-
async def get_container_client(url=os.environ.get('
|
| 33 |
try:
|
| 34 |
blob_service_client = await get_blob_service_client(url=url)
|
| 35 |
blob_client = blob_service_client.get_container_client(container_name)
|
|
|
|
| 20 |
# CONNECTION_STRING = os.getenv("AZURE_STORAGE_CONNECTION_STRING")
|
| 21 |
# CONTAINER_NAME = "pdf-uploads" # The name of your Azure Blob Storage container
|
| 22 |
|
| 23 |
+
async def get_blob_service_client(url=os.environ.get('azureai__container__endpoint')) -> BlobServiceClient:
|
| 24 |
try:
|
| 25 |
default_credential = DefaultAzureCredential()
|
| 26 |
blob_service_client = BlobServiceClient(url, credential=default_credential)
|
|
|
|
| 29 |
logger.error(f'❌ Getting blob service client error, {E}')
|
| 30 |
|
| 31 |
|
| 32 |
+
async def get_container_client(url=os.environ.get('azureai__container__endpoint'), container_name=os.environ.get("azureai__container__name")) -> ContainerClient:
|
| 33 |
try:
|
| 34 |
blob_service_client = await get_blob_service_client(url=url)
|
| 35 |
blob_client = blob_service_client.get_container_client(container_name)
|