ishaq101 commited on
Commit
85d8eae
·
1 Parent(s): ea3ee44

[NOTICKET] Update env key name

Browse files
.env.example CHANGED
@@ -8,15 +8,15 @@ causalogy--pg--host = ""
8
  causalogy--pg--port = ""
9
 
10
  # Local Postgresql
11
- ss--postgresql--url = ""
12
 
13
 
14
  ## -------------------- OBSERVABILITY -------------------- ##
15
 
16
  # Langfuse (dev)
17
- ss--langfuse--secret-key = ""
18
- ss--langfuse--public-key = ""
19
- langfuse--host = ""
20
 
21
 
22
  ## -------------------- LLM CONFIGURATION -------------------- ##
@@ -25,22 +25,22 @@ langfuse--host = ""
25
  GOOGLE_API_KEY= ""
26
 
27
  # GPT-4O-MINI
28
- azureai--api-key--4omini = ""
29
- azureai--endpoint--url--4omini = ""
30
- azureai--deployment--name--4omini = ""
31
- azureai--api--version--4omini = ""
32
 
33
  # GPT-5-MINI
34
- azureai--api-key--5mini = ""
35
- azureai--endpoint--url--5mini = ""
36
- azureai--deployment--name--5mini = ""
37
- azureai--api--version--5mini = ""
38
 
39
  # GPT-4O-2
40
- azureai--api-key--4o-2 = ""
41
- azureai--endpoint--url--4o-2 = ""
42
- azureai--deployment--name--4o-2 = ""
43
- azureai--api--version--4o-2 = ""
44
 
45
  # Qdrant
46
  ss--qdrant--api-key = ""
@@ -48,20 +48,20 @@ ss--qdrant--endpoint--url = ""
48
  ss--qdrant--collection--name = ""
49
 
50
  # Embedding Model
51
- azureai--embedmodel--endpoint = ""
52
- azureai--embedmodel--api-key = ""
53
- azureai--embedmodel--api--version = ""
54
- azureai--embedmodel--name = ""
55
 
56
  # AZURE BLOB
57
- azureai--search--sas = ""
58
- azureai--container--endpoint = ""
59
- azureai--container--name = ""
60
- azureai--container--account--name = ""
61
- azureai--max-file-size-mb = 5
62
 
63
 
64
  # JWT
65
- ss--jwt--secret-key = ""
66
- ss--jwt--algorithm = ""
67
- ss--jwt--access-token-expire-minutes = 60
 
8
  causalogy--pg--port = ""
9
 
10
  # Local Postgresql
11
+ ss__postgresql__url = ""
12
 
13
 
14
  ## -------------------- OBSERVABILITY -------------------- ##
15
 
16
  # Langfuse (dev)
17
+ ss__langfuse__secret_key = ""
18
+ ss__langfuse__public_key = ""
19
+ langfuse__host = ""
20
 
21
 
22
  ## -------------------- LLM CONFIGURATION -------------------- ##
 
25
  GOOGLE_API_KEY= ""
26
 
27
  # GPT-4O-MINI
28
+ azureai__api_key__4omini = ""
29
+ azureai__endpoint__url__4omini = ""
30
+ azureai__deployment__name__4omini = ""
31
+ azureai__api__version__4omini = ""
32
 
33
  # GPT-5-MINI
34
+ azureai__api_key__5mini = ""
35
+ azureai__endpoint__url__5mini = ""
36
+ azureai__deployment__name__5mini = ""
37
+ azureai__api__version__5mini = ""
38
 
39
  # GPT-4O-2
40
+ azureai__api_key__4o_2 = ""
41
+ azureai__endpoint__url__4o_2 = ""
42
+ azureai__deployment__name__4o_2 = ""
43
+ azureai__api__version__4o_2 = ""
44
 
45
  # Qdrant
46
  ss--qdrant--api-key = ""
 
48
  ss--qdrant--collection--name = ""
49
 
50
  # Embedding Model
51
+ azureai__embedmodel__endpoint = ""
52
+ azureai__embedmodel__api_key = ""
53
+ azureai__embedmodel__api__version = ""
54
+ azureai__embedmodel__name = ""
55
 
56
  # AZURE BLOB
57
+ azureai__search__sas = ""
58
+ azureai__container__endpoint = ""
59
+ azureai__container__name = ""
60
+ azureai__container__account__name = ""
61
+ azureai__max_file_size_mb = 5
62
 
63
 
64
  # JWT
65
+ ss__jwt__secret_key = ""
66
+ ss__jwt__algorithm = ""
67
+ ss__jwt__access_token_expire_minutes = 60
config/constant.py CHANGED
@@ -8,24 +8,24 @@ load_dotenv(EnvFilepath.ENVPATH)
8
 
9
  @dataclass
10
  class EnvPostgresConstants:
11
- CONSTRING: str = os.environ.get("ss--postgresql--url")
12
 
13
  @dataclass(frozen=True)
14
  class SecurityConstants:
15
  BCRYPT_MAX_BYTES: int = 72
16
- JWT_SECRET_KEY: str = os.environ.get("ss--jwt--secret-key")
17
- JWT_ALGORITHM: str = os.environ.get("ss--jwt--algorithm")
18
- JWT_EXPIRE_MINUTES: int = int(os.environ.get("ss--jwt--access-token-expire-minutes", 60))
19
 
20
  @dataclass(frozen=True)
21
  class AzureBlobConstants:
22
  BLOB_PREFIX: str = "dev" # DOCS: change accordingly
23
- ENDPOINT: str = os.environ.get("azureai--container--endpoint")
24
- CONTAINER_NAME: str = os.environ.get("azureai--container--name")
25
- SAS_KEY: str = os.environ.get("azureai--search--sas")
26
- MAX_FILE_SIZE_MB: int = int(os.getenv("azureai--max-file-size-mb", "5"))
27
  CHUNK_SIZE: int = 4 * 1024 * 1024
28
- ACCOUNT_NAME: str = os.environ.get("azureai--container--account--name")
29
 
30
  @dataclass
31
  class ProfileFieldTypes:
 
8
 
9
  @dataclass
10
  class EnvPostgresConstants:
11
+ CONSTRING: str = os.environ.get("ss__postgresql__url")
12
 
13
  @dataclass(frozen=True)
14
  class SecurityConstants:
15
  BCRYPT_MAX_BYTES: int = 72
16
+ JWT_SECRET_KEY: str = os.environ.get("ss__jwt__secret_key")
17
+ JWT_ALGORITHM: str = os.environ.get("ss__jwt__algorithm")
18
+ JWT_EXPIRE_MINUTES: int = int(os.environ.get("ss__jwt__access_token_expire_minutes", 60))
19
 
20
  @dataclass(frozen=True)
21
  class AzureBlobConstants:
22
  BLOB_PREFIX: str = "dev" # DOCS: change accordingly
23
+ ENDPOINT: str = os.environ.get("azureai__container__endpoint")
24
+ CONTAINER_NAME: str = os.environ.get("azureai__container__name")
25
+ SAS_KEY: str = os.environ.get("azureai__search__sas")
26
+ MAX_FILE_SIZE_MB: int = int(os.getenv("azureai__max_file_size_mb", "5"))
27
  CHUNK_SIZE: int = 4 * 1024 * 1024
28
+ ACCOUNT_NAME: str = os.environ.get("azureai__container__account__name")
29
 
30
  @dataclass
31
  class ProfileFieldTypes:
externals/observability/langfuse.py CHANGED
@@ -7,9 +7,9 @@ from langchain_core.callbacks.base import BaseCallbackHandler
7
 
8
 
9
  langfuse = Langfuse(
10
- secret_key=os.environ.get('ss--langfuse--secret-key'),
11
- public_key=os.environ.get('ss--langfuse--public-key'),
12
- host=os.environ.get('langfuse--host'),
13
  )
14
 
15
 
 
7
 
8
 
9
  langfuse = Langfuse(
10
+ secret_key=os.environ.get('ss__langfuse__secret_key'),
11
+ public_key=os.environ.get('ss__langfuse__public_key'),
12
+ host=os.environ.get('langfuse__host'),
13
  )
14
 
15
 
services/embed_model/embed_model.py CHANGED
@@ -8,10 +8,10 @@ from langchain_openai import AzureOpenAIEmbeddings
8
  def serve_embed_model():
9
  try:
10
  embed_model = AzureOpenAIEmbeddings(
11
- model=os.environ.get('azureai--embedmodel--name'),
12
- azure_endpoint=os.environ.get('azureai--embedmodel--endpoint'),
13
- api_key=os.environ.get('azureai--embedmodel--api-key'),
14
- api_version=os.environ.get('azureai--embedmodel--api--version')
15
  )
16
  # asyncio.run(embed_model.aembed_documents("Hello my name is EMA"))
17
  print(f"✅ [INFO] Initialized embedding model succeeded")
 
8
  def serve_embed_model():
9
  try:
10
  embed_model = AzureOpenAIEmbeddings(
11
+ model=os.environ.get('azureai__embedmodel__name'),
12
+ azure_endpoint=os.environ.get('azureai__embedmodel__endpoint'),
13
+ api_key=os.environ.get('azureai__embedmodel__api_key'),
14
+ api_version=os.environ.get('azureai__embedmodel__api__version')
15
  )
16
  # asyncio.run(embed_model.aembed_documents("Hello my name is EMA"))
17
  print(f"✅ [INFO] Initialized embedding model succeeded")
services/llms/LLM.py CHANGED
@@ -1,6 +1,7 @@
1
  import os
2
  from dotenv import load_dotenv
3
- load_dotenv()
 
4
  from langchain_openai import AzureChatOpenAI
5
  # from langchain_google_genai import ChatGoogleGenerativeAI
6
  # from langchain.callbacks import AsyncIteratorCallbackHandler
@@ -10,7 +11,7 @@ from langchain_openai import AzureChatOpenAI
10
  # langfuse_handler = CallbackHandler(
11
  # secret_key=os.environ.get('buma--langfuse--secret-key'),
12
  # public_key=os.environ.get('buma--langfuse--public-key'),
13
- # host=os.environ.get('buma--langfuse--host')
14
  # )
15
 
16
 
@@ -29,30 +30,30 @@ from langchain_openai import AzureChatOpenAI
29
  # )
30
 
31
  model_4o_2 = AzureChatOpenAI(
32
- azure_endpoint=os.environ.get("azureai--endpoint--url--4o-2"), # or your deployment
33
- openai_api_version=os.environ.get("azureai--api--version--4o-2"), # or your api version
34
- deployment_name=os.environ.get("azureai--deployment--name--4o-2"),
35
- openai_api_key=os.environ.get("azureai--api-key--4o-2"),
36
  openai_api_type="azure",
37
  max_retries=2,
38
  disable_streaming=True
39
  )
40
 
41
  model_4omini = AzureChatOpenAI(
42
- azure_endpoint=os.environ.get("azureai--endpoint--url--4omini"), # or your deployment
43
- openai_api_version=os.environ.get("azureai--api--version--4omini"), # or your api version
44
- deployment_name=os.environ.get("azureai--deployment--name--4omini"),
45
- openai_api_key=os.environ.get("azureai--api-key--4omini"),
46
  openai_api_type="azure",
47
  max_retries=2,
48
  disable_streaming=True
49
  )
50
 
51
  model_5mini = AzureChatOpenAI(
52
- azure_endpoint=os.environ.get("azureai--endpoint--url--5mini"), # or your deployment
53
- openai_api_version=os.environ.get("azureai--api--version--5mini"), # or your api version
54
- deployment_name=os.environ.get("azureai--deployment--name--5mini"),
55
- openai_api_key=os.environ.get("azureai--api-key--5mini"),
56
  openai_api_type="azure",
57
  max_retries=2,
58
  disable_streaming=True
 
1
  import os
2
  from dotenv import load_dotenv
3
+ from config.env_constant import EnvFilepath
4
+ load_dotenv(EnvFilepath.ENVPATH)
5
  from langchain_openai import AzureChatOpenAI
6
  # from langchain_google_genai import ChatGoogleGenerativeAI
7
  # from langchain.callbacks import AsyncIteratorCallbackHandler
 
11
  # langfuse_handler = CallbackHandler(
12
  # secret_key=os.environ.get('buma--langfuse--secret-key'),
13
  # public_key=os.environ.get('buma--langfuse--public-key'),
14
+ # host=os.environ.get('buma--langfuse__host')
15
  # )
16
 
17
 
 
30
  # )
31
 
32
  model_4o_2 = AzureChatOpenAI(
33
+ azure_endpoint=os.environ.get("azureai__endpoint__url__4o_2"), # or your deployment
34
+ openai_api_version=os.environ.get("azureai__api__version__4o_2"), # or your api version
35
+ deployment_name=os.environ.get("azureai__deployment__name__4o_2"),
36
+ openai_api_key=os.environ.get("azureai__api_key__4o_2"),
37
  openai_api_type="azure",
38
  max_retries=2,
39
  disable_streaming=True
40
  )
41
 
42
  model_4omini = AzureChatOpenAI(
43
+ azure_endpoint=os.environ.get("azureai__endpoint__url__4omini"), # or your deployment
44
+ openai_api_version=os.environ.get("azureai__api__version__4omini"), # or your api version
45
+ deployment_name=os.environ.get("azureai__deployment__name__4omini"),
46
+ openai_api_key=os.environ.get("azureai__api_key__4omini"),
47
  openai_api_type="azure",
48
  max_retries=2,
49
  disable_streaming=True
50
  )
51
 
52
  model_5mini = AzureChatOpenAI(
53
+ azure_endpoint=os.environ.get("azureai__endpoint__url__5mini"), # or your deployment
54
+ openai_api_version=os.environ.get("azureai__api__version__5mini"), # or your api version
55
+ deployment_name=os.environ.get("azureai__deployment__name__5mini"),
56
+ openai_api_key=os.environ.get("azureai__api_key__5mini"),
57
  openai_api_type="azure",
58
  max_retries=2,
59
  disable_streaming=True
services/uploader/azure_blob_service.py CHANGED
@@ -20,7 +20,7 @@ logger = get_logger("azure blob")
20
  # CONNECTION_STRING = os.getenv("AZURE_STORAGE_CONNECTION_STRING")
21
  # CONTAINER_NAME = "pdf-uploads" # The name of your Azure Blob Storage container
22
 
23
- async def get_blob_service_client(url=os.environ.get('azureai--container--endpoint')) -> BlobServiceClient:
24
  try:
25
  default_credential = DefaultAzureCredential()
26
  blob_service_client = BlobServiceClient(url, credential=default_credential)
@@ -29,7 +29,7 @@ async def get_blob_service_client(url=os.environ.get('azureai--container--endpoi
29
  logger.error(f'❌ Getting blob service client error, {E}')
30
 
31
 
32
- async def get_container_client(url=os.environ.get('azureai--container--endpoint'), container_name=os.environ.get("azureai--container--name")) -> ContainerClient:
33
  try:
34
  blob_service_client = await get_blob_service_client(url=url)
35
  blob_client = blob_service_client.get_container_client(container_name)
 
20
  # CONNECTION_STRING = os.getenv("AZURE_STORAGE_CONNECTION_STRING")
21
  # CONTAINER_NAME = "pdf-uploads" # The name of your Azure Blob Storage container
22
 
23
+ async def get_blob_service_client(url=os.environ.get('azureai__container__endpoint')) -> BlobServiceClient:
24
  try:
25
  default_credential = DefaultAzureCredential()
26
  blob_service_client = BlobServiceClient(url, credential=default_credential)
 
29
  logger.error(f'❌ Getting blob service client error, {E}')
30
 
31
 
32
+ async def get_container_client(url=os.environ.get('azureai__container__endpoint'), container_name=os.environ.get("azureai__container__name")) -> ContainerClient:
33
  try:
34
  blob_service_client = await get_blob_service_client(url=url)
35
  blob_client = blob_service_client.get_container_client(container_name)