Yaswanth-Bolla commited on
Commit
7f52668
·
1 Parent(s): bbbb404

update base url

Browse files
Files changed (2) hide show
  1. README.md +1 -1
  2. inference.py +5 -5
README.md CHANGED
@@ -151,7 +151,7 @@ curl -X POST http://localhost:8000/step -H "Content-Type: application/json" -d '
151
  ```bash
152
  export OPENAI_API_KEY=sk-...
153
  export MODEL_NAME=gpt-4o-mini
154
- export API_BASE_URL=http://localhost:8000
155
 
156
  python inference.py
157
  ```
 
151
  ```bash
152
  export OPENAI_API_KEY=sk-...
153
  export MODEL_NAME=gpt-4o-mini
154
+ export ENV_BASE_URL=http://localhost:8000
155
 
156
  python inference.py
157
  ```
inference.py CHANGED
@@ -26,7 +26,7 @@ from openai import OpenAI
26
  # Config
27
  # ------------------------------------------------------------------
28
 
29
- API_BASE_URL = os.environ.get("API_BASE_URL", "http://localhost:8000")
30
  MODEL_NAME = os.environ.get("MODEL_NAME", "gpt-4o-mini")
31
  HF_TOKEN = os.environ.get("HF_TOKEN", "")
32
  OPENAI_API_KEY = os.environ.get("OPENAI_API_KEY", "")
@@ -112,10 +112,10 @@ class EnvClient:
112
  def create_openai_client() -> OpenAI:
113
  """Create OpenAI client with appropriate config."""
114
  api_key = OPENAI_API_KEY or HF_TOKEN or "no-key"
115
- base_url = None
116
 
117
  # If using HF inference endpoint, set base_url
118
- if HF_TOKEN and not OPENAI_API_KEY:
119
  base_url = f"https://api-inference.huggingface.co/models/{MODEL_NAME}/v1"
120
 
121
  return OpenAI(api_key=api_key, base_url=base_url)
@@ -269,10 +269,10 @@ def main():
269
  print("=" * 60)
270
  print("SRE Incident Response — OpenEnv Inference")
271
  print(f"Model: {MODEL_NAME}")
272
- print(f"Environment: {API_BASE_URL}")
273
  print("=" * 60)
274
 
275
- env = EnvClient(API_BASE_URL)
276
  llm = create_openai_client()
277
 
278
  results = []
 
26
  # Config
27
  # ------------------------------------------------------------------
28
 
29
+ ENV_BASE_URL = os.environ.get("ENV_BASE_URL", "http://localhost:8000")
30
  MODEL_NAME = os.environ.get("MODEL_NAME", "gpt-4o-mini")
31
  HF_TOKEN = os.environ.get("HF_TOKEN", "")
32
  OPENAI_API_KEY = os.environ.get("OPENAI_API_KEY", "")
 
112
  def create_openai_client() -> OpenAI:
113
  """Create OpenAI client with appropriate config."""
114
  api_key = OPENAI_API_KEY or HF_TOKEN or "no-key"
115
+ base_url = os.environ.get("API_BASE_URL")
116
 
117
  # If using HF inference endpoint, set base_url
118
+ if HF_TOKEN and not OPENAI_API_KEY and not base_url:
119
  base_url = f"https://api-inference.huggingface.co/models/{MODEL_NAME}/v1"
120
 
121
  return OpenAI(api_key=api_key, base_url=base_url)
 
269
  print("=" * 60)
270
  print("SRE Incident Response — OpenEnv Inference")
271
  print(f"Model: {MODEL_NAME}")
272
+ print(f"Environment: {ENV_BASE_URL}")
273
  print("=" * 60)
274
 
275
+ env = EnvClient(ENV_BASE_URL)
276
  llm = create_openai_client()
277
 
278
  results = []