Commit ·
7f52668
1
Parent(s): bbbb404
update base url
Browse files- README.md +1 -1
- inference.py +5 -5
README.md
CHANGED
|
@@ -151,7 +151,7 @@ curl -X POST http://localhost:8000/step -H "Content-Type: application/json" -d '
|
|
| 151 |
```bash
|
| 152 |
export OPENAI_API_KEY=sk-...
|
| 153 |
export MODEL_NAME=gpt-4o-mini
|
| 154 |
-
export
|
| 155 |
|
| 156 |
python inference.py
|
| 157 |
```
|
|
|
|
| 151 |
```bash
|
| 152 |
export OPENAI_API_KEY=sk-...
|
| 153 |
export MODEL_NAME=gpt-4o-mini
|
| 154 |
+
export ENV_BASE_URL=http://localhost:8000
|
| 155 |
|
| 156 |
python inference.py
|
| 157 |
```
|
inference.py
CHANGED
|
@@ -26,7 +26,7 @@ from openai import OpenAI
|
|
| 26 |
# Config
|
| 27 |
# ------------------------------------------------------------------
|
| 28 |
|
| 29 |
-
|
| 30 |
MODEL_NAME = os.environ.get("MODEL_NAME", "gpt-4o-mini")
|
| 31 |
HF_TOKEN = os.environ.get("HF_TOKEN", "")
|
| 32 |
OPENAI_API_KEY = os.environ.get("OPENAI_API_KEY", "")
|
|
@@ -112,10 +112,10 @@ class EnvClient:
|
|
| 112 |
def create_openai_client() -> OpenAI:
|
| 113 |
"""Create OpenAI client with appropriate config."""
|
| 114 |
api_key = OPENAI_API_KEY or HF_TOKEN or "no-key"
|
| 115 |
-
base_url =
|
| 116 |
|
| 117 |
# If using HF inference endpoint, set base_url
|
| 118 |
-
if HF_TOKEN and not OPENAI_API_KEY:
|
| 119 |
base_url = f"https://api-inference.huggingface.co/models/{MODEL_NAME}/v1"
|
| 120 |
|
| 121 |
return OpenAI(api_key=api_key, base_url=base_url)
|
|
@@ -269,10 +269,10 @@ def main():
|
|
| 269 |
print("=" * 60)
|
| 270 |
print("SRE Incident Response — OpenEnv Inference")
|
| 271 |
print(f"Model: {MODEL_NAME}")
|
| 272 |
-
print(f"Environment: {
|
| 273 |
print("=" * 60)
|
| 274 |
|
| 275 |
-
env = EnvClient(
|
| 276 |
llm = create_openai_client()
|
| 277 |
|
| 278 |
results = []
|
|
|
|
| 26 |
# Config
|
| 27 |
# ------------------------------------------------------------------
|
| 28 |
|
| 29 |
+
ENV_BASE_URL = os.environ.get("ENV_BASE_URL", "http://localhost:8000")
|
| 30 |
MODEL_NAME = os.environ.get("MODEL_NAME", "gpt-4o-mini")
|
| 31 |
HF_TOKEN = os.environ.get("HF_TOKEN", "")
|
| 32 |
OPENAI_API_KEY = os.environ.get("OPENAI_API_KEY", "")
|
|
|
|
| 112 |
def create_openai_client() -> OpenAI:
|
| 113 |
"""Create OpenAI client with appropriate config."""
|
| 114 |
api_key = OPENAI_API_KEY or HF_TOKEN or "no-key"
|
| 115 |
+
base_url = os.environ.get("API_BASE_URL")
|
| 116 |
|
| 117 |
# If using HF inference endpoint, set base_url
|
| 118 |
+
if HF_TOKEN and not OPENAI_API_KEY and not base_url:
|
| 119 |
base_url = f"https://api-inference.huggingface.co/models/{MODEL_NAME}/v1"
|
| 120 |
|
| 121 |
return OpenAI(api_key=api_key, base_url=base_url)
|
|
|
|
| 269 |
print("=" * 60)
|
| 270 |
print("SRE Incident Response — OpenEnv Inference")
|
| 271 |
print(f"Model: {MODEL_NAME}")
|
| 272 |
+
print(f"Environment: {ENV_BASE_URL}")
|
| 273 |
print("=" * 60)
|
| 274 |
|
| 275 |
+
env = EnvClient(ENV_BASE_URL)
|
| 276 |
llm = create_openai_client()
|
| 277 |
|
| 278 |
results = []
|