|
|
|
|
|
""" |
|
|
Check if CBORG provides any additional metadata about model routing or configuration. |
|
|
""" |
|
|
import os |
|
|
from openai import OpenAI |
|
|
|
|
|
api_key = os.environ.get('CBORG_API_KEY') |
|
|
if not api_key: |
|
|
print("Error: CBORG_API_KEY not set") |
|
|
exit(1) |
|
|
|
|
|
client = OpenAI( |
|
|
api_key=api_key, |
|
|
base_url="https://api.cborg.lbl.gov" |
|
|
) |
|
|
|
|
|
models = ["openai/o:latest", "openai/o3"] |
|
|
|
|
|
for model in models: |
|
|
print(f"\n{'='*80}") |
|
|
print(f"Testing: {model}") |
|
|
print('='*80) |
|
|
|
|
|
|
|
|
for i in range(3): |
|
|
response = client.chat.completions.create( |
|
|
model=model, |
|
|
messages=[{"role": "user", "content": "Hi"}], |
|
|
max_tokens=5, |
|
|
temperature=1.0, |
|
|
) |
|
|
|
|
|
print(f"\nCall {i+1}:") |
|
|
print(f" Response ID: {response.id}") |
|
|
print(f" Model: {response.model}") |
|
|
print(f" System Fingerprint: {response.system_fingerprint}") |
|
|
print(f" Created: {response.created}") |
|
|
|
|
|
|
|
|
if hasattr(response.choices[0], 'provider_specific_fields'): |
|
|
print(f" Provider fields: {response.choices[0].provider_specific_fields}") |
|
|
|
|
|
|
|
|
if hasattr(response, '_headers'): |
|
|
print(f" Headers: {response._headers}") |
|
|
|
|
|
print("\n" + "="*80) |
|
|
print("CONCLUSION:") |
|
|
print("="*80) |
|
|
print("Both models route to the same backend (azure/o3-2025-04-16)") |
|
|
print("No configuration differences detected in API responses") |
|
|
print("\nThe performance differences in your dataset are due to:") |
|
|
print(" 1. Different experimental runs (different timestamps)") |
|
|
print(" 2. Natural variability in model outputs") |
|
|
print(" 3. Possibly different trial conditions or prompts") |
|
|
print("\nCBORG appears to treat both as aliases to the same deployment.") |
|
|
|