frdel commited on
Commit
8054905
·
1 Parent(s): 0e2acd3

model api_base, litellm finalizing

Browse files
Files changed (5) hide show
  1. agent.py +10 -3
  2. initialize.py +4 -0
  3. models.py +66 -130
  4. python/helpers/settings.py +45 -27
  5. python/tools/browser_agent.py +21 -8
agent.py CHANGED
@@ -206,6 +206,7 @@ class AgentContext:
206
  class ModelConfig:
207
  provider: models.ModelProvider
208
  name: str
 
209
  ctx_length: int = 0
210
  limit_requests: int = 0
211
  limit_input: int = 0
@@ -581,23 +582,29 @@ class Agent:
581
  return models.get_chat_model(
582
  self.config.chat_model.provider,
583
  self.config.chat_model.name,
584
- **self.config.chat_model.kwargs,
585
  )
586
 
587
  def get_utility_model(self):
588
  return models.get_chat_model(
589
  self.config.utility_model.provider,
590
  self.config.utility_model.name,
591
- **self.config.utility_model.kwargs,
592
  )
593
 
594
  def get_embedding_model(self):
595
  return models.get_embedding_model(
596
  self.config.embeddings_model.provider,
597
  self.config.embeddings_model.name,
598
- **self.config.embeddings_model.kwargs,
599
  )
600
 
 
 
 
 
 
 
601
  async def call_utility_model(
602
  self,
603
  system: str,
 
206
  class ModelConfig:
207
  provider: models.ModelProvider
208
  name: str
209
+ api_base: str = ""
210
  ctx_length: int = 0
211
  limit_requests: int = 0
212
  limit_input: int = 0
 
582
  return models.get_chat_model(
583
  self.config.chat_model.provider,
584
  self.config.chat_model.name,
585
+ **self._get_model_kwargs(self.config.chat_model),
586
  )
587
 
588
  def get_utility_model(self):
589
  return models.get_chat_model(
590
  self.config.utility_model.provider,
591
  self.config.utility_model.name,
592
+ **self._get_model_kwargs(self.config.utility_model),
593
  )
594
 
595
  def get_embedding_model(self):
596
  return models.get_embedding_model(
597
  self.config.embeddings_model.provider,
598
  self.config.embeddings_model.name,
599
+ **self._get_model_kwargs(self.config.embeddings_model),
600
  )
601
 
602
+ def _get_model_kwargs(self, model_config: ModelConfig):
603
+ kwargs = model_config.kwargs.copy() or {}
604
+ if model_config.api_base and "api_base" not in kwargs:
605
+ kwargs["api_base"] = model_config.api_base
606
+ return kwargs
607
+
608
  async def call_utility_model(
609
  self,
610
  system: str,
initialize.py CHANGED
@@ -29,6 +29,7 @@ def initialize_agent():
29
  chat_llm = ModelConfig(
30
  provider=models.ModelProvider[current_settings["chat_model_provider"]],
31
  name=current_settings["chat_model_name"],
 
32
  ctx_length=current_settings["chat_model_ctx_length"],
33
  vision=current_settings["chat_model_vision"],
34
  limit_requests=current_settings["chat_model_rl_requests"],
@@ -41,6 +42,7 @@ def initialize_agent():
41
  utility_llm = ModelConfig(
42
  provider=models.ModelProvider[current_settings["util_model_provider"]],
43
  name=current_settings["util_model_name"],
 
44
  ctx_length=current_settings["util_model_ctx_length"],
45
  limit_requests=current_settings["util_model_rl_requests"],
46
  limit_input=current_settings["util_model_rl_input"],
@@ -51,6 +53,7 @@ def initialize_agent():
51
  embedding_llm = ModelConfig(
52
  provider=models.ModelProvider[current_settings["embed_model_provider"]],
53
  name=current_settings["embed_model_name"],
 
54
  limit_requests=current_settings["embed_model_rl_requests"],
55
  kwargs=_normalize_model_kwargs(current_settings["embed_model_kwargs"]),
56
  )
@@ -58,6 +61,7 @@ def initialize_agent():
58
  browser_llm = ModelConfig(
59
  provider=models.ModelProvider[current_settings["browser_model_provider"]],
60
  name=current_settings["browser_model_name"],
 
61
  vision=current_settings["browser_model_vision"],
62
  kwargs=_normalize_model_kwargs(current_settings["browser_model_kwargs"]),
63
  )
 
29
  chat_llm = ModelConfig(
30
  provider=models.ModelProvider[current_settings["chat_model_provider"]],
31
  name=current_settings["chat_model_name"],
32
+ api_base=current_settings["chat_model_api_base"],
33
  ctx_length=current_settings["chat_model_ctx_length"],
34
  vision=current_settings["chat_model_vision"],
35
  limit_requests=current_settings["chat_model_rl_requests"],
 
42
  utility_llm = ModelConfig(
43
  provider=models.ModelProvider[current_settings["util_model_provider"]],
44
  name=current_settings["util_model_name"],
45
+ api_base=current_settings["util_model_api_base"],
46
  ctx_length=current_settings["util_model_ctx_length"],
47
  limit_requests=current_settings["util_model_rl_requests"],
48
  limit_input=current_settings["util_model_rl_input"],
 
53
  embedding_llm = ModelConfig(
54
  provider=models.ModelProvider[current_settings["embed_model_provider"]],
55
  name=current_settings["embed_model_name"],
56
+ api_base=current_settings["embed_model_api_base"],
57
  limit_requests=current_settings["embed_model_rl_requests"],
58
  kwargs=_normalize_model_kwargs(current_settings["embed_model_kwargs"]),
59
  )
 
61
  browser_llm = ModelConfig(
62
  provider=models.ModelProvider[current_settings["browser_model_provider"]],
63
  name=current_settings["browser_model_name"],
64
+ api_base=current_settings["browser_model_api_base"],
65
  vision=current_settings["browser_model_vision"],
66
  kwargs=_normalize_model_kwargs(current_settings["browser_model_kwargs"]),
67
  )
models.py CHANGED
@@ -59,19 +59,18 @@ class ModelType(Enum):
59
 
60
  class ModelProvider(Enum):
61
  ANTHROPIC = "Anthropic"
62
- CHUTES = "Chutes"
63
  DEEPSEEK = "DeepSeek"
64
- GOOGLE = "Google"
65
  GROQ = "Groq"
66
  HUGGINGFACE = "HuggingFace"
67
- LMSTUDIO = "LM Studio"
68
- MISTRALAI = "Mistral AI"
69
  OLLAMA = "Ollama"
70
  OPENAI = "OpenAI"
71
  AZURE = "OpenAI Azure"
72
  OPENROUTER = "OpenRouter"
73
  SAMBANOVA = "Sambanova"
74
- OTHER = "Other"
75
 
76
 
77
  class ChatChunk(TypedDict):
@@ -84,42 +83,6 @@ class ChatChunk(TypedDict):
84
  rate_limiters: dict[str, RateLimiter] = {}
85
 
86
 
87
- def configure_litellm_environment():
88
- env_mappings = {
89
- "API_KEY_OPENAI": "OPENAI_API_KEY",
90
- "API_KEY_ANTHROPIC": "ANTHROPIC_API_KEY",
91
- "API_KEY_GROQ": "GROQ_API_KEY",
92
- "API_KEY_GOOGLE": "GOOGLE_API_KEY",
93
- "API_KEY_MISTRAL": "MISTRAL_API_KEY",
94
- "API_KEY_OLLAMA": "OLLAMA_API_KEY",
95
- "API_KEY_HUGGINGFACE": "HUGGINGFACE_API_KEY",
96
- "API_KEY_OPENAI_AZURE": "AZURE_AI_API_KEY",
97
- "API_KEY_DEEPSEEK": "DEEPSEEK_API_KEY",
98
- "API_KEY_SAMBANOVA": "SAMBANOVA_API_KEY",
99
- "API_KEY_GOOGLE": "GEMINI_API_KEY",
100
- }
101
- base_url_mappings = {
102
- "OPENAI_BASE_URL": "OPENAI_API_BASE",
103
- "ANTHROPIC_BASE_URL": "ANTHROPIC_API_BASE",
104
- "GROQ_BASE_URL": "GROQ_API_BASE",
105
- "GOOGLE_BASE_URL": "GOOGLE_API_BASE",
106
- "MISTRAL_BASE_URL": "MISTRAL_API_BASE",
107
- "OLLAMA_BASE_URL": "OLLAMA_API_BASE",
108
- "HUGGINGFACE_BASE_URL": "HUGGINGFACE_API_BASE",
109
- "AZURE_BASE_URL": "AZURE_AI_API_BASE",
110
- "DEEPSEEK_BASE_URL": "DEEPSEEK_API_BASE",
111
- "SAMBANOVA_BASE_URL": "SAMBANOVA_API_BASE",
112
- }
113
- for a0, llm in env_mappings.items():
114
- val = dotenv.get_dotenv_value(a0)
115
- if val and not os.getenv(llm):
116
- os.environ[llm] = val
117
- for a0_base, llm_base in base_url_mappings.items():
118
- val = dotenv.get_dotenv_value(a0_base)
119
- if val and not os.getenv(llm_base):
120
- os.environ[llm_base] = val
121
-
122
-
123
  def get_api_key(service: str) -> str:
124
  return (
125
  dotenv.get_dotenv_value(f"API_KEY_{service.upper()}")
@@ -140,26 +103,6 @@ def get_rate_limiter(
140
  return limiter
141
 
142
 
143
- def _parse_chunk(chunk: Any) -> ChatChunk:
144
- delta = chunk["choices"][0].get("delta", {})
145
- message = chunk["choices"][0].get("model_extra", {}).get("message", {})
146
- response_delta = (
147
- delta.get("content", "")
148
- if isinstance(delta, dict)
149
- else getattr(delta, "content", "")
150
- ) or (
151
- message.get("content", "")
152
- if isinstance(message, dict)
153
- else getattr(message, "content", "")
154
- )
155
- reasoning_delta = (
156
- delta.get("reasoning_content", "")
157
- if isinstance(delta, dict)
158
- else getattr(delta, "reasoning_content", "")
159
- )
160
- return ChatChunk(reasoning_delta=reasoning_delta, response_delta=response_delta)
161
-
162
-
163
  class LiteLLMChatWrapper(SimpleChatModel):
164
  model_name: str
165
  provider: str
@@ -284,7 +227,7 @@ class LiteLLMChatWrapper(SimpleChatModel):
284
  self,
285
  system_message="",
286
  user_message="",
287
- messages: List[BaseMessage]|None = None,
288
  response_callback: Callable[[str, str], Awaitable[None]] | None = None,
289
  reasoning_callback: Callable[[str, str], Awaitable[None]] | None = None,
290
  tokens_callback: Callable[[str, int], Awaitable[None]] | None = None,
@@ -424,66 +367,86 @@ def _get_litellm_chat(
424
  provider_name: str = "",
425
  **kwargs: Any,
426
  ):
427
- provider_name = provider_name.lower()
428
-
429
- configure_litellm_environment()
430
- # Use original provider name for API key lookup, fallback to mapped provider name
431
  api_key = kwargs.pop("api_key", None) or get_api_key(provider_name)
432
 
433
- # litellm will pick up base_url from env. We just need to control the api_key.
434
- # base_url = dotenv.get_dotenv_value(f"{provider_name.upper()}_BASE_URL")
435
-
436
- # If a base_url is set, ensure api_key is not passed to litellm
437
- # > remove, this can be handled by api_key=None
438
- # if base_url:
439
- # if "api_key" in kwargs:
440
- # del kwargs["api_key"]
441
- # Only pass API key if no base_url is set and key is not a placeholder
442
  if api_key and api_key not in ("None", "NA"):
443
  kwargs["api_key"] = api_key
444
 
445
- # for openrouter add app reference
446
- if provider_name == "openrouter":
447
- kwargs["extra_headers"] = {
448
- "HTTP-Referer": "https://agent-zero.ai",
449
- "X-Title": "Agent Zero",
450
- }
451
-
452
- return cls(model=model_name, provider=provider_name, **kwargs)
453
 
454
 
455
- def get_litellm_embedding(model_name: str, provider: str, **kwargs: Any):
456
  # Check if this is a local sentence-transformers model
457
- if provider == "huggingface" and model_name.startswith("sentence-transformers/"):
 
 
458
  # Use local sentence-transformers instead of LiteLLM for local models
459
- return LocalSentenceTransformerWrapper(provider=provider, model=model_name, **kwargs)
460
-
461
- configure_litellm_environment()
462
- # Use original provider name for API key lookup, fallback to mapped provider name
463
- api_key = kwargs.pop("api_key", None) or get_api_key(provider)
 
464
 
465
- # litellm will pick up base_url from env. We just need to control the api_key.
466
- # base_url = dotenv.get_dotenv_value(f"{provider.upper()}_BASE_URL")
467
 
468
- # If a base_url is set, ensure api_key is not passed to litellm
469
- # > remove, this can be handled by api_key=None
470
- # if base_url:
471
- # if "api_key" in kwargs:
472
- # del kwargs["api_key"]
473
- # Only pass API key if no base_url is set and key is not a placeholder
474
  if api_key and api_key not in ("None", "NA"):
475
  kwargs["api_key"] = api_key
476
 
477
- return LiteLLMEmbeddingWrapper(model=model_name, provider=provider, **kwargs)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
478
 
479
 
480
  def get_model(type: ModelType, provider: ModelProvider, name: str, **kwargs: Any):
481
  provider_name = provider.name.lower()
482
- kwargs = _normalize_chat_kwargs(provider, kwargs)
483
  if type == ModelType.CHAT:
484
  return _get_litellm_chat(LiteLLMChatWrapper, name, provider_name, **kwargs)
485
  elif type == ModelType.EMBEDDING:
486
- return get_litellm_embedding(name, provider_name, **kwargs)
487
  else:
488
  raise ValueError(f"Unsupported model type: {type}")
489
 
@@ -491,8 +454,7 @@ def get_model(type: ModelType, provider: ModelProvider, name: str, **kwargs: Any
491
  def get_chat_model(
492
  provider: ModelProvider, name: str, **kwargs: Any
493
  ) -> LiteLLMChatWrapper:
494
- provider_name = _get_litellm_provider(provider)
495
- kwargs = _normalize_chat_kwargs(provider, kwargs)
496
  model = _get_litellm_chat(LiteLLMChatWrapper, name, provider_name, **kwargs)
497
  return model
498
 
@@ -501,7 +463,6 @@ def get_browser_model(
501
  provider: ModelProvider, name: str, **kwargs: Any
502
  ) -> BrowserCompatibleChatWrapper:
503
  provider_name = provider.name.lower()
504
- kwargs = _normalize_chat_kwargs(provider, kwargs)
505
  model = _get_litellm_chat(
506
  BrowserCompatibleChatWrapper, name, provider_name, **kwargs
507
  )
@@ -512,30 +473,5 @@ def get_embedding_model(
512
  provider: ModelProvider, name: str, **kwargs: Any
513
  ) -> LiteLLMEmbeddingWrapper | LocalSentenceTransformerWrapper:
514
  provider_name = provider.name.lower()
515
- kwargs = _normalize_embedding_kwargs(kwargs)
516
- model = get_litellm_embedding(name, provider_name, **kwargs)
517
  return model
518
-
519
-
520
- def _normalize_chat_kwargs(provider: ModelProvider, kwargs: Any) -> Any:
521
- # this prevents using openai api key for other providers
522
- if provider == ModelProvider.OTHER:
523
- if "api_key" not in kwargs:
524
- kwargs["api_key"] = "None"
525
- return kwargs
526
-
527
-
528
- def _normalize_embedding_kwargs(kwargs: Any) -> Any:
529
- return kwargs
530
-
531
-
532
- def _get_litellm_provider(provider: ModelProvider) -> str:
533
- name = provider.name.lower()
534
-
535
- # exceptions
536
- if name == "google":
537
- name = "gemini"
538
- elif name == "other":
539
- name = "openai"
540
-
541
- return name
 
59
 
60
  class ModelProvider(Enum):
61
  ANTHROPIC = "Anthropic"
 
62
  DEEPSEEK = "DeepSeek"
63
+ GEMINI = "Google"
64
  GROQ = "Groq"
65
  HUGGINGFACE = "HuggingFace"
66
+ LM_STUDIO = "LM Studio"
67
+ MISTRAL = "Mistral AI"
68
  OLLAMA = "Ollama"
69
  OPENAI = "OpenAI"
70
  AZURE = "OpenAI Azure"
71
  OPENROUTER = "OpenRouter"
72
  SAMBANOVA = "Sambanova"
73
+ OTHER = "Other OpenAI compatible"
74
 
75
 
76
  class ChatChunk(TypedDict):
 
83
  rate_limiters: dict[str, RateLimiter] = {}
84
 
85
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
86
  def get_api_key(service: str) -> str:
87
  return (
88
  dotenv.get_dotenv_value(f"API_KEY_{service.upper()}")
 
103
  return limiter
104
 
105
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
106
  class LiteLLMChatWrapper(SimpleChatModel):
107
  model_name: str
108
  provider: str
 
227
  self,
228
  system_message="",
229
  user_message="",
230
+ messages: List[BaseMessage] | None = None,
231
  response_callback: Callable[[str, str], Awaitable[None]] | None = None,
232
  reasoning_callback: Callable[[str, str], Awaitable[None]] | None = None,
233
  tokens_callback: Callable[[str, int], Awaitable[None]] | None = None,
 
367
  provider_name: str = "",
368
  **kwargs: Any,
369
  ):
370
+ # use api key from kwargs or env
 
 
 
371
  api_key = kwargs.pop("api_key", None) or get_api_key(provider_name)
372
 
373
+ # Only pass API key if key is not a placeholder
 
 
 
 
 
 
 
 
374
  if api_key and api_key not in ("None", "NA"):
375
  kwargs["api_key"] = api_key
376
 
377
+ provider_name, model_name, kwargs = _adjust_call_args(
378
+ provider_name, model_name, kwargs
379
+ )
380
+ return cls(provider=provider_name, model=model_name, **kwargs)
 
 
 
 
381
 
382
 
383
+ def _get_litellm_embedding(model_name: str, provider_name: str, **kwargs: Any):
384
  # Check if this is a local sentence-transformers model
385
+ if provider_name == "huggingface" and model_name.startswith(
386
+ "sentence-transformers/"
387
+ ):
388
  # Use local sentence-transformers instead of LiteLLM for local models
389
+ provider_name, model_name, kwargs = _adjust_call_args(
390
+ provider_name, model_name, kwargs
391
+ )
392
+ return LocalSentenceTransformerWrapper(
393
+ provider=provider_name, model=model_name, **kwargs
394
+ )
395
 
396
+ # use api key from kwargs or env
397
+ api_key = kwargs.pop("api_key", None) or get_api_key(provider_name)
398
 
399
+ # Only pass API key if key is not a placeholder
 
 
 
 
 
400
  if api_key and api_key not in ("None", "NA"):
401
  kwargs["api_key"] = api_key
402
 
403
+ provider_name, model_name, kwargs = _adjust_call_args(
404
+ provider_name, model_name, kwargs
405
+ )
406
+ return LiteLLMEmbeddingWrapper(model=model_name, provider=provider_name, **kwargs)
407
+
408
+
409
+ def _parse_chunk(chunk: Any) -> ChatChunk:
410
+ delta = chunk["choices"][0].get("delta", {})
411
+ message = chunk["choices"][0].get("model_extra", {}).get("message", {})
412
+ response_delta = (
413
+ delta.get("content", "")
414
+ if isinstance(delta, dict)
415
+ else getattr(delta, "content", "")
416
+ ) or (
417
+ message.get("content", "")
418
+ if isinstance(message, dict)
419
+ else getattr(message, "content", "")
420
+ )
421
+ reasoning_delta = (
422
+ delta.get("reasoning_content", "")
423
+ if isinstance(delta, dict)
424
+ else getattr(delta, "reasoning_content", "")
425
+ )
426
+ return ChatChunk(reasoning_delta=reasoning_delta, response_delta=response_delta)
427
+
428
+
429
+ def _adjust_call_args(provider_name: str, model_name: str, kwargs: dict):
430
+ # for openrouter add app reference
431
+ if provider_name == "openrouter":
432
+ kwargs["extra_headers"] = {
433
+ "HTTP-Referer": "https://agent-zero.ai",
434
+ "X-Title": "Agent Zero",
435
+ }
436
+
437
+ # remap other to openai for litellm
438
+ if provider_name == "other":
439
+ provider_name = "openai"
440
+
441
+ return provider_name, model_name, kwargs
442
 
443
 
444
  def get_model(type: ModelType, provider: ModelProvider, name: str, **kwargs: Any):
445
  provider_name = provider.name.lower()
 
446
  if type == ModelType.CHAT:
447
  return _get_litellm_chat(LiteLLMChatWrapper, name, provider_name, **kwargs)
448
  elif type == ModelType.EMBEDDING:
449
+ return _get_litellm_embedding(name, provider_name, **kwargs)
450
  else:
451
  raise ValueError(f"Unsupported model type: {type}")
452
 
 
454
  def get_chat_model(
455
  provider: ModelProvider, name: str, **kwargs: Any
456
  ) -> LiteLLMChatWrapper:
457
+ provider_name = provider.name.lower()
 
458
  model = _get_litellm_chat(LiteLLMChatWrapper, name, provider_name, **kwargs)
459
  return model
460
 
 
463
  provider: ModelProvider, name: str, **kwargs: Any
464
  ) -> BrowserCompatibleChatWrapper:
465
  provider_name = provider.name.lower()
 
466
  model = _get_litellm_chat(
467
  BrowserCompatibleChatWrapper, name, provider_name, **kwargs
468
  )
 
473
  provider: ModelProvider, name: str, **kwargs: Any
474
  ) -> LiteLLMEmbeddingWrapper | LocalSentenceTransformerWrapper:
475
  provider_name = provider.name.lower()
476
+ model = _get_litellm_embedding(name, provider_name, **kwargs)
 
477
  return model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
python/helpers/settings.py CHANGED
@@ -17,6 +17,7 @@ class Settings(TypedDict):
17
 
18
  chat_model_provider: str
19
  chat_model_name: str
 
20
  chat_model_kwargs: dict[str, str]
21
  chat_model_ctx_length: int
22
  chat_model_ctx_history: float
@@ -27,6 +28,7 @@ class Settings(TypedDict):
27
 
28
  util_model_provider: str
29
  util_model_name: str
 
30
  util_model_kwargs: dict[str, str]
31
  util_model_ctx_length: int
32
  util_model_ctx_input: float
@@ -36,12 +38,14 @@ class Settings(TypedDict):
36
 
37
  embed_model_provider: str
38
  embed_model_name: str
 
39
  embed_model_kwargs: dict[str, str]
40
  embed_model_rl_requests: int
41
  embed_model_rl_input: int
42
 
43
  browser_model_provider: str
44
  browser_model_name: str
 
45
  browser_model_vision: bool
46
  browser_model_kwargs: dict[str, str]
47
 
@@ -141,6 +145,16 @@ def convert_out(settings: Settings) -> SettingsOutput:
141
  }
142
  )
143
 
 
 
 
 
 
 
 
 
 
 
144
  chat_model_fields.append(
145
  {
146
  "id": "chat_model_ctx_length",
@@ -208,8 +222,7 @@ def convert_out(settings: Settings) -> SettingsOutput:
208
  {
209
  "id": "chat_model_kwargs",
210
  "title": "Chat model additional parameters",
211
- "description": """Any other parameters supported by the model. Format is KEY=VALUE on individual lines, just like .env file.
212
- For OpenAI compatible providers not listed here, select 'other' and specify api_base=https://... and api_key=... as additional parameters.""",
213
  "type": "textarea",
214
  "value": _dict_to_env(settings["chat_model_kwargs"]),
215
  }
@@ -245,6 +258,16 @@ def convert_out(settings: Settings) -> SettingsOutput:
245
  }
246
  )
247
 
 
 
 
 
 
 
 
 
 
 
248
  util_model_fields.append(
249
  {
250
  "id": "util_model_rl_requests",
@@ -279,8 +302,7 @@ def convert_out(settings: Settings) -> SettingsOutput:
279
  {
280
  "id": "util_model_kwargs",
281
  "title": "Utility model additional parameters",
282
- "description": """Any other parameters supported by the model. Format is KEY=VALUE on individual lines, just like .env file.
283
- For OpenAI compatible providers not listed here, select 'other' and specify api_base=https://... and api_key=... as additional parameters.""",
284
  "type": "textarea",
285
  "value": _dict_to_env(settings["util_model_kwargs"]),
286
  }
@@ -316,6 +338,16 @@ def convert_out(settings: Settings) -> SettingsOutput:
316
  }
317
  )
318
 
 
 
 
 
 
 
 
 
 
 
319
  embed_model_fields.append(
320
  {
321
  "id": "embed_model_rl_requests",
@@ -340,8 +372,7 @@ def convert_out(settings: Settings) -> SettingsOutput:
340
  {
341
  "id": "embed_model_kwargs",
342
  "title": "Embedding model additional parameters",
343
- "description": """Any other parameters supported by the model. Format is KEY=VALUE on individual lines, just like .env file.
344
- For OpenAI compatible providers not listed here, select 'other' and specify api_base=https://... and api_key=... as additional parameters.""",
345
  "type": "textarea",
346
  "value": _dict_to_env(settings["embed_model_kwargs"]),
347
  }
@@ -391,7 +422,7 @@ def convert_out(settings: Settings) -> SettingsOutput:
391
  {
392
  "id": "browser_model_kwargs",
393
  "title": "Web Browser model additional parameters",
394
- "description": "Any other parameters supported by the model. Format is KEY=VALUE on individual lines, just like .env file.",
395
  "type": "textarea",
396
  "value": _dict_to_env(settings["browser_model_kwargs"]),
397
  }
@@ -472,26 +503,9 @@ def convert_out(settings: Settings) -> SettingsOutput:
472
 
473
  # api keys model section
474
  api_keys_fields: list[SettingsField] = []
475
- api_keys_fields.append(_get_api_key_field(settings, "openai", "OpenAI API Key"))
476
- api_keys_fields.append(
477
- _get_api_key_field(settings, "anthropic", "Anthropic API Key")
478
- )
479
- api_keys_fields.append(_get_api_key_field(settings, "chutes", "Chutes API Key"))
480
- api_keys_fields.append(_get_api_key_field(settings, "deepseek", "DeepSeek API Key"))
481
- api_keys_fields.append(_get_api_key_field(settings, "google", "Google API Key"))
482
- api_keys_fields.append(_get_api_key_field(settings, "groq", "Groq API Key"))
483
- api_keys_fields.append(
484
- _get_api_key_field(settings, "huggingface", "HuggingFace API Key")
485
- )
486
- api_keys_fields.append(
487
- _get_api_key_field(settings, "mistralai", "MistralAI API Key")
488
- )
489
- api_keys_fields.append(
490
- _get_api_key_field(settings, "openrouter", "OpenRouter API Key")
491
- )
492
- api_keys_fields.append(
493
- _get_api_key_field(settings, "sambanova", "Sambanova API Key")
494
- )
495
 
496
  api_keys_section: SettingsSection = {
497
  "id": "api_keys",
@@ -965,6 +979,7 @@ def get_default_settings() -> Settings:
965
  version=_get_version(),
966
  chat_model_provider=ModelProvider.OPENROUTER.name,
967
  chat_model_name="openai/gpt-4.1",
 
968
  chat_model_kwargs={"temperature": "0"},
969
  chat_model_ctx_length=100000,
970
  chat_model_ctx_history=0.7,
@@ -974,6 +989,7 @@ def get_default_settings() -> Settings:
974
  chat_model_rl_output=0,
975
  util_model_provider=ModelProvider.OPENROUTER.name,
976
  util_model_name="openai/gpt-4.1-nano",
 
977
  util_model_ctx_length=100000,
978
  util_model_ctx_input=0.7,
979
  util_model_kwargs={"temperature": "0"},
@@ -982,11 +998,13 @@ def get_default_settings() -> Settings:
982
  util_model_rl_output=0,
983
  embed_model_provider=ModelProvider.HUGGINGFACE.name,
984
  embed_model_name="sentence-transformers/all-MiniLM-L6-v2",
 
985
  embed_model_kwargs={},
986
  embed_model_rl_requests=0,
987
  embed_model_rl_input=0,
988
  browser_model_provider=ModelProvider.OPENROUTER.name,
989
  browser_model_name="openai/gpt-4.1",
 
990
  browser_model_vision=True,
991
  browser_model_kwargs={"temperature": "0"},
992
  api_keys={},
 
17
 
18
  chat_model_provider: str
19
  chat_model_name: str
20
+ chat_model_api_base: str
21
  chat_model_kwargs: dict[str, str]
22
  chat_model_ctx_length: int
23
  chat_model_ctx_history: float
 
28
 
29
  util_model_provider: str
30
  util_model_name: str
31
+ util_model_api_base: str
32
  util_model_kwargs: dict[str, str]
33
  util_model_ctx_length: int
34
  util_model_ctx_input: float
 
38
 
39
  embed_model_provider: str
40
  embed_model_name: str
41
+ embed_model_api_base: str
42
  embed_model_kwargs: dict[str, str]
43
  embed_model_rl_requests: int
44
  embed_model_rl_input: int
45
 
46
  browser_model_provider: str
47
  browser_model_name: str
48
+ browser_model_api_base: str
49
  browser_model_vision: bool
50
  browser_model_kwargs: dict[str, str]
51
 
 
145
  }
146
  )
147
 
148
+ chat_model_fields.append(
149
+ {
150
+ "id": "chat_model_api_base",
151
+ "title": "Chat model API base URL",
152
+ "description": "API base URL for main chat model. Leave empty for default. Only relevant for Azure, local and custom (other) providers.",
153
+ "type": "text",
154
+ "value": settings["chat_model_api_base"],
155
+ }
156
+ )
157
+
158
  chat_model_fields.append(
159
  {
160
  "id": "chat_model_ctx_length",
 
222
  {
223
  "id": "chat_model_kwargs",
224
  "title": "Chat model additional parameters",
225
+ "description": "Any other parameters supported by <a href='https://docs.litellm.ai/docs/set_keys' target='_blank'>LiteLLM</a>. Format is KEY=VALUE on individual lines, just like .env file.",
 
226
  "type": "textarea",
227
  "value": _dict_to_env(settings["chat_model_kwargs"]),
228
  }
 
258
  }
259
  )
260
 
261
+ util_model_fields.append(
262
+ {
263
+ "id": "util_model_api_base",
264
+ "title": "Utility model API base URL",
265
+ "description": "API base URL for utility model. Leave empty for default. Only relevant for Azure, local and custom (other) providers.",
266
+ "type": "text",
267
+ "value": settings["util_model_api_base"],
268
+ }
269
+ )
270
+
271
  util_model_fields.append(
272
  {
273
  "id": "util_model_rl_requests",
 
302
  {
303
  "id": "util_model_kwargs",
304
  "title": "Utility model additional parameters",
305
+ "description": "Any other parameters supported by <a href='https://docs.litellm.ai/docs/set_keys' target='_blank'>LiteLLM</a>. Format is KEY=VALUE on individual lines, just like .env file.",
 
306
  "type": "textarea",
307
  "value": _dict_to_env(settings["util_model_kwargs"]),
308
  }
 
338
  }
339
  )
340
 
341
+ embed_model_fields.append(
342
+ {
343
+ "id": "embed_model_api_base",
344
+ "title": "Embedding model API base URL",
345
+ "description": "API base URL for embedding model. Leave empty for default. Only relevant for Azure, local and custom (other) providers.",
346
+ "type": "text",
347
+ "value": settings["embed_model_api_base"],
348
+ }
349
+ )
350
+
351
  embed_model_fields.append(
352
  {
353
  "id": "embed_model_rl_requests",
 
372
  {
373
  "id": "embed_model_kwargs",
374
  "title": "Embedding model additional parameters",
375
+ "description": "Any other parameters supported by <a href='https://docs.litellm.ai/docs/set_keys' target='_blank'>LiteLLM</a>. Format is KEY=VALUE on individual lines, just like .env file.",
 
376
  "type": "textarea",
377
  "value": _dict_to_env(settings["embed_model_kwargs"]),
378
  }
 
422
  {
423
  "id": "browser_model_kwargs",
424
  "title": "Web Browser model additional parameters",
425
+ "description": "Any other parameters supported by <a href='https://docs.litellm.ai/docs/set_keys' target='_blank'>LiteLLM</a>. Format is KEY=VALUE on individual lines, just like .env file.",
426
  "type": "textarea",
427
  "value": _dict_to_env(settings["browser_model_kwargs"]),
428
  }
 
503
 
504
  # api keys model section
505
  api_keys_fields: list[SettingsField] = []
506
+
507
+ for provider in ModelProvider:
508
+ api_keys_fields.append(_get_api_key_field(settings, provider.name.lower(), provider.value))
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
509
 
510
  api_keys_section: SettingsSection = {
511
  "id": "api_keys",
 
979
  version=_get_version(),
980
  chat_model_provider=ModelProvider.OPENROUTER.name,
981
  chat_model_name="openai/gpt-4.1",
982
+ chat_model_api_base="",
983
  chat_model_kwargs={"temperature": "0"},
984
  chat_model_ctx_length=100000,
985
  chat_model_ctx_history=0.7,
 
989
  chat_model_rl_output=0,
990
  util_model_provider=ModelProvider.OPENROUTER.name,
991
  util_model_name="openai/gpt-4.1-nano",
992
+ util_model_api_base="",
993
  util_model_ctx_length=100000,
994
  util_model_ctx_input=0.7,
995
  util_model_kwargs={"temperature": "0"},
 
998
  util_model_rl_output=0,
999
  embed_model_provider=ModelProvider.HUGGINGFACE.name,
1000
  embed_model_name="sentence-transformers/all-MiniLM-L6-v2",
1001
+ embed_model_api_base="",
1002
  embed_model_kwargs={},
1003
  embed_model_rl_requests=0,
1004
  embed_model_rl_input=0,
1005
  browser_model_provider=ModelProvider.OPENROUTER.name,
1006
  browser_model_name="openai/gpt-4.1",
1007
+ browser_model_api_base="",
1008
  browser_model_vision=True,
1009
  browser_model_kwargs={"temperature": "0"},
1010
  api_keys={},
python/tools/browser_agent.py CHANGED
@@ -57,7 +57,13 @@ class State:
57
  viewport={"width": 1024, "height": 2048},
58
  args=["--headless=new"],
59
  # Use a unique user data directory to avoid conflicts
60
- user_data_dir=str(Path.home() / ".config" / "browseruse" / "profiles" / f"agent_{self.agent.context.id}"),
 
 
 
 
 
 
61
  )
62
  )
63
 
@@ -119,11 +125,10 @@ class State:
119
  )
120
  return result
121
 
122
-
123
  model = models.get_browser_model(
124
  provider=self.agent.config.browser_model.provider,
125
  name=self.agent.config.browser_model.name,
126
- **self.agent.config.browser_model.kwargs,
127
  )
128
 
129
  try:
@@ -140,7 +145,9 @@ class State:
140
  # available_file_paths=[],
141
  )
142
  except Exception as e:
143
- raise Exception(f"Browser agent initialization failed. This might be due to model compatibility issues. Error: {e}") from e
 
 
144
 
145
  self.iter_no = get_iter_no(self.agent)
146
 
@@ -298,13 +305,17 @@ class BrowserAgent(Tool):
298
  f"Task reached step limit without completion. Last page: {current_url}. "
299
  f"The browser agent may need clearer instructions on when to finish."
300
  )
301
-
302
  # update the log (without screenshot path here, user can click)
303
  self.log.update(answer=answer_text)
304
 
305
  # add screenshot to the answer if we have it
306
- if self.log.kvps and "screenshot" in self.log.kvps and self.log.kvps['screenshot']:
307
- path = self.log.kvps['screenshot'].split('//', 1)[-1].split('&', 1)[0]
 
 
 
 
308
  answer_text += f"\n\nScreenshot: {path}"
309
 
310
  # respond (with screenshot path)
@@ -416,7 +427,9 @@ def get_use_agent_log(use_agent: browser_use.Agent | None):
416
  if item.success:
417
  short_log.append(f"✅ Done")
418
  else:
419
- short_log.append(f"❌ Error: {item.error or item.extracted_content or 'Unknown error'}")
 
 
420
 
421
  # progress messages
422
  else:
 
57
  viewport={"width": 1024, "height": 2048},
58
  args=["--headless=new"],
59
  # Use a unique user data directory to avoid conflicts
60
+ user_data_dir=str(
61
+ Path.home()
62
+ / ".config"
63
+ / "browseruse"
64
+ / "profiles"
65
+ / f"agent_{self.agent.context.id}"
66
+ ),
67
  )
68
  )
69
 
 
125
  )
126
  return result
127
 
 
128
  model = models.get_browser_model(
129
  provider=self.agent.config.browser_model.provider,
130
  name=self.agent.config.browser_model.name,
131
+ **self.agent._get_model_kwargs(self.agent.config.browser_model),
132
  )
133
 
134
  try:
 
145
  # available_file_paths=[],
146
  )
147
  except Exception as e:
148
+ raise Exception(
149
+ f"Browser agent initialization failed. This might be due to model compatibility issues. Error: {e}"
150
+ ) from e
151
 
152
  self.iter_no = get_iter_no(self.agent)
153
 
 
305
  f"Task reached step limit without completion. Last page: {current_url}. "
306
  f"The browser agent may need clearer instructions on when to finish."
307
  )
308
+
309
  # update the log (without screenshot path here, user can click)
310
  self.log.update(answer=answer_text)
311
 
312
  # add screenshot to the answer if we have it
313
+ if (
314
+ self.log.kvps
315
+ and "screenshot" in self.log.kvps
316
+ and self.log.kvps["screenshot"]
317
+ ):
318
+ path = self.log.kvps["screenshot"].split("//", 1)[-1].split("&", 1)[0]
319
  answer_text += f"\n\nScreenshot: {path}"
320
 
321
  # respond (with screenshot path)
 
427
  if item.success:
428
  short_log.append(f"✅ Done")
429
  else:
430
+ short_log.append(
431
+ f"❌ Error: {item.error or item.extracted_content or 'Unknown error'}"
432
+ )
433
 
434
  # progress messages
435
  else: