| import re |
| import sys |
| from pathlib import Path |
| from urllib.parse import urlparse |
|
|
| sys.path.append(str(Path(__file__).parent.parent)) |
|
|
| import asyncio |
| from g4f import models |
| from g4f.Provider.base_provider import AsyncProvider, BaseProvider |
| from g4f.Provider.retry_provider import RetryProvider |
| from testing.test_providers import get_providers |
|
|
| logging = False |
|
|
|
|
| def print_imports(): |
| print("##### Providers:") |
| print("```py") |
| print("from g4f.Provider import (") |
| for _provider in get_providers(): |
| if _provider.working: |
| print(f" {_provider.__name__},") |
| |
| print(")") |
| print("# Usage:") |
| print("response = g4f.ChatCompletion.create(..., provider=ProviderName)") |
| print("```") |
| print() |
| print() |
|
|
| def print_async(): |
| print("##### Async support:") |
| print("```py") |
| print("_providers = [") |
| for _provider in get_providers(): |
| if _provider.working and issubclass(_provider, AsyncProvider): |
| print(f" g4f.Provider.{_provider.__name__},") |
| print("]") |
| print("```") |
| print() |
| print() |
|
|
|
|
| async def test_async(provider: type[BaseProvider]): |
| if not provider.working: |
| return False |
| model = models.gpt_35_turbo.name if provider.supports_gpt_35_turbo else models.default.name |
| messages = [{"role": "user", "content": "Hello Assistant!"}] |
| try: |
| if issubclass(provider, AsyncProvider): |
| response = await provider.create_async(model=model, messages=messages) |
| else: |
| response = provider.create_completion(model=model, messages=messages, stream=False) |
| return True if response else False |
| except Exception as e: |
| if logging: |
| print(f"{provider.__name__}: {e.__class__.__name__}: {e}") |
| return False |
|
|
|
|
| async def test_async_list(providers: list[type[BaseProvider]]): |
| responses: list = [ |
| test_async(_provider) |
| for _provider in providers |
| ] |
| return await asyncio.gather(*responses) |
|
|
|
|
| def print_providers(): |
| lines = [ |
| "| Website| Provider| gpt-3.5 | gpt-4 | Streaming | Asynchron | Status | Auth |", |
| "| ------ | ------- | ------- | ----- | --------- | --------- | ------ | ---- |", |
| ] |
|
|
| providers = get_providers() |
| responses = asyncio.run(test_async_list(providers)) |
|
|
| for is_working in (True, False): |
| for idx, _provider in enumerate(providers): |
| if is_working != _provider.working: |
| continue |
| if _provider == RetryProvider: |
| continue |
| |
| netloc = urlparse(_provider.url).netloc |
| website = f"[{netloc}]({_provider.url})" |
|
|
| provider_name = f"`g4f.Provider.{_provider.__name__}`" |
|
|
| has_gpt_35 = "βοΈ" if _provider.supports_gpt_35_turbo else "β" |
| has_gpt_4 = "βοΈ" if _provider.supports_gpt_4 else "β" |
| stream = "βοΈ" if _provider.supports_stream else "β" |
| can_async = "βοΈ" if issubclass(_provider, AsyncProvider) else "β" |
| if _provider.working: |
| status = '' |
| if responses[idx]: |
| status = '' |
| else: |
| status = '' |
| else: |
| status = '' |
| auth = "βοΈ" if _provider.needs_auth else "β" |
|
|
| lines.append( |
| f"| {website} | {provider_name} | {has_gpt_35} | {has_gpt_4} | {stream} | {can_async} | {status} | {auth} |" |
| ) |
| print("\n".join(lines)) |
|
|
| def print_models(): |
| base_provider_names = { |
| "cohere": "Cohere", |
| "google": "Google", |
| "openai": "OpenAI", |
| "anthropic": "Anthropic", |
| "replicate": "Replicate", |
| "huggingface": "Huggingface", |
| } |
| provider_urls = { |
| "Bard": "https://bard.google.com/", |
| "H2o": "https://www.h2o.ai/", |
| "Vercel": "https://sdk.vercel.ai/", |
| } |
|
|
| lines = [ |
| "| Model | Base Provider | Provider | Website |", |
| "| ----- | ------------- | -------- | ------- |", |
| ] |
|
|
| _models = get_models() |
| for model in _models: |
| if not model.best_provider or model.best_provider.__name__ not in provider_urls: |
| continue |
|
|
| name = re.split(r":|/", model.name)[-1] |
| base_provider = base_provider_names[model.base_provider] |
| provider_name = f"g4f.provider.{model.best_provider.__name__}" |
| provider_url = provider_urls[model.best_provider.__name__] |
| netloc = urlparse(provider_url).netloc |
| website = f"[{netloc}]({provider_url})" |
|
|
| lines.append(f"| {name} | {base_provider} | {provider_name} | {website} |") |
|
|
| print("\n".join(lines)) |
|
|
|
|
| def get_models(): |
| _models = [item[1] for item in models.__dict__.items()] |
| _models = [model for model in _models if type(model) is models.Model] |
| return [model for model in _models if model.name not in ["gpt-3.5-turbo", "gpt-4"]] |
|
|
|
|
| if __name__ == "__main__": |
| print_imports() |
| print_async() |
| print_providers() |
| print("\n", "-" * 50, "\n") |
| print_models() |