| | |
| | |
| |
|
| | import sys, os |
| | import traceback |
| | import pytest |
| | sys.path.insert( |
| | 0, os.path.abspath("../..") |
| | ) |
| | import litellm |
| |
|
| | def test_update_model_cost(): |
| | try: |
| | litellm.register_model({ |
| | "gpt-4": { |
| | "max_tokens": 8192, |
| | "input_cost_per_token": 0.00002, |
| | "output_cost_per_token": 0.00006, |
| | "litellm_provider": "openai", |
| | "mode": "chat" |
| | }, |
| | }) |
| | assert litellm.model_cost["gpt-4"]["input_cost_per_token"] == 0.00002 |
| | except Exception as e: |
| | pytest.fail(f"An error occurred: {e}") |
| |
|
| | |
| |
|
| | def test_update_model_cost_map_url(): |
| | try: |
| | litellm.register_model(model_cost="https://raw.githubusercontent.com/BerriAI/litellm/main/model_prices_and_context_window.json") |
| | assert litellm.model_cost["gpt-4"]["input_cost_per_token"] == 0.00003 |
| | except Exception as e: |
| | pytest.fail(f"An error occurred: {e}") |
| |
|
| | |
| |
|
| | def test_update_model_cost_via_completion(): |
| | try: |
| | response = litellm.completion(model="gpt-3.5-turbo", messages=[{"role": "user", "content": "Hey, how's it going?"}], input_cost_per_token=0.3, output_cost_per_token=0.4) |
| | print(f"litellm.model_cost for gpt-3.5-turbo: {litellm.model_cost['gpt-3.5-turbo']}") |
| | assert litellm.model_cost["gpt-3.5-turbo"]["input_cost_per_token"] == 0.3 |
| | assert litellm.model_cost["gpt-3.5-turbo"]["output_cost_per_token"] == 0.4 |
| | except Exception as e: |
| | pytest.fail(f"An error occurred: {e}") |
| |
|
| | test_update_model_cost_via_completion() |