diff --git a/litellm/llms/OpenAI/openai.py b/litellm/llms/OpenAI/openai.py index 012b641ef56d..9f7b731bc5db 100644 --- a/litellm/llms/OpenAI/openai.py +++ b/litellm/llms/OpenAI/openai.py @@ -1466,7 +1466,9 @@ async def ahealth_check( elif mode == "audio_transcription": # Get the current directory of the file being run pwd = os.path.dirname(os.path.realpath(__file__)) - file_path = os.path.join(pwd, "../tests/gettysburg.wav") + file_path = os.path.join( + pwd, "../../../tests/gettysburg.wav" + ) # proxy address audio_file = open(file_path, "rb") completion = await client.audio.transcriptions.with_raw_response.create( file=audio_file, diff --git a/litellm/proxy/_new_secret_config.yaml b/litellm/proxy/_new_secret_config.yaml index b8d0aadf8ef0..b9dae49f1411 100644 --- a/litellm/proxy/_new_secret_config.yaml +++ b/litellm/proxy/_new_secret_config.yaml @@ -1,61 +1,7 @@ model_list: - - model_name: fake-claude-endpoint + - model_name: whisper litellm_params: - model: anthropic.claude-3-sonnet-20240229-v1:0 - api_base: https://exampleopenaiendpoint-production.up.railway.app - aws_secret_access_key: os.environ/AWS_SECRET_ACCESS_KEY - aws_access_key_id: os.environ/AWS_ACCESS_KEY_ID - - model_name: gemini-vision - litellm_params: - model: vertex_ai/gemini-1.0-pro-vision-001 - api_base: https://exampleopenaiendpoint-production.up.railway.app/v1/projects/adroit-crow-413218/locations/us-central1/publishers/google/models/gemini-1.0-pro-vision-001 - vertex_project: "adroit-crow-413218" - vertex_location: "us-central1" - - model_name: fake-azure-endpoint - litellm_params: - model: openai/429 - api_key: fake-key - api_base: https://exampleopenaiendpoint-production.up.railway.app - - model_name: fake-openai-endpoint - litellm_params: - model: gpt-3.5-turbo - api_base: https://exampleopenaiendpoint-production.up.railway.app - - model_name: o1-preview - litellm_params: - model: o1-preview - - model_name: rerank-english-v3.0 - litellm_params: - model: cohere/rerank-english-v3.0 - api_key: os.environ/COHERE_API_KEY - - model_name: azure-rerank-english-v3.0 - litellm_params: - model: azure_ai/rerank-english-v3.0 - api_base: os.environ/AZURE_AI_COHERE_API_BASE - api_key: os.environ/AZURE_AI_COHERE_API_KEY - - model_name: "databricks/*" - litellm_params: - model: "databricks/*" - api_key: os.environ/DATABRICKS_API_KEY - api_base: os.environ/DATABRICKS_API_BASE - - model_name: "anthropic/*" - litellm_params: - model: "anthropic/*" - - model_name: "*" - litellm_params: - model: "openai/*" - - model_name: "fireworks_ai/*" - litellm_params: - model: "fireworks_ai/*" - configurable_clientside_auth_params: ["api_base"] - - model_name: "gemini-flash-experimental" - litellm_params: - model: "vertex_ai/gemini-flash-experimental" - -litellm_settings: - json_logs: true - cache: true - cache_params: - type: "redis" - # namespace: "litellm_caching" - ttl: 900 - callbacks: ["batch_redis_requests"] + model: whisper-1 + api_key: os.environ/OPENAI_API_KEY + model_info: + mode: audio_transcription \ No newline at end of file diff --git a/litellm/proxy/proxy_cli.py b/litellm/proxy/proxy_cli.py index 9de5ad7f2d0a..9f889d2a232f 100644 --- a/litellm/proxy/proxy_cli.py +++ b/litellm/proxy/proxy_cli.py @@ -673,6 +673,9 @@ def _make_openai_completion(): import litellm + # DO NOT DELETE - enables global variables to work across files + from litellm.proxy.proxy_server import app # noqa + if run_gunicorn is False and run_hypercorn is False: if ssl_certfile_path is not None and ssl_keyfile_path is not None: print( # noqa diff --git a/tests/local_testing/test_bedrock_completion.py b/tests/local_testing/test_bedrock_completion.py index 64934d38107b..68b4232cef11 100644 --- a/tests/local_testing/test_bedrock_completion.py +++ b/tests/local_testing/test_bedrock_completion.py @@ -1229,7 +1229,7 @@ def test_not_found_error(): @pytest.mark.parametrize( "model", [ - # "bedrock/us.anthropic.claude-3-haiku-20240307-v1:0", + "bedrock/us.anthropic.claude-3-haiku-20240307-v1:0", "bedrock/us.meta.llama3-2-11b-instruct-v1:0", ], )