diff --git a/.stats.yml b/.stats.yml index 64f8716..3ea9b07 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,2 +1,2 @@ configured_endpoints: 21 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/anthropic-f5276eeef7512112e802c85530c51e0a971ee521eebe3a0db309621587b4973d.yml +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/anthropic-946022163d75ddc00f49883756b189412098aa07b600ee0890655eebcb3c440c.yml diff --git a/src/anthropic/types/model.py b/src/anthropic/types/model.py index 7115777..6fcafc5 100644 --- a/src/anthropic/types/model.py +++ b/src/anthropic/types/model.py @@ -6,7 +6,6 @@ __all__ = ["Model"] Model: TypeAlias = Union[ - str, Literal[ "claude-3-5-haiku-latest", "claude-3-5-haiku-20241022", @@ -20,4 +19,5 @@ "claude-2.1", "claude-2.0", ], + str, ] diff --git a/src/anthropic/types/model_param.py b/src/anthropic/types/model_param.py index 5d4a156..2698256 100644 --- a/src/anthropic/types/model_param.py +++ b/src/anthropic/types/model_param.py @@ -8,7 +8,6 @@ __all__ = ["ModelParam"] ModelParam: TypeAlias = Union[ - str, Literal[ "claude-3-5-haiku-latest", "claude-3-5-haiku-20241022", @@ -22,4 +21,5 @@ "claude-2.1", "claude-2.0", ], + str, ] diff --git a/tests/api_resources/beta/messages/test_batches.py b/tests/api_resources/beta/messages/test_batches.py index 6e532e9..d07a72a 100644 --- a/tests/api_resources/beta/messages/test_batches.py +++ b/tests/api_resources/beta/messages/test_batches.py @@ -37,7 +37,7 @@ def test_method_create(self, client: Anthropic) -> None: "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], @@ -58,7 +58,7 @@ def test_method_create_with_all_params(self, client: Anthropic) -> None: "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", "metadata": {"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, "stop_sequences": ["string"], "stream": True, @@ -128,7 +128,7 @@ def test_raw_response_create(self, client: Anthropic) -> None: "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], @@ -153,7 +153,7 @@ def test_streaming_response_create(self, client: Anthropic) -> None: "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], @@ -408,7 +408,7 @@ async def test_method_create(self, async_client: AsyncAnthropic) -> None: "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], @@ -429,7 +429,7 @@ async def test_method_create_with_all_params(self, async_client: AsyncAnthropic) "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", "metadata": {"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, "stop_sequences": ["string"], "stream": True, @@ -499,7 +499,7 @@ async def test_raw_response_create(self, async_client: AsyncAnthropic) -> None: "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], @@ -524,7 +524,7 @@ async def test_streaming_response_create(self, async_client: AsyncAnthropic) -> "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], diff --git a/tests/api_resources/beta/test_messages.py b/tests/api_resources/beta/test_messages.py index 1d6fae2..0377351 100644 --- a/tests/api_resources/beta/test_messages.py +++ b/tests/api_resources/beta/test_messages.py @@ -30,7 +30,7 @@ def test_method_create_overload_1(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) assert_matches_type(BetaMessage, message, path=["response"]) @@ -44,7 +44,7 @@ def test_method_create_with_all_params_overload_1(self, client: Anthropic) -> No "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, stop_sequences=["string"], stream=False, @@ -107,7 +107,7 @@ def test_raw_response_create_overload_1(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) assert response.is_closed is True @@ -125,7 +125,7 @@ def test_streaming_response_create_overload_1(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) as response: assert not response.is_closed assert response.http_request.headers.get("X-Stainless-Lang") == "python" @@ -145,7 +145,7 @@ def test_method_create_overload_2(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) message_stream.response.close() @@ -160,7 +160,7 @@ def test_method_create_with_all_params_overload_2(self, client: Anthropic) -> No "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, stop_sequences=["string"], @@ -223,7 +223,7 @@ def test_raw_response_create_overload_2(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) @@ -241,7 +241,7 @@ def test_streaming_response_create_overload_2(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) as response: assert not response.is_closed @@ -261,7 +261,7 @@ def test_method_count_tokens(self, client: Anthropic) -> None: "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) assert_matches_type(BetaMessageTokensCount, message, path=["response"]) @@ -274,7 +274,7 @@ def test_method_count_tokens_with_all_params(self, client: Anthropic) -> None: "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", system=[ { "text": "Today's date is 2024-06-01.", @@ -330,7 +330,7 @@ def test_raw_response_count_tokens(self, client: Anthropic) -> None: "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) assert response.is_closed is True @@ -347,7 +347,7 @@ def test_streaming_response_count_tokens(self, client: Anthropic) -> None: "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) as response: assert not response.is_closed assert response.http_request.headers.get("X-Stainless-Lang") == "python" @@ -371,7 +371,7 @@ async def test_method_create_overload_1(self, async_client: AsyncAnthropic) -> N "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) assert_matches_type(BetaMessage, message, path=["response"]) @@ -385,7 +385,7 @@ async def test_method_create_with_all_params_overload_1(self, async_client: Asyn "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, stop_sequences=["string"], stream=False, @@ -448,7 +448,7 @@ async def test_raw_response_create_overload_1(self, async_client: AsyncAnthropic "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) assert response.is_closed is True @@ -466,7 +466,7 @@ async def test_streaming_response_create_overload_1(self, async_client: AsyncAnt "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) as response: assert not response.is_closed assert response.http_request.headers.get("X-Stainless-Lang") == "python" @@ -486,7 +486,7 @@ async def test_method_create_overload_2(self, async_client: AsyncAnthropic) -> N "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) await message_stream.response.aclose() @@ -501,7 +501,7 @@ async def test_method_create_with_all_params_overload_2(self, async_client: Asyn "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, stop_sequences=["string"], @@ -564,7 +564,7 @@ async def test_raw_response_create_overload_2(self, async_client: AsyncAnthropic "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) @@ -582,7 +582,7 @@ async def test_streaming_response_create_overload_2(self, async_client: AsyncAnt "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) as response: assert not response.is_closed @@ -602,7 +602,7 @@ async def test_method_count_tokens(self, async_client: AsyncAnthropic) -> None: "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) assert_matches_type(BetaMessageTokensCount, message, path=["response"]) @@ -615,7 +615,7 @@ async def test_method_count_tokens_with_all_params(self, async_client: AsyncAnth "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", system=[ { "text": "Today's date is 2024-06-01.", @@ -671,7 +671,7 @@ async def test_raw_response_count_tokens(self, async_client: AsyncAnthropic) -> "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) assert response.is_closed is True @@ -688,7 +688,7 @@ async def test_streaming_response_count_tokens(self, async_client: AsyncAnthropi "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) as response: assert not response.is_closed assert response.http_request.headers.get("X-Stainless-Lang") == "python" diff --git a/tests/api_resources/messages/test_batches.py b/tests/api_resources/messages/test_batches.py index 079be16..e89f02d 100644 --- a/tests/api_resources/messages/test_batches.py +++ b/tests/api_resources/messages/test_batches.py @@ -37,7 +37,7 @@ def test_method_create(self, client: Anthropic) -> None: "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], @@ -58,7 +58,7 @@ def test_raw_response_create(self, client: Anthropic) -> None: "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], @@ -83,7 +83,7 @@ def test_streaming_response_create(self, client: Anthropic) -> None: "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], @@ -304,7 +304,7 @@ async def test_method_create(self, async_client: AsyncAnthropic) -> None: "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], @@ -325,7 +325,7 @@ async def test_raw_response_create(self, async_client: AsyncAnthropic) -> None: "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], @@ -350,7 +350,7 @@ async def test_streaming_response_create(self, async_client: AsyncAnthropic) -> "role": "user", } ], - "model": "claude-3-5-sonnet-20241022", + "model": "claude-3-5-haiku-latest", }, } ], diff --git a/tests/api_resources/test_completions.py b/tests/api_resources/test_completions.py index 24374a1..828900d 100644 --- a/tests/api_resources/test_completions.py +++ b/tests/api_resources/test_completions.py @@ -21,7 +21,7 @@ class TestCompletions: def test_method_create_overload_1(self, client: Anthropic) -> None: completion = client.completions.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", ) assert_matches_type(Completion, completion, path=["response"]) @@ -30,7 +30,7 @@ def test_method_create_overload_1(self, client: Anthropic) -> None: def test_method_create_with_all_params_overload_1(self, client: Anthropic) -> None: completion = client.completions.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, stop_sequences=["string"], @@ -45,7 +45,7 @@ def test_method_create_with_all_params_overload_1(self, client: Anthropic) -> No def test_raw_response_create_overload_1(self, client: Anthropic) -> None: response = client.completions.with_raw_response.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", ) @@ -58,7 +58,7 @@ def test_raw_response_create_overload_1(self, client: Anthropic) -> None: def test_streaming_response_create_overload_1(self, client: Anthropic) -> None: with client.completions.with_streaming_response.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", ) as response: assert not response.is_closed @@ -73,7 +73,7 @@ def test_streaming_response_create_overload_1(self, client: Anthropic) -> None: def test_method_create_overload_2(self, client: Anthropic) -> None: completion_stream = client.completions.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", stream=True, ) @@ -83,7 +83,7 @@ def test_method_create_overload_2(self, client: Anthropic) -> None: def test_method_create_with_all_params_overload_2(self, client: Anthropic) -> None: completion_stream = client.completions.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", stream=True, metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, @@ -98,7 +98,7 @@ def test_method_create_with_all_params_overload_2(self, client: Anthropic) -> No def test_raw_response_create_overload_2(self, client: Anthropic) -> None: response = client.completions.with_raw_response.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", stream=True, ) @@ -111,7 +111,7 @@ def test_raw_response_create_overload_2(self, client: Anthropic) -> None: def test_streaming_response_create_overload_2(self, client: Anthropic) -> None: with client.completions.with_streaming_response.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", stream=True, ) as response: @@ -131,7 +131,7 @@ class TestAsyncCompletions: async def test_method_create_overload_1(self, async_client: AsyncAnthropic) -> None: completion = await async_client.completions.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", ) assert_matches_type(Completion, completion, path=["response"]) @@ -140,7 +140,7 @@ async def test_method_create_overload_1(self, async_client: AsyncAnthropic) -> N async def test_method_create_with_all_params_overload_1(self, async_client: AsyncAnthropic) -> None: completion = await async_client.completions.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, stop_sequences=["string"], @@ -155,7 +155,7 @@ async def test_method_create_with_all_params_overload_1(self, async_client: Asyn async def test_raw_response_create_overload_1(self, async_client: AsyncAnthropic) -> None: response = await async_client.completions.with_raw_response.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", ) @@ -168,7 +168,7 @@ async def test_raw_response_create_overload_1(self, async_client: AsyncAnthropic async def test_streaming_response_create_overload_1(self, async_client: AsyncAnthropic) -> None: async with async_client.completions.with_streaming_response.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", ) as response: assert not response.is_closed @@ -183,7 +183,7 @@ async def test_streaming_response_create_overload_1(self, async_client: AsyncAnt async def test_method_create_overload_2(self, async_client: AsyncAnthropic) -> None: completion_stream = await async_client.completions.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", stream=True, ) @@ -193,7 +193,7 @@ async def test_method_create_overload_2(self, async_client: AsyncAnthropic) -> N async def test_method_create_with_all_params_overload_2(self, async_client: AsyncAnthropic) -> None: completion_stream = await async_client.completions.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", stream=True, metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, @@ -208,7 +208,7 @@ async def test_method_create_with_all_params_overload_2(self, async_client: Asyn async def test_raw_response_create_overload_2(self, async_client: AsyncAnthropic) -> None: response = await async_client.completions.with_raw_response.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", stream=True, ) @@ -221,7 +221,7 @@ async def test_raw_response_create_overload_2(self, async_client: AsyncAnthropic async def test_streaming_response_create_overload_2(self, async_client: AsyncAnthropic) -> None: async with async_client.completions.with_streaming_response.create( max_tokens_to_sample=256, - model="string", + model="claude-3-5-haiku-latest", prompt="\n\nHuman: Hello, world!\n\nAssistant:", stream=True, ) as response: diff --git a/tests/api_resources/test_messages.py b/tests/api_resources/test_messages.py index 06eea7e..4047ba0 100644 --- a/tests/api_resources/test_messages.py +++ b/tests/api_resources/test_messages.py @@ -30,7 +30,7 @@ def test_method_create_overload_1(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) assert_matches_type(Message, message, path=["response"]) @@ -44,7 +44,7 @@ def test_method_create_with_all_params_overload_1(self, client: Anthropic) -> No "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, stop_sequences=["string"], stream=False, @@ -105,7 +105,7 @@ def test_raw_response_create_overload_1(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) assert response.is_closed is True @@ -123,7 +123,7 @@ def test_streaming_response_create_overload_1(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) as response: assert not response.is_closed assert response.http_request.headers.get("X-Stainless-Lang") == "python" @@ -143,7 +143,7 @@ def test_method_create_overload_2(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) message_stream.response.close() @@ -158,7 +158,7 @@ def test_method_create_with_all_params_overload_2(self, client: Anthropic) -> No "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, stop_sequences=["string"], @@ -219,7 +219,7 @@ def test_raw_response_create_overload_2(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) @@ -237,7 +237,7 @@ def test_streaming_response_create_overload_2(self, client: Anthropic) -> None: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) as response: assert not response.is_closed @@ -257,7 +257,7 @@ def test_method_count_tokens(self, client: Anthropic) -> None: "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) assert_matches_type(MessageTokensCount, message, path=["response"]) @@ -270,7 +270,7 @@ def test_method_count_tokens_with_all_params(self, client: Anthropic) -> None: "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", system=[ { "text": "Today's date is 2024-06-01.", @@ -324,7 +324,7 @@ def test_raw_response_count_tokens(self, client: Anthropic) -> None: "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) assert response.is_closed is True @@ -341,7 +341,7 @@ def test_streaming_response_count_tokens(self, client: Anthropic) -> None: "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) as response: assert not response.is_closed assert response.http_request.headers.get("X-Stainless-Lang") == "python" @@ -365,7 +365,7 @@ async def test_method_create_overload_1(self, async_client: AsyncAnthropic) -> N "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) assert_matches_type(Message, message, path=["response"]) @@ -379,7 +379,7 @@ async def test_method_create_with_all_params_overload_1(self, async_client: Asyn "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, stop_sequences=["string"], stream=False, @@ -440,7 +440,7 @@ async def test_raw_response_create_overload_1(self, async_client: AsyncAnthropic "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) assert response.is_closed is True @@ -458,7 +458,7 @@ async def test_streaming_response_create_overload_1(self, async_client: AsyncAnt "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) as response: assert not response.is_closed assert response.http_request.headers.get("X-Stainless-Lang") == "python" @@ -478,7 +478,7 @@ async def test_method_create_overload_2(self, async_client: AsyncAnthropic) -> N "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) await message_stream.response.aclose() @@ -493,7 +493,7 @@ async def test_method_create_with_all_params_overload_2(self, async_client: Asyn "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, metadata={"user_id": "13803d75-b4b5-4c3e-b2a2-6f21399b021b"}, stop_sequences=["string"], @@ -554,7 +554,7 @@ async def test_raw_response_create_overload_2(self, async_client: AsyncAnthropic "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) @@ -572,7 +572,7 @@ async def test_streaming_response_create_overload_2(self, async_client: AsyncAnt "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", stream=True, ) as response: assert not response.is_closed @@ -592,7 +592,7 @@ async def test_method_count_tokens(self, async_client: AsyncAnthropic) -> None: "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) assert_matches_type(MessageTokensCount, message, path=["response"]) @@ -605,7 +605,7 @@ async def test_method_count_tokens_with_all_params(self, async_client: AsyncAnth "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", system=[ { "text": "Today's date is 2024-06-01.", @@ -659,7 +659,7 @@ async def test_raw_response_count_tokens(self, async_client: AsyncAnthropic) -> "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) assert response.is_closed is True @@ -676,7 +676,7 @@ async def test_streaming_response_count_tokens(self, async_client: AsyncAnthropi "role": "user", } ], - model="string", + model="claude-3-5-haiku-latest", ) as response: assert not response.is_closed assert response.http_request.headers.get("X-Stainless-Lang") == "python" diff --git a/tests/test_client.py b/tests/test_client.py index 5c1f075..f60e2fb 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -816,7 +816,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) assert response.retries_taken == failures_before_success @@ -849,7 +849,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", extra_headers={"x-stainless-retry-count": Omit()}, ) @@ -882,7 +882,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", extra_headers={"x-stainless-retry-count": "42"}, ) @@ -915,7 +915,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) as response: assert response.retries_taken == failures_before_success assert int(response.http_request.headers.get("x-stainless-retry-count")) == failures_before_success @@ -1697,7 +1697,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) assert response.retries_taken == failures_before_success @@ -1731,7 +1731,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", extra_headers={"x-stainless-retry-count": Omit()}, ) @@ -1765,7 +1765,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", extra_headers={"x-stainless-retry-count": "42"}, ) @@ -1799,7 +1799,7 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: "role": "user", } ], - model="claude-3-5-sonnet-20241022", + model="claude-3-5-haiku-latest", ) as response: assert response.retries_taken == failures_before_success assert int(response.http_request.headers.get("x-stainless-retry-count")) == failures_before_success