As the error says, you must provide the ID of an inference profile and not the model for this particular model. The easiest way to do this is to provide the ID of a system-defined inference profile for this model. You can find it by invoking this awscli command with the correct credentials defined in the environment (or set via standard flags):

aws bedrock list-inference-profiles

You will see this one in the JSON list:

{
  "inferenceProfileName": "US Anthropic Claude 3.5 Sonnet v2",
  "description": "Routes requests to Anthropic Claude 3.5 Sonnet v2 in us-west-2, us-east-1 and us-east-2.",
  "inferenceProfileArn": "arn:aws:bedrock:us-east-1:381492273274:inference-profile/us.anthropic.claude-3-5-sonnet-20241022-v2:0",
  "models": [
    {
      "modelArn": "arn:aws:bedrock:us-west-2::foundation-model/anthropic.claude-3-5-sonnet-20241022-v2:0"
    },
    {
      "modelArn": "arn:aws:bedrock:us-east-1::foundation-model/anthropic.claude-3-5-sonnet-20241022-v2:0"
    },
    {
      "modelArn": "arn:aws:bedrock:us-east-2::foundation-model/anthropic.claude-3-5-sonnet-20241022-v2:0"
    }
  ],
  "inferenceProfileId": "us.anthropic.claude-3-5-sonnet-20241022-v2:0",
  "status": "ACTIVE",
  "type": "SYSTEM_DEFINED"
}

Modify the invoke_model line in your code to specify the ID or ARN of the inference profile instead:

response = bedrock_runtime.invoke_model(
  body=body,
  modelId="us.anthropic.claude-3-5-sonnet-20241022-v2:0",
)
Answer from Asfand Qazi on Stack Overflow
🌐
AWS
aws.amazon.com › blogs › aws › upgraded-claude-3-5-sonnet-from-anthropic-available-now-computer-use-public-beta-and-claude-3-5-haiku-coming-soon-in-amazon-bedrock
Announcing three new capabilities for the Claude 3.5 model family in Amazon Bedrock | Amazon Web Services
November 4, 2024 - Back in the Amazon Bedrock console, I choose Chat/text under Playgrounds in the navigation pane. For the model, I select Anthropic as the model provider and then Claude 3.5 Sonnet V2.
🌐
GitHub
github.com › boto › boto3 › issues › 4335
Bedrock Claude 3.5 Sonnet v2 is not supporting new attachments (PDF) · Issue #4335 · boto/boto3
October 3, 2024 - Describe the bug Recently Anthropic has released a new version of Claude 3.5 Sonnet v2 and it's also available via Amazon Bedrock. Breaking changes in it includes the support of PDF attachments (earlier it was limited to images). This is...
Published   Nov 06, 2024
🌐
Amazon Web Services
aws.amazon.com › generative ai › amazon bedrock › anthropic
Claude by Anthropic - Models in Amazon Bedrock – AWS
1 week ago - Hear from Neerav Kingsland, Head of Global Accounts at Anthropic, as he discusses what Claude’s availability in Amazon Bedrock can provide for businesses around the world. ... By default, Anthropic's Claude models have a 200,000 token context window enabling you to relay a large volume of information to Claude. This translates to roughly 150,000 words, or over 500 pages of material. Sonnet 4 and 4.5 now offers an expanded context length of 1 million tokens (preview).
🌐
GitHub
github.com › Aider-AI › aider › issues › 2120
Claude Sonnet 3.5 V2 not working with AWS Bedrock · Issue #2120 · Aider-AI/aider
September 21, 2024 - Model: anthropic.claude-3-5-sonnet-20241022-v2:0 with diff edit format, infinite output Git repo: .git with 111 files Repo-map: using 1024 tokens, auto refresh VSCode terminal detected, pretty output has been disabled.
Published   Oct 22, 2024
Top answer
1 of 3
11

As the error says, you must provide the ID of an inference profile and not the model for this particular model. The easiest way to do this is to provide the ID of a system-defined inference profile for this model. You can find it by invoking this awscli command with the correct credentials defined in the environment (or set via standard flags):

aws bedrock list-inference-profiles

You will see this one in the JSON list:

{
  "inferenceProfileName": "US Anthropic Claude 3.5 Sonnet v2",
  "description": "Routes requests to Anthropic Claude 3.5 Sonnet v2 in us-west-2, us-east-1 and us-east-2.",
  "inferenceProfileArn": "arn:aws:bedrock:us-east-1:381492273274:inference-profile/us.anthropic.claude-3-5-sonnet-20241022-v2:0",
  "models": [
    {
      "modelArn": "arn:aws:bedrock:us-west-2::foundation-model/anthropic.claude-3-5-sonnet-20241022-v2:0"
    },
    {
      "modelArn": "arn:aws:bedrock:us-east-1::foundation-model/anthropic.claude-3-5-sonnet-20241022-v2:0"
    },
    {
      "modelArn": "arn:aws:bedrock:us-east-2::foundation-model/anthropic.claude-3-5-sonnet-20241022-v2:0"
    }
  ],
  "inferenceProfileId": "us.anthropic.claude-3-5-sonnet-20241022-v2:0",
  "status": "ACTIVE",
  "type": "SYSTEM_DEFINED"
}

Modify the invoke_model line in your code to specify the ID or ARN of the inference profile instead:

response = bedrock_runtime.invoke_model(
  body=body,
  modelId="us.anthropic.claude-3-5-sonnet-20241022-v2:0",
)
2 of 3
0

You can add the ARN of an Inference profile to your modelID itself while invoking the model.

response = bedrock_client.invoke_model(

       modelId="arn:aws:bedrock:us-east-1::model/your-bedrock-model-arn"  

       prompt="Your prompt here"

   )
🌐
Reddit
reddit.com › r/aws › bedrock anthropic sonnet claude 3.5 v2 forces you to region us-west-2
r/aws on Reddit: Bedrock Anthropic Sonnet Claude 3.5 v2 forces you to region us-west-2
December 5, 2024 -

Hey, i have been using `http://anthropic.claude-3-5-sonnet-20240620-v1:0/` with great success from us-east-1. Today, I tried to pivot to v2 which has an inference profile ID of `us.anthropic.claude-3-5-sonnet-20241022-v2:0`. The model seems to be available in both us-east-1 and in us-west-2.

I am getting denied with a error code of 403 (not authorised) to access the model. Which doesn't make sense as I have access to the model in us-east-1. However, I noticed that the API tries to access the model from us-west-2, `arn:aws:bedrock:us-west-2::foundation-model/anthropic.claude-3-5-sonnet-20241022-v2:0` even though I am manually selecting aws_region of "us-east-1".

The same problem when I am using the Inference profile ARN (arn:aws:bedrock:us-east-1:422697756335:inference-profile/us.anthropic.claude-3-5-sonnet-20241022-v2:0). Any ideas why the API seems to force a us-west-2 access?

By the way, I don't have access to us-west-2

Find elsewhere
🌐
Anthropic
anthropic.com › news › claude-3-5-sonnet
Introducing Claude 3.5 Sonnet
Claude 3.5 Sonnet is now available for free on Claude.ai and the Claude iOS app, while Claude Pro and Team plan subscribers can access it with significantly higher rate limits. It is also available via the Anthropic API, Amazon Bedrock, and Google Cloud’s Vertex AI.
🌐
Anthropic
anthropic.com › claude › sonnet
Claude Sonnet 4.5
Sonnet 3.7 is the first hybrid reasoning model and our most intelligent model to date. It’s state-of-the art for coding and delivers significant improvements in content generation, data analysis, and planning. ... Anyone can chat with Claude using Sonnet 4.5 on Claude.ai, available on web, iOS, and Android. For developers interested in building agents, Sonnet 4.5 is available on the Claude Developer Platform natively, and in Amazon Bedrock, Google Cloud's Vertex AI, and Microsoft Foundry.
🌐
Amazon
aboutamazon.com › news › aws › amazon-bedrock-anthropic-ai-claude-3-5-sonnet
Amazon Bedrock introduces Claude 3.5 Haiku and an upgraded Claude 3.5 Sonnet, Anthropic’s most intelligent AI models to date
November 4, 2024 - Claude 3.5 Haiku and Claude 3.5 Sonnet, the latest and most intelligent models in the Claude model family from artificial intelligence (AI) safety and research company Anthropic, are now generally available in Amazon Bedrock.
🌐
AWS re:Post
repost.aws › questions › QUoB4cqmFgRG2-EY_KW_Ghdg › claude-3-5-sonnet-v2-on-demand-invokemodel-requests-per-minute-quota-increase
Claude 3.5 Sonnet V2 On-demand InvokeModel requests per minute quota increase | AWS re:Post
December 4, 2024 - Consider optimizing your application to work within the current quota limits, such as batching requests or implementing client-side rate limiting to stay within the 50 requests per minute limit. Sources AWS Bedrock Claude 3.5 Sonnet throttled randomly "Too many tokens, please wait before trying again." | AWS re:Post Understand why there are service quotas on my account | AWS re:Post Submit quota increase | AWS re:Post
🌐
GitHub
github.com › posit-dev › positron › issues › 9339
Should remove Claude 3.5 Sonnet v1 and v2 Bedrock models · Issue #9339 · posit-dev/positron
September 8, 2025 - I got an email from AWS that says that they're sunsetting Claude 3.5 Sonnet v1 in Bedrock, so we should remove it. (v2 is still available.) We are contacting you because you have used model ID 'ant...
Published   Sep 08, 2025
🌐
AWS re:Post
repost.aws › questions › QUdUN4vu-FT4y4u_QnrhGB_g › which-claude-sonnet-model-is-better-3-5-v1-or-3-5-v2-features-cost-comparison
Which Claude Sonnet Model is Better: 3.5 v1 or 3.5 v2? (Features & Cost Comparison) | AWS re:Post
November 13, 2024 - Sources Upgraded Claude 3.5 Sonnet ... Claude 3.5 Sonnet model and computer use now in Amazon Bedrock - AWS ... According to leaderboards v2 is better....
🌐
Reddit
reddit.com › r/aws › aws bedrock - claude 3.5 sonnet update
r/aws on Reddit: AWS Bedrock - Claude 3.5 Sonnet Update
March 15, 2024 -

Hi,

I have been utilizing AWS Bedrock hosted version of Claude 3.5 extensively for my clients and the convenience and ease of use is unmatched.

However, there seems to be an issue with Amazon or Anthropic updating the model to the latest version meaning last month (July) Anthropic updated Sonnet to have an output token length of 8192 tokens, but it requires extra headers to be passed (when using their native API).

I tried it with Bedrock, but to my surprise I realized Bedrock models are hardly being updated from the first time they are launched and it still doesn't have the latest update of the models which is a but frustrating as I already have a use case for the 8000 context window and the client doesn't want to use LLM outside of Bedrock.

Does anyone know if AWS or Anthropic will update the model to July version?

X discussion (for native Anthropic API): https://x.com/alexalbert__/status/1812921642143900036?t=NH_I5sEF2Z7UjQecleEaug&s=19

🌐
AWS Builder Center
builder.aws.com › content › 2uun3vJOQHdwN9l4ZEqbkbnyUFG › claude-35-sonnet-v2-double-output-tokens-on-aws-bedrock
Claude 3.5 Sonnet v2: Double Output Tokens on ...
Connect with builders who understand your journey. Share solutions, influence AWS product development, and access useful content that accelerates your growth. Your community starts here.
🌐
IOD
iamondemand.com › home › implementing claude 3.5 sonnet on aws: a practical guide – part 2
Implementing Claude 3.5 Sonnet on AWS, Part 2: A Practical Guide
January 7, 2025 - You can either use the UI or interact programmatically via the Amazon Bedrock Converse API or the Anthropic Claude Messages API using any AWS SDK. For this guide, I’ll be using the Anthropic Claude Messages API through the AWS SDK because it offers greater flexibility for integration into existing systems and workflows. Note: Different AWS regions may offer different models. To get the best performance out of Claude 3.5 Sonnet, I use the US-East region.