By using AWS re:Post, you agree to the AWS re:Post Terms of Use

Issue with Bedrock- Claude Sonnet 3.5

0

I was checking out Claude Sonnet 3.5 using bedrock. But I seem to get unfinished answers from Sonnet 3.5 It states the end reason as max token limit or something similar.

Why so and also I am facing a lot of throttling issues with this model only. All other models are working fine. Any ideas on how to fix this ?

asked 5 months ago3.4K views
2 Answers
1

Input context window is 200,000 tokens but as you are generating tokens to answer your question you are most likely hitting output token limits which are 4096 tokens see Anthropic user guide for more details.

You should ensure the model is giving you the most concise answer possible e.g. use prompt engineering to minimize preamble.

You would need to give more context on why you are being throttled. For Claude 3.5 Sonnet on Bedrock if you are using on-demand provisioning you have both a requests per minute and tokens per minute limit you need to remain under. More details on Bedrock quota limits available in Quotas for Bedrock documentation.

If you are being throttled the best course of action will depend on your use case. Provisioned throughput would offer dedicated capacity or depending on your use case a batch process may make sense.

AWS
EXPERT
answered 5 months ago
0

In AI models, a token is a chunk of text that the model reads at a time. For Claude Sonnet 3.5, the context window is 200,000 tokens, which is equivalent to approximately 150,000 words or 300 pages of text. If your input exceeds this limit, the model might return incomplete answers or errors indicating that the max token limit has been reached.

https://claude101.com/claude-3-5-sonnet/

profile picture
EXPERT
answered 5 months ago

You are not logged in. Log in to post an answer.

A good answer clearly answers the question and provides constructive feedback and encourages professional growth in the question asker.

Guidelines for Answering Questions