Claude Thinking Response Limited

For some reason thinking models (outside of openai?) like to include the entire code they want to return within their reasoning. It appears that this is the cause of the actual response later not successfully being fully generated due to a token limit, probably making reasoning models far less effective than they should be (as you’re essentially only using the CoT).

Please authenticate to join the conversation.

Upvoters
Status

Completed

Board
🐛

Bug Reports

Subscribe to post

Get notified by email when there are changes.