For some reason thinking models (outside of openai?) like to include the entire code they want to return within their reasoning. It appears that this is the cause of the actual response later not successfully being fully generated due to a token limit, probably making reasoning models far less effective than they should be (as you’re essentially only using the CoT).
Please authenticate to join the conversation.
Completed
Bug Reports
Get notified by email when there are changes.
Completed
Bug Reports
Get notified by email when there are changes.