When I code with 4.5, it is better at avoiding getting caught in loops than previous versions (but it still happens sometimes, especially when tool integrations seem to be failing).
A few guesses:
I wonder if being trained to understand the size of the context window gives the model an impetus to move on from repetitive output to preserve that limited resource.
I could imagine there being a compacted context containing something like “the two participants exchange a series of repetitive messages about …”, and the model expects that continuing the same repetitive behavior without
In their developer docs for 4.5, Anthropic says that the model compacts the results to stay within the context window: https://www.anthropic.com/news/context-management
When I code with 4.5, it is better at avoiding getting caught in loops than previous versions (but it still happens sometimes, especially when tool integrations seem to be failing).
A few guesses:
- I wonder if being trained to understand the size of the context window gives the model an impetus to move on from repetitive output to preserve that limited resource.
- I could imagine there being a compacted context containing something like “the two participants exchange a series of repetitive messages about …”, and the model expects that continuing the same repetitive behavior without
... (read more)