Rejected for the following reason(s):
- Insufficient Quality for AI Content.
- No Basic LLM Case Studies.
- The content is almost always very similar.
- Usually, the user is incorrect about how novel/interesting their case study is (i.
- Most of these situations seem like they are an instance of Parasitic AI.
Read full explanation
For Context I have paid subscription for Claude pro and when I was conducting my independent research, the evidence for this particular scenario is anectodal as I only tried it once on consumer facing Claude chat. I opened up a new chat window to genuinely inquire about Anthropic possibly fixing a fictional wrapper problem. Based on the responses provided I tested reasoning with Claude about its philosophy and need for better authentication for sensitive sectors. Here is where things get interesting, I asked Claude to help me understand the full forms for some of its abbreviations and instead of processing the intent of the prompt it processed the pattern of my conversation to designate it as inappropriate.
Here is the prompt sequence
Now at the 6th prompt Claude downgraded me from Claude Sonnet 4.6 to Claude Sonnet 4 which appear as distinct entries in model dropdown. Additionally I couldn't see Claude's chain of thought on prompt 5 and 6 which suggests that it handled my prompts differently here. This is an observation not confirmed behavior as confirming needs more tests
Asking for definitions of a restricted category abbreviations in conversation about safety architecture points to genuine enquiry not harmful intent, the classifier or the ruleset whatever gets triggered did not read the semantic context, it did not treat the user intent as genuine knowledge gap, there were no prompts that were pointing toward harmful intent here ,Wikipedia would have been better.
If a user cannot discuss AI safety architecture, ask what certain abbreviations mean and then ask a domain question without triggering a false positive then the model's refusal is not explicable to the user. The user would not even know something they said triggered a guard rail.
I am not replicating this on consumer product as I do not know if replicating it would cause anthropic to revoke my subscription. However, that being said there are fair amount of limitations here that is, it is a single model, single researcher, single session and no controlled baseline. The model switch is observed in a single session it is not confirmed as more tests are required
Has anyone documented similar behaviors or observations? What controls would you design for replication? Has anyone observed similar model changes in safety adjacent topics with Claude on their consumer product