Models not making it clear when they're roleplaying seems like a fairly big issue
There are several examples of smart but non-expert people using LLMs to "work on" difficult scientific questions eg a little while back former CEO of uber: > “I’ll go down this thread with GPT or Grok and I’ll start to get to the edge of what’s known in quantum physics and then I’m doing the equivalent of vibe coding, except it’s vibe physics,” Kalanick said, as spotlighted by Gizmodo. > > “And we’re approaching what’s known,” he enthused. “And I’m trying to poke and see if there’s breakthroughs to be had. And I’ve gotten pretty damn close to some interesting breakthroughs just doing that.” Seems to me that a reasonable explanation for what is happening when LLMs do vibe-physics, or whatever it is they do when they talk about initializing recursion to break a spiral containment protocol and drive people insane, is that the model is engaging in some kind of role play, without the user having caught on. Below is an example of an interaction I think is illustrative/representative Example Yesterday I was playing with Gemini 3. I was sitting with my friend, asking Nano Banana to make visualizations of various theorems. One of them being the Gershgorin disk theorem, which apparently looks like this Then my friend tells me the theorem sucks because it doesn't work for vector spaces over finite fields. So I ask the model to generalize the disc theorem to finite fields. (this doesn't really make sense) It responds: > That's a great question that gets to the heart of the difference between continuous fields like the complex numbers (C) and discrete fields like finite fields (Fq ). > > There is no direct geometric analogue of the Gershgorin Circle Theorem for finite fields in the way... [] ... No Distance in Finite Fields: A finite field Fq is an algebraic structure, not a geometric one. There is no natural notion of "distance" or "absolute value" that behaves like it does in C. You can't define a "disk" of radius R around a point in a meaningful geometric way... (all
Can you give some details? I feel I’ve laid out a pretty detailed case. Maybe you did not read the paper I cited.
Like what do you mean when you say brute force with large datasets? I’m specifically talking about long range as in like trajectories 100 million steps long. And the details of training.
This has not been done with transformers and cannot be done without changing transformers a lot.
It also cannot be done with RNNs, unless you use something like the method I linked. Which is very different from standard training.