are you forgetting specification gap between what someone actually wants vs what they say they want vs what someone (else) will actually write in Lean? ..and what they should have wanted if they knew better.. those are still kinds of "semantic gap", no?
Observation about working with coding agents: as I review outputs of claude code today (reading the plan and git diff before commit), my threshold for asking for a second opinion is much MUCH out of the way, I just perceive stuff as both "clearly questionable" and "good enough" at the same time - I used to ask other stakeholders (other devs sitting next to me in the office, the tester assigned to the issue, or the PO depending on who was more conveniently available and how much my "best guess" could have been different from "other plausible interpretations of the known requirements") for their concrete opinions about concrete "unimportant trivia" from time to time, but now it feels like I am the second person, as if claude would ask other people if I said "I don't know, sounds good enough" about something... but it won't. And I know it won't but it still doesn't feel like my job somehow to follow-up on those feelings when something is "good enough" but I might have followed up on some of that during a coffee break or something, but now I know I won't.
On a scale from 2025-04-01 to 2026-04-01, after meditating on it for a moment, I decided to take this post 73% seriously.
My perception of time is like sampling from a continuous 2D plane of conceptual space, something akin to git railways but with a thickness to the lines, like electron orbital probability clouds that are dense around plausible points of view of what happened in the past and thin around conspiracy theories, as different linear mind-map perspectives of people standing around a sculpture, each only inferring what's on the other side, but together they can prune down non-overlapping minority reports, sticking to the consensus but never deleting (git) history.
My sense of beauty finds it displeasing to read articles with only point measurements and only n-point predictions and to look at charts from a single interpretation / scenario, I have to hallucinate the gaps, infer the systemic bias and imagine the size of the error bars due to "randomness" as if the authors were good intentioned and as if the authors had an agenda to prove a point, would they stumble upon convenient evidence before they stopped looking?
But alternative timelines that are infinitesimally thin and split only on known unknowns would imply perfect Bayesian approximators, an impossible standard, uncomputable. No one has ever made that kind of precise prediction, why do we allow prediction-readers to behave as if prediction-writers could have made an infinitely precise measurable decidable statements with completely non-ambiguous semantics that will be evaluated to a non-reversible Boolean?
...we aphantasiacs see nothing but darkness.
Instead of seeing images, I think exclusively in terms of concepts. For example, I know what my dog looks like, but not from visualizing her in my mind. Instead, I can describe my dog by listing her characteristics stored in my memory—like I’m accessing words from a txt file.
Hm, it sounds like as if you were grieving a loss... How did you come to the conclusion that there is anything of substance that is desirable to have that you cannot have? I reached the opposite conclusion, that "normal" people are misfortunate if they auto-sample a concrete instance of abstract concepts like "a red apple" so prematurely.
While I myself I don't "see", "hear", "smell", or "touch" my imagined concepts, those are are much richer objects for me than mere sensory input hallucinations would have been - and I have very rich sensory inputs to start with - when I close my eyes, I don't see darkness, I see a red-ish background with dynamic blue-green "pixel noise". And I still know where the windows are in the 3D space around me as I rotate in a room with my eyes closed, while I can also see afterimages from the screen but those stay 2D-projected in front where I point my eyes as I move them inside my eye sockets. I perceive the shape of the room from my other senses as distinct from the visual-only layers, it doesn't go away by merely closing my eyes, but I also cannot help but integrate the visual info of the direction of the light source to inform my sense of the shape of the room - putting my hands in front of my closed eyes makes it harder to walk through the door if I turn around a few times.
..it's just the imagined apple comes from neither of those senses, it floats on a different plane of existence. I can give it colors, rotate it, I can give it smells, I can give it a past, I can give it a future, I know what sound it would make if I eat it and it was still fresh and crisp or if it was old and soggy. I know all that the instant I focus on any aspect of that apple that exists only in my mind, I am not accessing words from a txt file, I am accessing the concept that is much richer than I could put into words.
And if you can shuffle a key into a door even when holding a bunch of bags that obscure the lock, you have evidence that you must be able to imagine the 3D space that you learned from your vision even without seeing it directly with your eyes - you are not describing how to find the lock with just your fingers and your ears in words, I would define the situation as "imagining" how the real world really is even without "seeing" it.
(Though now that I paid too much attention to my vision, I will see the stupid tangled protein floaters in front of the screen until I distract myself with something, aaah...)
first... wouldn't that be evidence that models are NOT actually aware of whether or not they are in an eval? second, how do you figure out whether they are aware of the real situation vs always think they are in an eval with some probability BEFORE releasing to production? I am trying to understand how "eval awareness" is a useful concept for doing something about it and I feel like I'm missing something here
hm, most wannabe new games won't get to the publishing phase tho.. or even attempt to get to steam -📱? (...also probably best to count roblox as "1 game" if we are interested in sales, but...)
Incentives and institutions.
enjoy the process
yup, that's going to be my strategy next year, taking a couple of months without doing any software engineering at all last summer might not have solved alignment, but I realized I don't care for humanity that much and that staying away from AI safety for a few years didn't help to keep it out of my mind anyway... so I collected a few points what makes the process more enjoyable for me in My burnout journey and will try to keep up with those points when helping to build infrastructure on various projects 🤞
(and even if the same person(a) fulfills all roles in a principal-agent thought experiment, humans are not coherent VNM agents, a person at different instances of time can be modelled as diverging agents, some being more tired, distracted, wiser, skilled, etc as if they didn't have the same values and goals ... doesn't that fuck with all hope about scalable practicality of formal proofs?)