Former safety researcher & TPM at OpenAI, 2020-24
https://www.linkedin.com/in/sjgadler
stevenadler.substack.com
I really appreciate that the charts show which models are frontier; I'd like to see more groups adopt that convention
Yeah I agree that works and feel slightly sheepish not to have already internalized that as the term to use?
I guess there’s still some distinction between an objective as a single thing, vs drives as, like, heuristics that will tend to contribute toward shaping the overall objective? I’m not sure, still feel a bit fuzzy and should probably sit with it more
Maybe instead of talking about AI having 'goals' or 'drives,' which sound biological, it would be helpful to use a more mechanical-sounding term, like a 'target state,' or 'termination conditions,' or 'prioritization heuristics'?
A surprising number of people seem to bounce off the idea of 'AI being dangerous' once they encounter anything that feels like anthropomorphism.
My suggestions above are all kind of clunky (and 'termination conditions' has an unfortunate collision with The Terminator). But I think the spirit is correct of 'we should find ways to describe AI's behavioral tendencies with as little anthropomorphizing as possible'?
Hmm I don't think so? If you buy land for $X, that's the floor on what you could reasonably assess it at, which is basically the status quo world. So we're in the status quo until someone comes along and bids up the price to their willingness-to-pay: Then, the asset either moves to someone who values it more, or you start paying higher taxes on it. I think either branch is preferable to the status quo?
Yup this makes more sense imo, basically having a right of refusal on the sale, but reflecting the now-assessed-higher tax rate
“should you be trying to dispose of all of your money before the singularity makes it worthless”
This is pretty different than my model of what would happen? Though I admittedly haven’t spent a ton of time thinking through it. I just don’t see why money would lose value though; I expect that some goods would still remain scarce, positional, etc (land in high-demand cities being a strong example), which would seem to cut against that happening?
Maybe I’ve been misusing it or seeing it misused, but I thought it meant something more like “called a thing ahead of time” or “made a good prediction” and therefore treated as more credible in the future?
Presumably you'd still feel productivity effects from not having a monitor, having worse ergonomics, etc?
I was surprised to see you say above that you'd anticipate flying way more often! Are there times you've wanted to fly recently but held off because you couldn't spare the lost hours of flying? (I would have expected the bigger barrier to be the loss of productive hours from, say, being out-of-the-office in the destination itself)
I've been wondering about this in terms of my own writing, whether I should be working on multiple pieces at once to a greater degree than I am. Thinking aloud a bit:
I guess part of the question is, what are the efficiency effects of batch-processing, vs the more diluted feedback signal from multiple 'coming off the production line' at once? Though in my case, I'd probably still stagger the publication, and so maybe that's less of a concern (though there may still be some dilution from having shallower focus on each piece-in-process).
At first I wondered how quickly a human could do this, with tooling?
The thing I was trying to get at is, like, distinguishing reading-speed from reasoning-speed, though in retrospect I think these may not be very separable in this case.
I guess there's the degenerate case of feeding those words to an AI and saying "what's the needle?"
I had meant something that still involved human cognition, just with faster rifling through the text. Like maybe a method that embedded the text, and then you could search through it more quickly.
But in retrospect, the "still uses cognition" version is probably just asking the model "What are a few possible needles?" and then using your judgment among the options.