In 2015, I didn't write much about AI on Hacker News because even just explaining why it is dangerous will tend to spark enthusiasm for it in some people (people attracted to power, who notice that since it is dangerous, it must be powerful). These days, I don't let that consideration stop me from write about AI.
Yeah, 10 years ago it was: "we need smarter people, so that they can make the AI safe".
Today it is: "not enough time left to educate people, this is the crunch time".
There's lots of coding training data and not very much training data for creating documents of a specific length. I think if we added a bunch of "Write ### words about X" training data the LLM's would suddenly be good at it.
There's lots of coding training data and not very much training data for creating documents of a specific length. I think if we added a bunch of "Write ### words about X" training data the LLM's would suddenly be good at it.
My point was that it's surprising that AI is so bad at generalizing to tasks that it hasn't been trained on. I would've predicted that generalization would be much better (I also added a link to a post with more examples). This is also why I think creating AGI will be very hard, unless there will be a massive paradigm shift (some new NN architecture or a new way to train NNs).
EDIT: It's not "Gemini can't count how many words it has in its output" that surprises me, it's "Gemini can't count how many words it has in its output, given that it can code in Python and in a dozen other languages and can also do calculus".
"Alignment" is badly defined. It's ambiguous between "making AI safe, in anyway whatsoever" and "make Agentive AI safe by making it share human values" (As opposed to some other way, such as Control, or building only non Agentive AIs).
Alignment with "human values" is probably difficult, not least because "human value" isn't a well defined target. We can't define it as the output of CEV, because CEV isn't a method that's defined mechanically.
my take is that they haven't changed enough. People often still seem to be talking about agents and concepts that only make sense in the context of agents all the time - but LLMs aren't agents, they don't work that way. if often feels like the Agenda for the field got set 10+ years ago and now people are shaping the narrative around it regardless of how good a fit it is for the tech that actually came along.
But frontier labs are deliberately working on making LLMs more agentic. Why wouldn't they - AI that can do work autonomously is more economically valuable than a chatbot.
but they're not agents in the same way as the models in the thought experiments, even if they're more agentic. The base-level thing they do is not "optimise for a goal". We need to be thinking in terms of models that are shaped like the ones we actually have instead of holding on to old theories so hard we instantiate them in reality
I wasn't on LessWrong 10 years ago, during the pre-ChatGPT era, so I'd like to hear from people who were here back in the day. I want to know how AI-related discussions have changed and how people's opinions have changed.