LESSWRONG
LW

Expertium
3476330
Message
Dialogue
Subscribe

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
Questions for old LW members: how have discussions about AI changed compared to 10+ years ago?
Expertium1mo20

Let me put it another way - do you expect that "LLMs do not optimize for a goal" will still be a valid objection in 2030? If yes, then I guess we have a very different idea of how progress will go.

Reply
Questions for old LW members: how have discussions about AI changed compared to 10+ years ago?
Expertium1mo10

But frontier labs are deliberately working on making LLMs more agentic. Why wouldn't they - AI that can do work autonomously is more economically valuable than a chatbot.

Reply
METR: How Does Time Horizon Vary Across Domains?
Expertium1mo10

Another suggestion: https://cybench.github.io/

Reply
IMO challenge bet with Eliezer
Expertium1mo90

https://x.com/alexwei_/status/1946477742855532918

I believe this qualifies as "technical capability existing by end of 2025".

Reply
Do confident short timelines make sense?
Expertium2mo10

For example, did any of the examples derive their improvement by some way other than chewing through bits of algebraicness? 

I don't think so.

Reply
Do confident short timelines make sense?
Expertium2mo30

https://deepmind.google/discover/blog/alphaevolve-a-gemini-powered-coding-agent-for-designing-advanced-algorithms/

https://arxiv.org/pdf/2506.13131

What did the system invent?

Example: matrix multiplication using fewer multiplication operations.

There were also combinatorics problems, "packing" problems (like multiple hexagons inside a bigger hexagon), and others. All of that is in the paper.

Also, "This automated approach enables AlphaEvolve to discover a heuristic that yields an average 23% kernel speedup across all kernels over the existing expert-designed heuristic, and a corresponding 1% reduction in Gemini’s overall training time."

How did the system work?

It's essentially an evolutionary/genetic algorithm, with LLMs providing "mutations" for the code. Then the code is automatically evaluated, bad solutions are discarded, and good solutions are kept.

What makes you think it's novel?

These solutions weren't previously discovered by humans. Unless the authors just couldn't find the right references, of course, but I assume the authors were diligent.

Would it have worked without the LLM?

You mean, "could humans have discovered them, given enough time and effort?". Yes, most likely.

Reply
Do confident short timelines make sense?
Expertium2mo10

I'm surprised to see zero mentions of AlphaEvolve. AlphaEvolve generated novel solutions to math problems, "novel" in the "there are no records of any human ever proposing those specific solutions" sense. Of course, LLMs didn't generate them unprompted, humans had to do a lot of scaffolding. And it was for problems where it's easy to verify that the solution is correct; "low messiness" problems if you will. Still, this means that LLMs can generate novel solutions, which seems like a crux for "Can we get to AGI just by incrementally improving LLMs?".

Reply
An Opinionated Guide to Using Anki Correctly
Expertium2mo30

Sounds like you could benefit either from Easy Days (available natively in the newer versions of Anki) or from Advance/Postpone from the FSRS Helper add-on

Reply
What does 10x-ing effective compute get you?
Expertium2mo50

https://www.virologytest.ai/

This benchmark has human expert percentiles, which makes it very convenient for exactly the kind of stuff you are doing (though I decided to calculate SDs as a function of release date rather than compute, just because it's more intuitive).

I wrote down SOTA models, their release dates, and performance:

ModelRelease dateNormalized dateAccuracyExpert percentilez-score
GPT-4 Turbo2023-06-01016.8%43%-0.18
Gemini 1.5 Pro2024-02-1525925.4%61%0.28
Sonnet 3.52024-06-2038526.9%69%0.50
Sonnet 3.5 v22024-10-2250933.6%75%0.67
o12024-12-0555335.4%89%1.23
o32025-04-1668543.8%94%1.55

Z-scores are based on expert percentiles. This gives roughly 0.90 SD/year for LLMs. So we should expect an LLM as good as a +6 SD human virology expert around 2030.

I wish more benchmarks had human percentiles.

Reply1
Foom & Doom 2: Technical alignment is hard
Expertium2mo*10

I'm curious why it seems better to you.

Because it's not rewarding AI's outward behavior. Any technique that just rewards the outward behavior is doomed once we get to AIs capable of scheming and deception. Self-other overlap may still be doomed in some other way, though.

It might choose to go along with its initial behavioral and ethical habits, or it might choose to deliberately undo the effects of the self-other overlap training once it is reflective and largely rational and able to make decisions about what goals/values to follow

That seems like a fully general argument that aligning a self-modifying superintelligence is impossible.

Reply
Load More
4810x more training compute = 5x greater task length (kind of)
2mo
8
2Let's look at another "LLMs lack true understanding" paper
2mo
0
6Smarter Models Lie Less
2mo
0
207Intelligence Is Not Magic, But Your Threshold For "Magic" Is Pretty Low
3mo
27
3Let's stop making "Intelligence scale" graphs with humans and AI
4mo
15
11Questions for old LW members: how have discussions about AI changed compared to 10+ years ago?
4mo
12