LESSWRONG
LW

Tao Lin
913Ω511660
Message
Dialogue
Subscribe

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
2Tao Lin's Shortform
4y
6
No wikitag contributions to display.
Yudkowsky on "Don't use p(doom)"
Tao Lin9d20

I'm surprised that's the question. I would guess that's not what Eliezer means because he says Dath Ilan is responding sufficiently to AI risk but also hints at Dath Ilan still spending a significant fraction of its resources on AI safety (I've only read a fraction of the work here, maybe wrong). I have a background belief that the largest problems don't change that much, and it's rare for problems to go from #1 problem to not-in-top-10 problems, and that most things have diminishing returns such that it's not worthwhile to solve them so thoroughly. An alternative definition that's spiritually similar that I like more is; "What policy could governments implement such that the improving the AI x-risk policy would now not be the #1 priority, if the governments were wise.". This isolates AI / puts it in context of other global problems, such that the AI solution doesn't need to prevent governments from changing their minds over the next 100 years or whatever needs to happen for the next 100 years to go well.

Reply
Thomas Kwa's Shortform
Tao Lin23d40

I would expect aerodynamic maneuvering MIRVS to work and not be prohibitively expensive. The closest deployed version appears to be https://en.wikipedia.org/wiki/Pershing_II which has 4 large fins. You likely don't need that much steering force

Reply
Consider chilling out in 2028
Tao Lin1mo80

I really struggle to think of problems you want to wait 2.5 years to solve - when you identify a problem, you usually want to work on solving it within the month. Just update most of the way now + a tiny bit over time as evidence comes in. As others commented, no doom by 2028 is very little evidence

Reply
How Fast Can Algorithms Advance Capabilities? | Epoch Gradient Update
Tao Lin1mo10

I heard some rumors that gpt 4.5 got good pretraining loss but bad downstream performance. If that's true the loss scaling laws may have worked correctly. If not, yeah a lot of things can go wrong and something did, whether that's hardware issues, software bugs, or machine learning problems or problems with their earlier experiments

Reply
OpenAI Claims IMO Gold Medal
Tao Lin1mo51

This is OpenAI cot style. See it in the original o1 blog post. https://openai.com/index/learning-to-reason-with-llms/

Reply
Video and transcript of talk on "Can goodness compete?"
Tao Lin1mo20

I can imagine scenarios where you could end up with more resources from causing vacuum decay without extortion. Like if you care about doing something with resources quickly and other agents want to use resources slowly, then if you cause vacuum decay inside your region, the non collapsed shell of your region becomes more valuable to you relative to other agents because it only exists for a short duration, and maybe that makes other agents fight over it less. Or maybe you can vacuum decay into a state that still supports life and you value that

Reply
Video and transcript of talk on "Can goodness compete?"
Tao Lin1mo41

Whether you can cause various destructive chain reactions is pretty important. If locusts could benefit from causing vacuum collapse, or could trigger star supernova, or could efficiently collapse various bodies into black holes, that could easily eat up large fractions of the universe.

Reply
The Cult of Pain
Tao Lin2mo153

No, AC actually moves 2-3x as much heat as it's input power, so a 1500W AC will extract an additional 3000W from inside and dump 4500W outside

Reply
When is it important that open-weight models aren't released? My thoughts on the benefits and dangers of open-weight models in response to developments in CBRN capabilities.
Tao Lin3moΩ684

This overestimates the impact of large models on external safety research. My impression is that the AI safety community has barely used deepseek r1 and v3 open source weights at all. I checked again and still see little evidence of v3/r1 weights in safety research. People use r1 distill 8b, and qwq 32b, but the decision to open source the most capable small model is different than the decision to open source the frontier. So then it matters when 8b or 32b models can assist with bioterrorism, which happens a bit later, and we get most of the benefits of open source until then. It's also cheaper to filter virology or even all biology data out of training for a small models pre training data because it wouldn't cause customers to switch providers (customers prefer large model anyway) and small models are more often narrowly focused on math or coding.

Reply
Season Recap of the Village: Agents raise $2,000
Tao Lin3mo180

What are your API costs, and how do they compare to the $ raised?

Reply
Load More
77Send us example gnarly bugs
Ω
2y
Ω
10
34Causal scrubbing: results on induction heads
Ω
3y
Ω
1
34Causal scrubbing: results on a paren balance checker
Ω
3y
Ω
2
2Tao Lin's Shortform
4y
6