LESSWRONG
LW

567
Alexander Gietelink Oldenziel
5953Ω1962410040
Message
Dialogue
Subscribe

" (...) the term technical is a red flag for me, as it is many times used not for the routine business of implementing ideas but for the parts, ideas and all, which are just hard to understand and many times contain the main novelties." 
                                                                                                           - Saharon Shelah
"A little learning is a dangerous thing ;
Drink deep, or taste not the Pierian spring"                                                                                                           - Alexander Pope

 

As a true-born Dutchman I endorse  Crocker's rules.

For my most of my writing see my short-forms (new shortform, old shortform)

Twitter: @FellowHominid

Personal website: https://sites.google.com/view/afdago/home

Sequences

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
Singular Learning Theory
5Alexander Gietelink Oldenziel's Shortform
Ω
3y
Ω
605
Daniel Kokotajlo's Shortform
Alexander Gietelink Oldenziel6dΩ564

Drone countermeasures are an idle hope. The only real counter to drones is more drones. 

Lasers, shotguns, tank redesign [no holes!], nets, counter-drones, flak etc will all be part of the arsenal surely but thinking drone countermeasures are going to restore the previous generation's war doctrine is as silly as thinking that metallurgy innovations will reverse the gunpowder age. 

The future present of warfare is drones, drones, drones. 

Reply
Daniel Kokotajlo's Shortform
Alexander Gietelink Oldenziel6dΩ342

It seems nobody outside Ukraine/Russia is actually at the leading edge of where the reality of military technology is. That includes Hamas. Even using the drone doctrine from two years ago would be devastating to the Israelis. Probably they don't have the resources, organization to do so. 

[Even Ukraine itself is not really there - there are clearly many simple ways drones and drone manufacturing could be improved they haven't had the time and resources to focus on yet. ]

Expect terror/resistance groups to start utilizing drones en masse in the next few years. 

Reply
Alexander Gietelink Oldenziel's Shortform
Alexander Gietelink Oldenziel6d47

Referring to particular models and particular deployment plans and particular catastrophe doesn't help - the answer is the same. 

We don't know how to scientifically quantify any of these probabilities. 

Reply
Jacob_Hilton's Shortform
Alexander Gietelink Oldenziel7d197

Couldn't agree more! 

One of the saddest ways we would die is if we fail to actually deploy enough theory/math people to work on prompting AIs to solve alignment in the next two-three years, even if it would be totally possible to solve alignment this way. 

P.S. Please get into contact if you are interested in this. 

Reply
Alexander Gietelink Oldenziel's Shortform
Alexander Gietelink Oldenziel7d614

I'm having trouble seeing that  (2) is actually a thing?

The whole problem is that there is no generally agreed " chance of catastrophe" so " same change of catastrophe" has no real meaning. 
It seems this kind of talk is being backchained from what governance people want as opposed to the reality of safety guarantees or safety/risk probabilities - which is that they don't meaningfully exist [outside of super heuristic guesses]. 

Indeed, to estimate this probability in a non-bullshit way we exactly need fundamental scientific progress, i.e. (1).

EA has done this exercise a dozen times: if you ask experts the probabilities of doom it ranges all the way from 99% to 0.0001 %. 

Will that change? Will expert judgement converge? Maybe. Maybe not. I don't have a crystal ball.

Even if they do [outside of meaningful progress on (1)] those probabilities won't actually reflect reality as opposed to political reality. 

The problem is there is no ' scientific'  way to estimate p(doom) and as long as we don't make serious progress on 1. there won't be. 

I don't see how cot/activation/control monitoring will have any significant and scientifically -grounded [as opposed to purely story-telling/politics] influence in a way that can be measured and can be utilized to make risk-tradeoffs. 

Reply1
Alexander Gietelink Oldenziel's Shortform
Alexander Gietelink Oldenziel7d60

I didn't intend it that way, though admittedly that is a valid reading. From my own point of view both functions seem significant. 

Reply
Alexander Gietelink Oldenziel's Shortform
Alexander Gietelink Oldenziel7d*268

Additive versus Multiplicative model of AI-assisted research

Occasionally one hears somebody say "most of the relevant AI-safety work will be done at crunch time. Most work being done now at present is marginal". 

One cannot shake the suspicion that this statement merely reflects the paucity of ideas & vision of the speaker. Yet it cannot be denied that their reasoning has a certain logic: if, as seems likely, AI will become more and more dominant in AI alignment research than maybe we should be focusing on how to safely extract work from future superintelligent machines rather than hurting our painfully slow mammalian walnuts to crack AI safety research today. I understand this to be a key motivation for several popular agendas AI safety.

Similarly, many Pause advocates argue that pause advocacy  is more impactful than direct research. Most will admit that a Pause cannot be maintained indefinitely. The aim of a Pause would be to buy time to figure out alignment. Unless one believes in very long pauses, implicitly it seems there is an assumption that research progress will be faster in the future. 

Implicitly, we might say there is underlying " Additive" model of AI-assisted research: there is a certain amount of research that is to be done to solve alignment. Humans do some of it, AI does some (more). If the sum is large enough we are golden.

Contrast this with a " multiplicative" model of AI-assisted research:

Humans increasingly resort to a supervisory and directing role for AI alignment research by AIs. Human experts become 'research managers' of AI grad students. The key bottleneck increasingly is the capacity for humans to effectively oversee, supervise, direct, and judge the AI's research output. In this model research effort by humans and AI is multiplicative. The better the understanding of the humans, the more AI can be effectively leveraged.

These are highly simplified models of course. I still think it's worthwhile to keep them in mind since they imply very different courses of action. In the additive framework direct research is somewhat marginal. In the multiplicative model every research XP point that humans build today is worth 10x tomorrow in effective research output at crunch time. [1]


 

  1. ^

    Moreover, in the latter model it is especially important to start thinking today about very ambitious plans/ideas/methods that are less likely to be obsoleted the moment the next generation of AI comes along. 

Reply11
MakoYass's Shortform
Alexander Gietelink Oldenziel8d20

I wonder what the evidence is that politicians trading on legislative decisions is very harmful. 

It seems distasteful to be sure. A moral failure. But how bad is it really?

Reply
adamzerner's Shortform
Alexander Gietelink Oldenziel8d1111

>> Anecdotally I feel like there are a fair amount of examples of people who have kids and end up regretting it, at least to a meaningful degree.

The social desirability bias against saying one regrets having kids is very intense. So although I would very much not say that the following observation is anything but a highly biased sample: I personally have never met a person regretting having children in real life. I've heard parents complain for sure --- true regrets seem to be rare.  

Reply
johnswentworth's Shortform
Alexander Gietelink Oldenziel10d270

... and so at long last John found the answer to alignment

The answer was Love 

and it had always has been

Reply3
Load More
78Proceedings of ILIAD: Lessons and Progress
5mo
5
80Announcing ILIAD2: ODYSSEY
Ω
6mo
Ω
1
99Timaeus in 2024
Ω
7mo
Ω
1
90Agent Foundations 2025 at CMU
Ω
8mo
Ω
10
67Timaeus is hiring!
Ω
1y
Ω
6
163Announcing ILIAD — Theoretical AI Alignment Conference
Ω
1y
Ω
18
20Are extreme probabilities for P(doom) epistemically justifed?
2y
12
173Timaeus's First Four Months
Ω
2y
Ω
6
59What's next for the field of Agent Foundations?
Ω
2y
Ω
23
188Announcing Timaeus
Ω
2y
Ω
15
Load More