LESSWRONG
LW

1719
Oliver Sourbut
1088Ω120212530
Message
Dialogue
Subscribe

oliversourbut.net

  • Autonomous Systems @ UK AI Safety Institute (AISI)
  • DPhil AI Safety @ Oxford (Hertford college, CS dept, AIMS CDT)
  • Former senior data scientist and software engineer + SERI MATS

I'm particularly interested in sustainable collaboration and the long-term future of value. I'd love to contribute to a safer and more prosperous future with AI! Always interested in discussions about axiology, x-risks, s-risks.

I enjoy meeting new perspectives and growing my understanding of the world and the people in it. I also love to read - let me know your suggestions! In no particular order, here are some I've enjoyed recently

  • Ord - The Precipice
  • Pearl - The Book of Why
  • Bostrom - Superintelligence
  • McCall Smith - The No. 1 Ladies' Detective Agency (and series)
  • Melville - Moby-Dick
  • Abelson & Sussman - Structure and Interpretation of Computer Programs
  • Stross - Accelerando
  • Graeme - The Rosie Project (and trilogy)

Cooperative gaming is a relatively recent but fruitful interest for me. Here are some of my favourites

  • Hanabi (can't recommend enough; try it out!)
  • Pandemic (ironic at time of writing...)
  • Dungeons and Dragons (I DM a bit and it keeps me on my creative toes)
  • Overcooked (my partner and I enjoy the foody themes and frantic realtime coordination playing this)

People who've got to know me only recently are sometimes surprised to learn that I'm a pretty handy trumpeter and hornist.

Sequences

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
Breaking Down Goal-Directed Behaviour
No wikitag contributions to display.
6Oliver Sourbut's Shortform
3y
4
Trust me bro, just one more RL scale up, this one will be the real scale up with the good environments, the actually legit one, trust me bro
Oliver Sourbut2d*20

I do think that this sort of "AIs generate their own training environments" flywheel could cause superexponential progress via the same sort of mechanism as AIs automating AI R&D, though I don't expect to see this data generation effect show up much in overall AI progress.

Confused by this. "AIs generate their own training environments" is (a central part of) AI R&D, right? (Besides this there's what, compute optimisation of various kinds, which shades into architecture and training algo and hyperparameter innovation/exploration, plus scaffolding, prompting/input design, and tooling/integrations? I'd weight data+environment as ~a third of AI R&D.)

Separately, how to square '...could cause superexponential...' with 'I don't expect... to show up much'? Is it just that you think it's conceivable but low probability?

Reply
An epistemic advantage of working as a moderate
Oliver Sourbut16d*70

Mostly unfiltered blurting

Counterfactual?

Control is super obvious and not new conceptually; rather it's a bit new that someone is actually trying to do the faffy thing of making something maybe work. I think it's pretty likely they'd be doing it anyway?

Counterpoint: companies as group actors (in spite of intelligent and even caring constituent humans) are mostly myopic and cut as many corners as possible by default (either due to vicious leadership, corporate myopia, or (perceived) race incentives), so maybe even super obvious things get skipped without external parties picking up the slack?

The same debate could perhaps be had about dangerous capability evaluations.

Reply
Yudkowsky on "Don't use p(doom)"
Oliver Sourbut18d20

As a minor anecdatal addition: people (at least in the bay) were annoyingly and unproductively asking about p(doom) before Death with Dignity was published.

Reply1
How quick and big would a software intelligence explosion be?
Oliver Sourbut1mo*20

This is nicely put together as usual! Sadly still leaves me feeling uncomfortable. Trying to put my finger on why, and I think it's at least two things.

Mainly: lumping so many things together as a single scalar 'software' just smells really off to me! Perhaps I'm 'too close' to the problem, having got deep on the nitty gritty of pretty well all aspects of software here at one time or another. You definitely flag that admirably, and have some discussion on how to make those things sort of commensurable. I do think it's important to at least distinguish efficiency from 'quality', and perhaps to go further (e.g. distinguishing training from runtime efficiencies, or even speed from parallel efficiencies).

I also think in treatment of R&D it's important to distinguish steady/stock 'quality' from learning/accrual 'quality', and to acknowledge that all of these things deprecate as you move through scale regimes: today's insights may or may not stand up to a 10x or a 100x of your system parameters. This makes sample efficient generalisation and exploratory heuristics+planning really key.

Related to this (and I'm less sure, having not deeply interrogated the details), I feel like some double counting of factors is going into the estimates of parameters, especially r. But I can imagine retracting this on further scrutiny.

Reply
Civil Service: a Victim or a Villain?
Oliver Sourbut1mo40

Anecdata: I quit the UK civil service (AISI, quite atypical) in small part due to feeling micromanagement creeping in, but also I mostly experienced a good amount of trust and autonomy. People sometimes said 'startup in government', perhaps to convey this (among other features). I don't know how long that's sustainable.

Reply
Civil Service: a Victim or a Villain?
Oliver Sourbut1mo2-1

It's strange... I have a sense that civil services used to be enviable, admirable places to work. Elite selection and talented operators getting impressive things done. At least in the UK and China, and also in the US (presumably also other places but I'm less familiar). What went wrong?

My limited experience (c. 14 months as a 'civil servant' in the UK AISI, really more like an imported technical expert pretending to be a civil servant) was that I encountered a predominance of very driven, intelligent, and well-motivated people in my immediate vicinity (including, perhaps especially, the 'real' civil servants who'd pre-existed AISI). I expect there may have been substantial selection effects. I heard complaints about interfacing with other parts of govt, and certain kinds of moves (like publication or sending messages 'up the chain' to MPs) were highly bureaucratised and a bit Procrustean in ways that sometimes limited our autonomy and ability to achieve the mission.

Reply
Will compute bottlenecks prevent a software intelligence explosion?
Oliver Sourbut1mo20

I also noticed this! And wondered how much evidence it is (and of what). I don't think of Meta as especially rational in its AI-related behaviour. Maybe this is Zuckerberg trying to make up for Le Cun's years of bad judgement.

Reply
Jemist's Shortform
Oliver Sourbut1mo50

(Psst: a lot of AISI's work is this, and they have sufficient independence and expertise credentials to be quite credible; this doesn't go for all of their work, some of which is indeed 'try for a better plan')

Reply
Better than logarithmic returns to reasoning?
Oliver Sourbut1mo30

Indeed. von Neumann

All stable processes we shall predict. All unstable processes we shall control.

Though some chaotic processes are quite hard to control (and others might not take kindly to your attempting to control them)!

Reply
Better than logarithmic returns to reasoning?
Oliver Sourbut1mo*20

Some gestures which didn't make the cut as they're too woolly or not quite the right shape:

  • 'iteration' or 'refinement' of proposals is the serial counterpart to best-of-k
    • if per-cycle marginal improvement to proposals is proportional to 1/t (diminishing), then overall return is logarithmic
    • we have to demand constant marginal improvement to get overall linear returns, though there are intermediates
    • why and when are marginal reasoning iteration improvements constant vs diminishing around 1/t vs something else?
  • adversarial exponentials might force exponential expense per gain
    • e.g. combatting replicators
    • e.g. brute forcing passwords
  • many empirical 'learning curve' effects appear to consume exponential observations per increment
    • Wright's Law (which is the more general cousin of Moore's Law) requires exponentially many production iterations per incremental efficiency gain
    • Deep learning scaling laws appear to consume exponential inputs per incremental gain
    • AlphaCode and AlphaZero appear to make uniform gains per runtime compute doubling
    • OpenAI's o-series 'reasoning models' appear to improve accuracy on many benchmarks with logarithmic returns to more 'test time' compute
    • (in all of these examples, there's some choice of what scale to represent 'output' on, which affects whether the gains look uniform or not, so the thesis rests on whether the choices made are 'natural' in some way)
Reply
Load More
14Better than logarithmic returns to reasoning?
2mo
5
50Do LLMs know what they're capable of? Why this matters for AI safety, and initial findings
Ω
2mo
Ω
4
20You Can’t Skip Exploration: Why understanding experimentation and taste is key to understanding AI
4mo
0
75FLF Fellowship on AI for Human Reasoning: $25-50k, 12 weeks
4mo
1
26Deceptive Alignment and Homuncularity
8mo
12
8Cooperation and Alignment in Delegation Games: You Need Both!
1y
0
17Terminology: <something>-ware for ML?
Q
2y
Q
27
6Alignment, conflict, powerseeking
2y
1
16Careless talk on US-China AI competition? (and criticism of CAIS coverage)
2y
3
12Invading Australia (Endless Formerlies Most Beautiful, or What I Learned On My Holiday)
2y
1
Load More