Logan Zoellner

Wiki Contributions

Comments

Sorted by

Einstein didn't write a half-assed NYT op-ed about how vague 'advances in science' might soon lead to new weapons of war and the USA should do something about that; he wrote a secret letter hand-delivered & pitched to President Roosevelt by a trusted advisor.

Strongly agree.

What other issues might there be with this new ad hoced strategy...?

I am not a China Hawk.  I do not speak for the China Hawks.  I 100% concede your argument that these conversations should be taking place in a room that neither you our I are in right now.

I would like to see them state things a little more clearly than commentators having to guess 'well probably it's supposed to work sorta like this idk?'

Meh.  I want the national security establishment to act like a national security establishment.  I admit it is frustratingly opaque from the outside, but that does not mean I want more transparency at the cost of it being worse.  Tactical Surprise and Strategic Ambiguity are real things with real benefits.

A great example, thank you for reminding me of it as an illustration of the futility of these weak measures which are the available strategies to execute.

I think both can be true true: Stuxnet did not stop the Iranian nuclear program and if there was a "destroy all Chinese long-range weapons and High Performance Computing clusters" NATSEC would pound that button.  

Is your argument that a 1-year head start on AGI is not enough to build such a button, or do you really think it wouldn't be pressed?

It is a major, overt act of war and utter alarming shameful humiliating existential loss of national sovereignty which crosses red lines so red that no one has even had to state them - an invasion that no major power would accept lying down and would likely trigger a major backlash

The game theory implications of China waking up to finding all of their long-range military assets and GPUs have been destroyed are not what you are suggesting.  A very telling current example being the current Iranian non-response to Israel's actions against Hamas/Hezbollah.

Nukes were a hyper-exponential curve too.

While this is a clever play on words, it is not a good argument.  There are good reasons to expect AGI to affect the offense-defense balance in ways that are fundamentally different from nuclear weapons. 

Because the USA has always looked at the cost of using that 'robust military superiority', which would entail the destruction of Seoul and possibly millions of deaths and the provoking of major geopolitical powers - such as a certain CCP - and decided it was not worth the candle, and blinked, and kicked the can down the road, and after about three decades of can-kicking, ran out of road.

 

I can't explicitly speak for the China Hawks (not being one myself), but I believe one of the working assumptions is that AGI will allow the "league of free nations" to disarm China without the messiness of millions of deaths.  Probably this is supposed to work like EY's "nanobot swarm that melts all of the GPUs".  

I agree that the details are a bit fuzzy, but from an external perspective "we don't publicly discuss capabilities" and "there are no adults in the room" are indistinguishable.  OpenAI openly admits the plan is "we'll as the AGI what to do".  I suspect NATSEC's position is more like "amateurs discuss tactics, experts discuss logistics" (i.e. securing decisive advantage is more important that planning out exactly how to melt the GPUs)

To believe that the same group that pulled of Stuxnet and this lack the imagination or will to use AGI enabled weapons strikes me as naive, however.

 

The USA, for example, has always had 'robust military superiority' over many countries it desired to not get nukes, and yet, which did get nukes.

It's also worth nothing AGI is not a zero-to-one event but rather a hyper-exponential curve.  Theoretically it may be possible to always stay far-enough-ahead to have decisive advantage (unlike nukes where even a handful is enough to establish MAD).

Okay, this at least helps me better understand your position.  Maybe you should have opened with "China Hawks won't do the thing they've explicitly and repeatedly said they are going to do"
 


 

What does winning look like? What do you do next?

 

This question is a perfect mirror of the brain-dead "how is AGI going to kill us?" question.  I could easily make a list of 100 things you might do if you had AGI supremacy and wanted to suppress the development of AGI in China.  But the whole point of AGI is that it will be smarter than me, so anything I put on the list would be redundant.

Playing the AIs definitely seems like the most challenging role

 

Seems like a missed opportunity not having the AIs be played bi AIs

This is a bad argument, and to understand why it is bad, you should consider why you don't routinely have the thought "I am probably in a simulation, and since value is fragile the people running the simulation probably have values wildly different than human values so I should do something insane right now"

Chinese companies explicitly have a rule not to release things that are ahead of SOTA (I've seen comments of the form "trying to convince my boss this isn't SOTA so we can release it" on github repos).  So "publicly release Chinese models are always slightly behind American ones" doesn't prove much.

Answer by Logan Zoellner1-1

Current AI methods are basically just fancy correlations, so unless the thing you are looking for is in the dataset (or is a simple combination of things in the dataset) you won't be able to find it.

This means "can we use AI to translate between humans and dolphins" is mostly a question of "how much data do you have?"

Suppose, for example that we had 1 billion hours of audio/video of humans/dolphins doing things.  In this case, AI could almost certainly find correlations like: when dolphins pick up the seashell, they make the <<dolphin word for seashell>> sound, when humans pick up the seashell they make the <<human word for seashell>> sound.  You could then do something like CLIP to find a mapping between <<human word for seashell>> and <<dolphin word for seashell>>.  The magic step here is because we use the same embedding model for video in both cases, <<seashell>> is located at the same position in both our dolphin and human CLIP models.

But notice that I am already simplifying here.  There is no such thing as <<human word for seashell>>.  Instead, humans have many different languages.  For example Papua New Guinea has over 800 languages in a land area of a mere 400k square kilometers.  Because dolphins are living in what is essentially a hunter-gatherer existence, none of the pressures (trade, empire building) that cause human languages to span widespread areas exist.  Most likely each pod of dolphins has at a minimum its own dialect. (one pastime I noticed when visiting the UK was that people there liked to compare how towns only a few miles apart had different words for the same things)

Dolphin lives are also much simpler than human lives, so their language is presumably also much simpler.  Maybe like Eskimos have 100 words for snow, dolphins have 100 words for water.  But it's much more likely that without the need to coordinate resources for complex tasks like tool-making, dolphins simply don't have as complex a grammar as humans do.  Less complex grammar means less patterns means less for the machine learning to pick up on (machine learning loves patterns).  

So, perhaps the correct analogy is: if we had a billion hours of audio/video of a particular tribe of humans and billion hours of a particular pod of dolphins we could feed it into a model like CLIP and find sounds with similar embeddings in both languages.  As pointed out in other comments, it would help if the humans and dolphins were doing similar things, so for the humans you might want to pick a group that focused on underwater activities.

In reality (assuming AGI doesn't get there first, which seems quite likely), the fastest path to human-dolphin translation will take a hybrid approach.  AI will be used to identify correlations in dolphin language.  For example this study that claims to have identified vowels in whale speech.  Once we have a basic mapping: dolphin sounds -> symbols humans can read, some very intelligent and very persistent human being will stare at those symbols, make guesses about what they mean, and then do experiments to verify those guesses.  For example, humans might try replaying the sounds they think represent words/sentences to dolphins and seeing how they respond.  This closely matches how new human languages are translated: a human being lives in contact with the speakers of the language for an extended period of time until they figure out what various words mean.

What would it take for an only-AI approach to replicate the path I just talked about (AI generates a dictionary of symbols that a human then uses to craft a clever experiment that uses the least amount of data possible)?  Well, it would mean overcoming the data inefficiency of current machine learning algorithms.  Comparing how many "input tokens" it takes to train a human child vs GPT-3, we can estimate that humans are ~1000x more data efficient than modern AI techniques.  

Overcoming this barrier will likely require inference+search techniques where the AI uses a statistical model to "guess" at an answer and then checks that answer against a source of truth.  One important metric to watch is the ARC prize, which intentionally has far less data than traditional machine learning techniques require.  If ARC is solved, it likely means that AI-only dolphin-to-human translation is on its way (but it also likely means that AGI is immanent).

So, to answer your original question: "Could we use current AI methods to understand dolphins?"  Yes, but doing so would require an unrealistically large amount of data and most likely other techniques will get there sooner.

Load More