I want everyone to be able to achieve world they would really like; guess the best way to do that is to help people learn, build one's strengths, build small-scale and large-scale projects, and also to cooperate.
Any gift must be accepted with gratitude and, if possible, with grace.
As of 2024-12-21, I have signed no contracts I cannot mention exist. I finally got to adding this notice thanks to the one in the-gears-to-ascension's bio.
Aidan McLaughlin (OpenAI): ignore literally all the benchmarks the biggest o3 feature is tool use. Ofc it’s smart, but it’s also just way more useful. >deep research quality in 30 seconds >debugs by googling docs and checking stackoverflow >writes whole python scripts in its CoT for fermi estimates McKay Wrigley: 11/10
Newline formatting is off (and also for many previous posts).
I'm not sure if this will be of any use, since your social skills will surely be warped when you expect iterating on them (in a manner like radical transparency reduces awareness of feelings ).
Most cases, you could build a mini-quadcopter, "teach" it some tricks and try showcasing it, having video as a side effect!
Browsing through recent rejects, I found an interesting comment that suggests an automatic system "build a prior on whether something is true by observing social interactions around it vs competing ideas".
@timur-sadekov While it will fail most certainly, I do remember that the ultimate arbiter is experiment, so I shall reserve the judgement. Instead, I'm calling for you to test the idea over prediction markets (Manifold, for instance) and publish the results.
The AI industry is different—more like biology labs testing new pathogens: The biolab must carefully monitor and control conditions like air pressure, or the pathogen might leak into the external world.
This looks like a cached analogy, because given previous paragraph this would fit more: "the AI industry is different, more like testing planes; their components can be internally tested, but eventually the full plane must be assembled and fly, and then it can damage the surroundings".
Same goes for AI: If we don’t keep a watchful eye, the AI might be able to escape—for instance, by exploiting its role in shaping the security code meant to keep the AI locked inside the company’s computers.
This seems very likely to happen! We've already seen that AIs improving code can remove its essential features, so it wouldn't be surprising for some security checks to disappear.
Based on this, I've decided to upvote this post weakly only.
I suggest additional explanation.
The bigger the audience is, the more people there are who won't know a specific idea/concept/word (xkcd's comic #1053 "Ten Thousand" captures this quite succinctly), so you'll simply have to shorten.
I took logarithm of sentence length and linearly fitted it against logarithm of world population (that shouldn't really be precise since authors presumably mostly cared about their society, but that would be more time-expensive to check).
Relevant lines of Python REPL
>>> import math
>>> wps = [49, 50, 42, 20, 21, 14, 18, 12]
>>> pop = [600e6, 700e6, 1e9, 1.4e9, 1.5e9, 2.3e9, 3.5e9, 6e9]
>>> [math.log(w) for w in wps]
[3.8918202981106265, 3.912023005428146, 3.7376696182833684, 2.995732273553991, 3.044522437723423, 2.6390573296152584, 2.8903717578961645, 2.4849066497880004]
>>> [math.log(p) for p in pop]
[20.21244021318042, 20.36659089300768, 20.72326583694641, 21.059738073567623, 21.128730945054574, 21.556174959881517, 21.97602880544178, 22.515025306174465]
>>> 22.51-20.21
2.3000000000000007
>>> 3.89-2.48
1.4100000000000001
>>> 2.3/1.41
1.6312056737588652
>>> [round(math.exp(26.41 - math.log(w)*1.63)/1e9, 3) for w,p in zip(wps,pop)] # predicted population, billion
[0.518, 0.502, 0.667, 2.234, 2.063, 3.995, 2.652, 5.136]
>>> [round(math.exp(26.41 - math.log(w)*1.63)/1e9 - p/1e9, 3) for w,p in zip(wps,pop)] # prediction off by, billion
[-0.082, -0.198, -0.333, 0.834, 0.563, 1.695, -0.848, -0.864]
Given the gravity of Sam Altman's position at the helm of the company leading the development of an artificial superintelligence which it does not yet know how to align -- to imbue with morality and ethics -- I feel Annie's claims warrant a far greater level of investigation than they've received thus far.
Then there's a bit of shortage of something public... and ratsphere-adjacent... maybe prediction markets?
I can create (subsidize) a few, given resolution criteria.
This concludes my April the 1st!
That's screened off by actual evidence, which is, top labs don't publish much no matter where they are, so I'd only agree with "equally opaque".