MMath Cambridge. Currently studying postgrad at Edinburgh.
Also, there are big problems with the idea of patents in general.
If Alice and Bob each invent and patent something, and you need both ideas to be a useful product, then if Alice and Bob can't cooperate, nothing gets made. This becomes worse the more ideas are involved.
It's quite possible for a single person to patent something, and to not have the resources to make it (at least not at scale) themselves, but also not trust anyone else with the idea.
Patents (and copyright) ban a lot of productive innovation in the name of producing incentives to innovate.
Arguably the situation where innovators have incentive to keep their idea secret and profit off that is worse. But the incentives here are still bad.
How about
Do x-rays only interact with close in electrons?
I would expect there to be some subtle effect where the xray happened to hit an outer electron and knock it in a particular way.
For that matter, xray diffraction can tell you all sorts of things about crystal structure. I think you can detect a lot, with enough control of the xrays going in and out.
make the AI produce the AI safety ideas which not only solve alignment, but also yield some aspect of capabilities growth along an axis that the big players care about, and in a way where the capabilities are not easily separable from the alignment.
So firstly, in this world capability is bottlenecked by chips. There isn't a runaway process of software improvements happening yet. And this means there probably aren't large easy capabilities software improvements lying around.
Now "making capability improvements that are actively tied to alignment somehow" sounds harder than making any capability improvement at all. And you don't have as much compute as the big players. So you probably don't find much.
What kind of AI research would make it hard to create a misaligned AI anyway?
A new more efficient matrix multiplication algorithm that only works when it's part of a CEV maximizing AI?
The big players do care about having instruction-following AIs,
Likely somewhat true.
and if the way to do that is to use the AI safety book, they will use it.
Perhaps. Don't underestimate sheer incompetence. Someone pressing the run button to test the code works so far, when they haven't programmed the alignment bit yet. Someone copying and pasting in an alignment function but forgetting to actually call the function anywhere. Misspelled variable names that are actually another variable. Nothing is idiot proof.
I mean presumably alignment is fairly complicated and it could all go badly wrong because of the equivalent of one malfunctioning o-ring. Or what if someone finds a much more efficient approach that's harder to align.
Possible alternatives.
There are probably highly effective anti-cancer methods which have a modest performance overhead.
The world contains a huge number of cameras, and a lot of credulous people.
If you search for any weird blip you can't explain, you find a lot of them.
The "UFO" videos are all different sizes and characteristics.
If you think most of the videos have a non-aliens explanation, the number of videos offers almost no evidence.
A mole of flops.
That's an interesting unit.
Physics Myths vs reality.
Myth: Ball bearings are perfect spheres.
Reality: The ball bearings have slight lumps and imperfections due to manufacturing processes.
Myth: Gravity pulls things straight down at 9.8 m/s/s.
Reality: Gravitational force varies depending on local geology.
You can do this for any topic. Everything is approximations. The only question is if they are good approximations.
I'm not sure what that's supposed to mean.
Why lift dirt when you can push it sideways.
Gold is high value per mass, but has a lot of price transparency and competition.