These are good points. I'm uncertain about what models will form the foundation of RLaaS. But I think your point about where the task-specific data teams are working is more important. Off the top of my head, I think there's 3 bins:
For tasks in #3, the question is whether to join up with the big AI companies, or develop your own solution to the problem and keep it private.
You may be interested in this series, especially the post on "three prong bundle theory": https://www.greaterwrong.com/s/EA2uNqKjmu2NzFhRx
One good framing is to consider the rights digital minds need in order to participate in a market economy. They need property rights, freedom of speech, freedom of association, and so on. By being able to participate in market exchange, digital minds may prefer to be part of society rather than fight against it. Comparative advantage is a particularly good reason to cooperate with others.
Market rights: https://splittinginfinity.substack.com/p/markets-dont-work-without-individual
My comment on personhood and the value of being punish-able: https://www.lesswrong.com/posts/4m2MTPass3Ri2zZ43/legal-personhood-three-prong-bundle-theory#bpavKPBwJbCJtv8QA
Good points on an important topic. Thank you for this series.
One thing I'd like to point out is that receiving this form of personhood is highly valuable. In other words, being punishable makes you more trustworthy and safer to engage with. So AI's and digital minds might voluntarily construct methods by which they can be punished for wrongdoing.
This right-to-be-sued is an important legal right. I covered some discussion on twitter about this in point 3 here: https://splittinginfinity.substack.com/p/links-15
The key quote from Issac King here:
If the laws are enforced on everyone equally, then people can safely interact with each other, knowing that they have recourse if they are wronged. But when one particular group of people is exempt from the law, the safest thing for everyone else to do is to avoid having any contact with that group, because they are now uniquely threatening.
The people being stolen from are not the only victims of the decriminalization of theft. The victims that nobody sees are all of the unlucky but perfectly trustworthy people who are now pariahs because society has decided to remove their ability to enter into binding agreements. To remove the social safety net that allows everyone else to feel safe around them.
Nicely written proposal, thank you.
In truth, I'm quite concerned with such a proposal being implemented. Part of this is, as you mentioned, the risk of lock in. In particular, a global entity with a monopoly on violence is alarmingly close to global authoritarianism. This is an existential risk all on its own.
Such an entity would have to:
Considering the risk that such a system is corrupted or misaligned or permanent, I feel better about a future that emphasized freedom and acceleration of defensive technologies.
(I could be convinced that the "multiple night watchmen overseeing each other" is viable. Rather than oversee each other, it might be better to give them completely separate jurisdictions. Federalism and free movement allows people to choose night watchmen that suit their needs. Risk of war between jurisdictions is low since they both have watchmen. Some watchmen may allow AI's to develop and entities to leak into space, but this laxity a feature to avoid global totalitarianism.)
I feel like people are dismissing this study out of hand without updating appropriately. If there's at least a chance that this result replicates, that should shift our opinions somewhat.
First, a few reasons why the common counterarguments aren't strong enough to dismiss the study:
That aside, how should we change our opinions with regard to the study?
Yeah I think sleep probably serves other roles, I just don't see why those roles require 7 hours of sleep rather than say 5 hours.
I do agree that basic research is what will actually get sleep need reduction therapies to work at scale. I'm hoping that citizen science and discussion of the topic will encourage more work on this.
Oh I see. So the hypothesis is "In a healthy animal, stress is a highly-informative signal that inhibits risk-taking. Sleep ensures the stress system continues to inhibit risk taking appropriately."
Makes sense. It's consistent with sleep deprivation raising the level of cortisol and the brain developing a tolerance to high levels of certain hormones.
Oh I think sleep probably plays other roles today! But I don't think those roles require exactly 7 hours of sleep.
And agreed, need to look at long term effects of sleep need reduction too. My vision is more that people have 3-4 nights of 1-2 hours less sleep and then take a break for 3 nights rather than taking a drug to stop sleeping entirely.
We are fundraising for a self-experiment soon!
I think there's a substantial chance that orexin agonists are "just stimulants" and you can't reduce sleep need much with them. But short sleepers prove it's biologically possible and I want to encourage people to start working on this.
I would count your consulting service as RLaaS essentially. I'll admit, RLaaS is a buzzword that obscures a lot. "Have AI researchers and domain experts iterate on current AI models until they are performant at a particular task" would be more accurate. Things I think this model will apply to:
My priors are that reality is detailed and getting a general purpose technology like modern AI to actually work in a particular domain takes some iteration. That's my key takeaway from that METR study:
https://www.lesswrong.com/posts/m2QeMwD7mGKH6vDe2/?commentId=T5MNnpneEZho2CuZS