Regarding "the joy of being read", I think that an AI reader can only provide that feeling for a human writer to the extent that the human feels, on a gut level, that the AI is a person. If they don't feel that way, then I think it wouldn't satisfy the human's social drives that cause them to enjoy it when people read their writing. And I'm not sure how many humans feel like current AIs are people.
About the headings on the transcripts - it looks like each heading is at the end of the corresponding section of the transcript, instead of being at the beginning of the section? That seems like a mistake.
Based on context, I presume there are supposed to be images (screenshots) in your post, but I can't see them? I just see empty newlines. (e.g. I can't see what "I asked it this" refers to.)
Some countries have laws against (anti-gay) conversion therapy.
It sounds like inter-branch communication would imply that you could do that. OP does mention that as an application:
7. Superintelligence jumps from one branch to another and also gains more computational power via branch-distributed computations. The exchange of ideas between branches would accelerate science and progress in AI algorithms. This all means that inter-branch communication and the singularity would happen almost simultaneously.
Superintelligent AI will likely appear before this technology matures in the current setup, but branch communication could help AI propagate between branches and increase its dominance on Earth and in the multiverse.
Yeah, that model sounds plausible to me (pending elaboration on how the friend-or-enemy parameter is updated). Thanks.
Thanks for writing this series.
I can see how Approval Reward explains norm-following behavior. If people approve of honesty, then being honest will make people approve of me.
But I'm not totally convinced that Approval Reward is enough to explain norm-enforcement behavior on its own?
For some action or norm X, it doesn't seem obvious to me that "doing X" and "punishing someone who does not-X" are equivalent in terms of human-approving, unless you already knew that humans punish others who do things they don't like.
If you know that humans punish others who ac...
I can only see the first image you posted. It sounds like there should be a second image (below "This is not happening for his other chats:") but I can't see it.
In the graphic in section 3.5.2, you mention
Groups that are plausibly aggressive enough to unilaterally (and flagrantly-illegally) deliberately release a sovereign AI into the wild
What kind of existing laws were you thinking that this would violate? When you said "into the wild", are you thinking of sending the AI to copy itself across the internet (by taking over other peoples' computers), and this would violate laws about hacking? If the AI was just accessing websites like a human, or if the AI had a robot body and it went out onto the streets, I can't i...
James Miller discussed similar ideas.
The "ideas" link doesn't seem to work.
About the example in section 6.1.3: Do you have an idea of how the Steering Subsystem can tell that Zoe is trying to get your attention with her speech? It seems to me like that requires both (a) identifying that the speech is trying to get someone's attention, and (b) identifying that the speech is directed at you. (Well, I guess (b) implies (a) if you weren't visibly paying attention to her beforehand.)
About (a): If the Steering Subsystem doesn't know the meaning of words, then how can it tell that Zoe is trying to get someone's attention? Is there some ...
I couldn't click into it from the front page if I tried to click on the zone where the text content would normally go, but I was able to click into this from the front page if I clicked on the reply-count icon in the top-right corner. (But that wouldn't have worked when there were zero replies.)
The UK government also heavily used AI chatbots to generate diagrams and citations for a report on the impact of AI on the labour market, some of which were hallucinated.
This link is broken.
Thank you for writing this series.
I have a couple of questions about conscious awareness, and a question about intuitive self-models in general. They might be out-of-scope for this series, though.
Questions 1 and 2 are just for my curiosity. Question 3 seems more important to me, but I can imagine that it might be a dangerous capabilities question, so I acknowledge you might not want to answer it for that reason.
On second thought, there must be a decent number of people for whom LLMs can fulfil social roles, given the people with AI boy/girlfriends, etc. I guess I was just replying to my own feeling that an LLM reading my writing (if I did writing) would not satisfy me, and so they wouldn't "solve the problem of unread writing" for me, but they could "solve the problem" for some people.