All of P.'s Comments + Replies

My vision of a good future, part I

Rot13: Vf gung cvrpr bs svpgvba Crefba bs Vagrerfg be Png Cvpgherf Cyrnfr?

2Jeffrey Ladish1mo
Rot13: Ab vg'f Jbegu gur Pnaqyr
Who is this MSRayne person anyway?

Then you should at least try to talk to 80,000 hours, you might eventually relocate somewhere where meeting people is easier.

It wasn't intended to make fun of you. When I say that you shouldn't start a religion I mean it literally, like most people here I don't hold a favorable view of religions.

Sentences like "But I am fundamentally a mystic, prone to ecstatic states of communion with an ineffable divine force immanent in the physical universe which I feel is moving towards incarnating as an AI god such as I called Anima" make me think that what you are t... (read more)

3Aleksi Liimatainen1mo
There's a trivial sense in which this is false: any experience or utterance, no matter how insensible, is as much the result of a real cognitive process as the more sensible ones are. There's another, less trivial sense that I feel is correct and often underappreciated: obfuscation of correspondence does not eliminate it. The frequency by which phenomena with shared features arise or persist is evidence of shared causal provenance, by some combination of universal principles or shared history. After puzzling over the commonalities found in mystical and religious claims, I've come to see them as having some basis in subtle but detectable real patterns. The unintelligibility comes from the fact that neither mystics nor their listeners have a workable theory to explain the pattern. The mystic confabulates and the listener's response depends on whether they're able to match the output to patterns they perceive. No match, no sense.
8MSRayne1mo
Yeah, that's the problem. I'm not a rationalist and this makes it obvious that I don't want to / can't be, and shouldn't be here. There are beliefs I would not update if an endless amount of evidence against them came my way, because they are how I keep myself from suicidal ideation. The belief that I have a unique spiritual vision etc is one of them. It is one of the pillars of my pretense of sanity. It is very frightening even to talk about this with anyone because the smallest poke at it sends me into spirals of depression and anxiety such as I felt for several hours (to the point of having to nap because I couldn't stand being awake any longer) as a result of just writing this post - even before you replied! It's actually almost physically sickening to be paid attention to like that. I'm not a mentally healthy person. I feel like everyone who reads this must think I'm stupid and pathetic and disgusting and childish and a lunatic, and I regret joining this site. And the worst thing is that I know I will regret even more it if I burn bridges yet again and leave - that the depression demon is what is making me believe these things, and they are not necessarily true - but I can't stop myself from wanting to because it hurts that much to be spoken to with simple blunt honesty as you have done. Sorry. This isn't your problem. You're just a commenter on a post.
Who is this MSRayne person anyway?

Have you considered:

Trying to find IRL friends through meetup.com

Going to nearby rationality meetups (https://www.lesswrong.com/community)

Using dating apps (and photofeeler.com)

Getting free career advice for your situation through 80,000 hours (https://80000hours.org/speak-with-us/)

Writing a few pages of your book and posting them on LW (but please don't start a religion)

?

2MSRayne1mo
Hey, sorry, this was actually a very good comment and set of suggestions and not-depressed-me thanks you for it. Depressed-me unfortunately overreacts to small triggers, as you've seen, but rereading this now I can tell you were just trying to help.
-4Flaglandbase1mo
Good advice but I recommend against dating apps unless you look like a celebrity. * EDIT: of course the above advice against dating sites only applies if you're male.
2MSRayne1mo
I live out in the middle of nowhere and going places is not generally an easy thing to do for me. Anything that requires that is kind of not viable right now. Dating is pointless. I'm not good looking and again, there's no way to go out and be around people. I'm already regretting writing this post. The "please don't start a religion" part of your comment to me feels like it was intended to make fun of or insult me. I didn't come here to be lectured by reddit atheists. And if that's the reaction I'll get for sharing something so important to me, I WON'T post anything further about it on this site.
Has anyone actually tried to convince Terry Tao or other top mathematicians to work on alignment?

The email to Demis has been there since the beginning, I even received feedback on it. I think I will send it next week, but will also try to get to him through some DeepMind employee if that doesn’t work.

Please don't send these types of emails, I expect that they're actively counterproductive for high-profile recipients.

If you want to outreach, there are clear channels which should be used for coordinating it. For example, you could contact DeepMind's alignment team, and ask them if there's anything which would be useful for them.

Scott Aaronson is joining OpenAI to work on AI safety

He says he will be doing alignment work, the worst thing I can think of that can realistically happen is that he gives OpenAI unwarranted confidence in how aligned their AIs are. Working at OpenAI isn’t intrinsically bad, publishing capabilities research is.

1[comment deleted]2mo
Scott Aaronson is joining OpenAI to work on AI safety

Thanks, I’ve added him to my list of people to contact. If someone else wants to do it instead, reply to this comment so that we don’t interfere with each other.

FYI: I’m working on a book about the threat of AGI/ASI for a general audience. I hope it will be of value to the cause and the community

No offense, but It's not obvious to me why communicating to a general audience could be a net positive. Exactly how do you expect this to help?

2Darren McKee2mo
None taken, it's a reasonable question to ask. It's part of the broader problem of knowing if anything will be good or bad (unintended consequences and such). To clarify a bit, by general audience, I don't mean everyone because most people don't read many books, let alone non-fiction books, let alone non-fiction books that aren't memoirs/biographies or the like. So, my loose model is that (1) there is a group of people who would care about this issue if they knew more about it and (2) their concerns will lead to interest from those with more power to (3) increase funding for AI safety and/or governance that might help. Expanding on 1, it could also increase those who want to work on the issue, in a wide range of domains beyond technical work. It's also possible that it is net-positive but still insufficient but was worth trying.
Can you MRI a deep learning model?
Answer by P.Jun 13, 20222

Most neural networks don’t have anything comparable to specialised brain areas, at least structurally, so you can’t see which areas light up given some stimulus to determine what that part does. You can do it with individual neurons or channels, though. The best UI I know of to explore this is the “Dataset Samples” option in the OpenAI Microscope, that shows which inputs activate each unit.

Has anyone actually tried to convince Terry Tao or other top mathematicians to work on alignment?

Please do! You can DM me their contact info, tell them about my accounts: either this one or my EA Forum one, or ask me for my email address.

Has anyone actually tried to convince Terry Tao or other top mathematicians to work on alignment?

Well, if he has, unbeknownst to me, already hired the “Terence Taos of the world” like he said on the podcast, that would be great, and I would move on to other tasks. But if he only has a regular alignment team, I don’t think either of us considers that to be enough. I’m just trying to convince him that it’s urgent and we can’t leave it for later.

Has anyone actually tried to convince Terry Tao or other top mathematicians to work on alignment?

I think I get it, but even if I didn’t now I know that’s how it sounds, and I think I know how to improve it. That will be for other mathematicians though (at least Maxim Kontsevich), see the edit to the post. Does the tone in the email to Demis seem like the right one to you?

2TekhneMakre2mo
In terms of tone it seems considerably less bad. I definitely like it more than the other one because it seems to make arguments rather than give social cues. It might be improved by adding links giving technical descriptions about the terms you use (e.g. inner alignment (Hubinger's paper), IRL (maybe a Russell paper on CIRL)). I still don't think it would work, simply because I would guess Hassabis gets a lot of email from randos who are confused and the email doesn't seem to distinguish you from that (this may be totally unfair to you, and I'm not saying it's correct or not, it's just what I expect to happen). I also feel nervous about talking about arms races like that, enforcing a narrative where they're not only real but the default (this is an awkward thing to think because it sounds like I'm trying to manage Hassabis's social environment deceptively, and usually I would think that worrying about "reinforcing narratives" isn't a main thing to worry about and instead one should just say what one thinks, but, still my instincts say to worry about that here, which might be incorrect).
Has anyone actually tried to convince Terry Tao or other top mathematicians to work on alignment?

I’ll rewrite it but I can’t just model other people after me. If I were writing it for someone like myself it would be a concise explanation of the main argument to make me want to spend time thinking about it followed by a more detailed explanation or links to further reading. As long as it isn’t mean I don’t think I would care if it’s giving me orders, begging for help or giving me information without asking for anything at all. But he at least already knows that unaligned AIs are a problem, I can only remind him of that, link to reading material or say ... (read more)

Has anyone actually tried to convince Terry Tao or other top mathematicians to work on alignment?

Do you mean website links about his plan? I found nothing.

I’m still not changing the deadlines but I’ve received information that made me want to change the order.

5Greg C2mo
No I mean links to him in person to talk to him (or for that matter, even an email address or any way of contacting him..).
Has anyone actually tried to convince Terry Tao or other top mathematicians to work on alignment?

Then do you think I should contact Jacob Steinhardt to ask him what I should write to interest Tao and avoid seeming like a crank?

There isn’t much I can do about SA other than telling him to work on the problem in his free time.

Unless something extraordinary happens I’m definitely not contacting anyone in politics. Politicians being interested in AGI is a nightmarish scenario and those news about Huawei don’t help my paranoia about the issue.

6SP2mo
I personally think the probability of success would be maximized if we were to first contact high-status members of the rationalist community, get them on board with this plan, and ask them to contact Scott Aaronson as well as contact professors who would be willing to contact other professors. The link to Scott Aaronson's blog says he provisionally would be willing to take a leave of absence from his job to work on alignment full-time for a year for $500k. I believe EA has enough funds that they could fund that if they deemed it to be worthwhile. I think the chance of success would be greatest if we contacted Eliezer and/or whoever is in charge of funds, asked them to make Scott a formal offer, and sent Scott an email with the offer and an invitation to talk to somebody (maybe Paul Christiano, his former student) working on alignment to see what kinds of things they think are worth working on. I think even with the perfect email from most members of this community, the chances that e.g. Terry Tao reads it, takes it seriously, and works on alignment are not very good, due to lack of easily verifiable credibility of the sender. Institutional affiliation at least partly remedies this, and so I think it would be preferable if an email came from another professor who directly tried to convince them. I think cold-emailing Jacob Steinhardt/Robin Hanson/etc. asking them to email other academics would have a better chance of succeeding given that the former indeed participate on this forum. However, even here, I think people are inclined to pay more attention to the views of those closer to them. My impression is that Eliezer and other high-ranked members of the rationalist community have closer connections to these alignment-interested professors (and know many more such professors) and could more successfully convince them to reach out to their colleagues about AI safety. I don't mean to suggest that these less-direct ways are necessarily better. If for instance Eliez
Has anyone actually tried to convince Terry Tao or other top mathematicians to work on alignment?

I might try that, but “community health” is not really what I’m optimising for. Maybe the name is misleading?

2Chris_Leong2mo
I don't think it's that surprising. Tasks in an organisation will naturally go to the team that has the most relevant experience.
Has anyone actually tried to convince Terry Tao or other top mathematicians to work on alignment?

I was trying to rely on Tao’s trust in Demis's judgement, since he is an AI researcher. Mentioning Eliezer is mainly so he has someone to contact if he wants to get hired.

I wanted his thinking to be “this competent entity has spent some of his computational resources verifying that it is important to solve this problem, and now that I’m reminded of that I should also throw mine at it”.

Is he truly mostly interested in what he considers to be mentally stimulating? Not in improving the world, or in social nonsense, or guaranteeing that his family is completel... (read more)

2TekhneMakre2mo
Does the stuff about pushiness make sense to you? What do you think of it? I think as is, the letter, if Tao reads it, would be mildly harmful, for the reasons described by other commenters.
6mukashi2mo
If you plan to rewrite that letter with a less pushy tone (I agree 100% with the comment from TechneMakre) I think it might be useful if you try to change the framework of the problem a bit. Imagine that a random guy is writing to you instead, and he is telling you to work on deviating possible meteorites reaching Earth. What sort of email would make you compelled to reply?
Has anyone actually tried to convince Terry Tao or other top mathematicians to work on alignment?

That’s part of the point of the post, to coordinate so that fewer emails are sent. I asked if anyone tried something similar and asked people not to send their own emails without telling the rest of us.

Announcing the Alignment of Complex Systems Research Group

You wrote this twice:

From here, protecting humans and their values looks different than in the classic single/single scenario. On the other hand, it seems quite likely that there is a deep connection between safety issues in the "AI in a box" case, and in the ecosystems case. 

From here, protecting humans and their values looks different than in the classic single/single scenario. On the other hand, it seems quite likely that there is a deep connection between safety issues in the "AI in a box" case, and in the ecosystems case.

2Jan_Kulveit2mo
Thanks, removed
Have You Tried Hiring People?

Did you change your mind about Demis being reasonable or a petition being a good idea? And why? I’m considering sending him an email.

Six Dimensions of Operational Adequacy in AGI Projects

Spoiler Warning. Tried hiding it with >! and :::spoiler but neither seems to work.

For those unaware, this is a story (worth reading) about anti-memes, ideas that cannot be spread, so researchers have a hard time working with them, not knowing they exist. So the point of the parent comment probably is that even if an adequate AGI project existed we wouldn’t know about it.

I don't think any AGI projects that exist today are anywhere near adequacy on all of these dimensions.

(And I don't think we live in the sort of world where there could plausibly be a secret-to-me adequate project, right now in 2022.)

I could imagine optimistic worlds where an adequate project exists in the future but I or others publicly glomarize about whether it's adequate (e.g., because we don't want to publicly talk about the state of the project's opsec or something)? Mostly, though, I think we'd want to loudly and clearly broadcast 'this is the Good P... (read more)

6lc3mo
Prize for Alignment Research Tasks

Task: Computing the relevance of a paper to solving a problem

Context: A researcher is looking at existing literature trying to solve some problem. However, it is not obvious what to search for. There might exist research whose relevance can only be determined once they read it and think about it. For this task a simple keyword-matching search engine isn’t useful.

Input type: A problem statement and an URL to a website or PDF. Optionally, text with reasoning about the problem and what has been tried before. The URLs in the training set can be papers that hav... (read more)

What would be the impact of cheap energy and storage?
Answer by P.May 03, 20225

You might like this video (from 9:18) by Isaac Arthur that talks about that

[$20K in Prizes] AI Safety Arguments Competition

To the extent that rationality has a purpose, I would argue that it is to do what it takes to achieve our goals, if that includes creating "propaganda", so be it. And the rules explicitly ask for submissions not to be deceiving, so if we use them to convince people it will be a pure epistemic gain.

Edit: If you are going to downvote this, at least argue why. I think that if this works like they expect, it truly is a net positive.

1hath4mo
Fair. Should've started with that. I think there's a difference between "rationality is systematized winning" and "rationality is doing whatever it takes to achieve our goals". That difference requires more time to explain than I have right now. I think that the whole AI alignment thing requires extraordinary measures, and I'm not sure what specifically that would take; I'm not saying we shouldn't do the contest. I doubt you and I have a substantial disagreement as to the severity of the problem or the effectiveness of the contest. My above comment was more "argument from 'everyone does this' doesn't work", not "this contest is bad and you are bad". Also, I wouldn't call this contest propaganda. At the same time, if this contest was "convince EAs and LW users to have shorter timelines and higher chances of doom", it would be reacted to differently. There is a difference, convincing someone to have a shorter timeline isn't the same as trying to explain the whole AI alignment thing in the first place, but I worry that we could take that too far. I think that (most of) the responses John's comment got were good, and reassure me that the OPs are actually aware of/worried about John's concerns. I see no reason why this particular contest will be harmful, but I can imagine a future where we pivot to mainly strategies like this having some harmful second-order effects (which need their own post to explain).
[$20K in Prizes] AI Safety Arguments Competition

To executives and researchers:

However far you think we are from AGI, do you think that aligning it with human values will be any easier? For intelligence we at least have formalisms (like AIXI) that tell us in principle how to achieve goals in arbitrary environments. For human values on the other hand, we have no such thing. If we don't seriously start working on that now (and we can, with current systems or theoretical models), there is no chance of solving the problem in time when we near AGI, and the default outcome of that will be very bad, to say the least.

Edit: typos

Calling for Student Submissions: AI Safety Distillation Contest

I have a few questions:

Do we need to study at a US university in order to participate? I’m in Europe.

Who should be the target audience for the posts? CS students? The average LW reader? People somewhat interested in AI alignment? The average Joe? How much do we need to dumb it down?

Can we publish the posts before the contest ends?

Will you necessarily post the winners' names? Can we go by a pseudonym instead?

How close to the source material should we stay? I might write a post about what value learning is, why it seems like the most promising approach and w... (read more)

2Aris4mo
You can participate from any country! The submission should just be in English (but, if you'd like to additionally submit in another language, that could be cool to increase accessibility). The target audience would ideally be a few levels lower than the audience of the original piece. If it was originally written for alignment researchers, then making the paper accessible to a CS student is good, but making it accessible to people who have never heard of alignment could be even better (although certainly harder). This is what the "accessibility" scoring point would be focusing on, but it's far from the only factor. You can publish the post before the contest ends, but please indicate that the post is intended to be part of the contest! We'd ideally like to be encouraging people to create content they wouldn't have without the contest. A pseudonym works! Publishing names would mainly be to help students build social capital, so it's an optional part of winning. I think that as long as you faithfully explain parts of the original posts you would be very welcome to add to the discussion. (Would also love to hear answers to the last question!)
Are deference games a thing?

How do you propose to have many votes every day? Voting that often (on laws or representatives) would probably significantly lower the quality of the votes. It would also force most people to delegate to others that do it full time; people that don’t have enough time to vote and don’t want to delegate won’t be happy with that. And what infrastructure would you use? The internet? That doesn’t seem safe.

What I Was Thinking About Before Alignment

I mean, just in case I wasn’t clear enough, you want a program that takes in a representation of some system and outputs something a human can understand, right? But even if you could automatically divide a system into a tree of submodules such that a human could in principle describe how any one works in terms of short descriptions of the function of its submodules, there is no obvious way of automatically computing those descriptions. So if you gave a circuit diagram of a CPU as the input to that universal translator, what do you want it to output?

DALL·E 2 by OpenAI

You replied to the wrong comment.

1Caridorc Tergilti4mo
Thanks, added the comment in the correct place now.
DALL·E 2 by OpenAI

My mom wants to see what this would look like: "An ant on a scooter on an interplanetary flight". Let's see how many concepts it can combine at once.

I could see your mother having a promising career in QA

What I Was Thinking About Before Alignment

What would such a representation look like for a computer? There might exist some method for computing how the circuits are divided into modules and submodules, but how would you represent what they do? You don’t expect it to be annotated in natural language, do you?

3P.4mo
I mean, just in case I wasn’t clear enough, you want a program that takes in a representation of some system and outputs something a human can understand, right? But even if you could automatically divide a system into a tree of submodules such that a human could in principle describe how any one works in terms of short descriptions of the function of its submodules, there is no obvious way of automatically computing those descriptions. So if you gave a circuit diagram of a CPU as the input to that universal translator, what do you want it to output?
DALL·E 2 by OpenAI

It seems they edited the paper, where now there is a close up of a handpalm with leaves growing from it, there once was a super saiyan sentient bag of potato chips:

0Caridorc Tergilti4mo
Let's have fun with recursion! A checkerboard where each square is itself a checkerboard. A cube with mirrors on both sides, the mirrors show multiple reflections of the cube. A person wearing a shirt with an image of that person wearing that shirt.
DALL·E 2 by OpenAI

We shouldn't be surprised by the quality of the images, but since this will become a commercial product and art is something that is stereotypically hard for computers, I wonder if for the general public (including world governments) this will be what finally makes them realize that AGI is coming. OpenAI could have at least refrained from publishing the paper, it wouldn’t have made any difference but would have been a nice symbolic gesture.

MIRI announces new "Death With Dignity" strategy

ELK itself seems like a potentially important problem to solve, the part that didn't make much sense to me was what they plan to do with the solution, their idea based on recursive delegation.

5Buck4mo
Ok, that's a very reasonable answer.
MIRI announces new "Death With Dignity" strategy

I will probably spend 4 days (from the 14th to the 17th, I’m somewhat busy until then) thinking about alignment to see whether there is any chance I might be able to make progress. I have read what is recommended as a starting point on the alignment forum, and can read the AGI Safety Fundamentals Course’s curriculum on my own. I will probably start by thinking about how to formalize (and compute) something similar to what we call human values, since that seems to be the core of the problem, and then turning that into something that can be evaluated ov... (read more)

3Chris_Leong4mo
Sounds like a great plan! Even if you end up deciding that you can't make research progress (not that you should give up after just 4 days!), I can suggest a bunch of other activities that might plausibly contribute towards this. I expect that this will change within the next year or so (for example, there are plans for a Longtermist Hotel in Berlin and I think it's very likely to happen).
Unfinished Projects Thread

Neural style transfer algorithms that rely on optimizing an image by gradient descent are painfully slow. You can make them faster by training a network to map the original image to a stylized one that minimizes the loss. I made them even faster by stacking a few Perlin noise channels to the input, using a very fast hand-written edge detection algorithm and stacking that too and then performing only local processing. Unfortunately this gives very poor results and is only not uselessly slow for low-res inputs.

So I gave up on that and started reading about n... (read more)

MIRI announces new "Death With Dignity" strategy

I think going to congress would be counterproductive and would just convince them to create an AGI before their enemies do.

MIRI announces new "Death With Dignity" strategy

For whatever it is worth, this post along with reading the unworkable alignment strategy on the ELK report has made me realize that we actually have no idea what to do and has finally convinced me to try to solve alignment, I encourage everyone else to do the same. For some people knowing that the world is doomed by default and that we can't just expect the experts to save it is motivating. If that was his goal, he achieved it.

3Aditya2mo
Same this post is what made me decide I can't leave it to the experts. It is just a matter of spending the required time to catch up on what we know and tried. As Keltham said - Diversity is in itself an asset. If we can get enough humans to think about this problem we can get some breakthroughs many some angles others have not thought of yet. For me, it was not demotivating. He is not a god, and it ain't over until the fat lady sings. Things are serious and it just means we should all try our best. In fact, I am kinda happy to imagine we might see a utopia happen in my lifetime. Most humans don't get a chance to literally save the world. It would be really sad if I died a few years before some AGI turned into a superintelligence.

I very much had the same experience, making me decide to somewhat radically re-orient my life.

5Buck4mo
What part of the ELK report are you saying felt unworkable?
5Chris_Leong4mo
Awesome. What are your plans? Have you considered booking a call with AI Safety Support [https://www.aisafetysupport.org/], registering your interest for the next AGI Safety Fundamentals Course [https://www.eacambridge.org/agi-safety-fundamentals] or applying to talk to 80,000 hours [https://80000hours.org/speak-with-us/#:~:text=If%20you%20have%20an%20urgent,email%20coaching%4080000hours.org.] ?

Certainly for some people (including you!), yes. For others, I expect this post to be strongly demotivating. That doesn’t mean it shouldn’t have been written (I value honestly conveying personal beliefs and are expressing diversity of opinion enough to outweigh the downsides), but we should realistically expect this post to cause psychological harm for some people, and could also potentially make interaction and PR with those who don’t share Yudkowsky’s views harder. Despite some claims to the contrary, I believe (through personal experience in PR) that ex... (read more)

Replacing Karma with Good Heart Tokens (Worth $1!)

Oh, and another bad faith strategy consists on writing many ok comments instead of just a few good ones, the Good Heart Tokens a user receives aren’t actually proportional to the total value added. Please keep the score of this one at 1.

Replacing Karma with Good Heart Tokens (Worth $1!)

I upvoted you (+1), but the point of my comment was to make sure that someone besides the author of a comment reads it so that the score more closely reflects its real value. If you do it recursively then we will have two kinds of comments to consider: replies to my comment and to your, so it will become harder to make comparisons.

I guess I should have said something like “and if you up/downvote a reply in this comment thread, please read and score all of its siblings accordingly”, since if it's just me upvoting the problem will persist. I can only give on... (read more)

1P.4mo
Oh, and another bad faith strategy consists on writing many ok comments instead of just a few good ones, the Good Heart Tokens a user receives aren’t actually proportional to the total value added. Please keep the score of this one at 1.

Why the downvotes? This is RSA-2048, which is in fact composite and is the product of two unknown large primes. Judging by the self-replies of goodestheart, this is what they are looking for, a non-prime that looks like a prime. It might be too well known or there might exist some simple primality test that it fails, but numbers like this one seem to be close to the perfect answer to the question.

4James Payor4mo
It's a great answer, but I think folks are turning up to defend the Good Hearts from being Goodharted

25195908475657893494027183240048398571429282126204032027777137836043662020707595556264018525880784406918290641249515082189298559149176184502808489120072844992687392807287776735971418347270261896375014971824691165077613379859095700097330459748808428401797429100642458691817195118746121515172654632282216869987549182422433637259085141865462043576798423387184774447920739934236584823824281198163815010674810451660377306056201619676256133844143603833904414952634432190114657544454178424020924616515723350778707749817125772467962926386356373289912154831438167899885040445364023527381951378636564391212010397122822120720357

1P.4mo
Why the downvotes? This isRSA-2048 [https://en.wikipedia.org/w/index.php?title=RSA_numbers#RSA-2048], which is in fact composite and is the product of two unknown large primes. Judging by the self-replies of goodestheart, this is what they are looking for, a non-prime that looks like a prime. It might be too well known or there might exist some simple primality test that it fails, but numbers like this one seem to be close to the perfect answer to the question.
Replacing Karma with Good Heart Tokens (Worth $1!)

Given the flood of comments that will inevitably result from this, it might be hard to get noticed and to surface the best ones to the top. So I am offering the following service: If you reply to this I guarantee that I will read your comment, and then will give you one or two upvotes (or none) depending on how insightful I consider it to be. Sadly, this only works if people get to see this comment, so it is in your best interest to upvote it. Let’s turn this into a new, better comment section!

5abramdemski4mo
This was very clever, so I strong-upvoted it, but it also feels like it violates good faith, so I then strong-downvoted it.
3Simon Skade4mo
Agreed, since now many people will probably comment in this thread, I make the same recursive offer: So please upvote this comment so it stays on top of this comment thread!
If AGI were coming in a year, what should we do?

I interpreted what they said as "As the pandemic clearly demonstrated, given even a minor crisis people will act selfishly and turn against each other, if the singularity were 2 years away it would be a bloodbath" instead of "The pandemic will kill most of us in less than 2 years".

3shminux4mo
What I meant was that people do not want to inconvenience themselves for a long time. After a while they will accept an elevated risk of death now rather than spending more time safe but mildly inconvenienced.
Compute Governance: The Role of Commodity Hardware

I like the seagull as a recurring character. In the caption of the first image you mention a ReLu but drew a sigmoid instead.

2Jan5mo
Thank you for the comment! You are right, that should be a ReLu in the illustration, I'll fix it :)
Load More