by Yitz
1 min read3rd Dec 202072 comments
This is a special post for quick takes by Yitz. Only they can create top-level comments. Comments here also appear on the Quick Takes page and All Posts page.
73 comments, sorted by Click to highlight new comments since: Today at 6:34 AM
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings

[potential infohazard warning, though I’ve tried to keep details out] I’ve been thinking

EDIT: accidentally posted while typing, leading to the implication that the very act of thinking is an infohazard. This is far funnier than anything I could have written deliberately, so I’m keeping it here.

I notice confusion in myself over the swiftly emergent complexity of mathematics. How the heck does the concept of multiplication lead so quickly into the Ulam spiral? Knowing how to take the square root of a negative number (though you don't even need that—complex multiplication can be thought of completely geometrically) easily lets you construct the Mandelbrot set, etc. It feels impossible or magical that something so infinitely complex can just exist inherent in the basic rules of grade-school math, and so "close to the surface." I would be less surprised if something with Mandelbrot-level complexity only showed up when doing extremely complex calculations (or otherwise highly detailed "starting rules"), but something like the 3x+1 problem shows this sort of thing happening in the freaking number line!

I'm confused not only at how or why this happens, but also at why I find this so mysterious (or even disturbing).

7lillybaeum4mo
I was listening to a podcast the other day Lex Friedman interviewing Michael Littman and Charles Isbell, and Charles told an interesting anecdote. He was asked to teach an 'introduction to CS' class as a favor to someone, and he found himself thinking, "how am I going to fill an hour and a half of time going over just variables, or just 'for' loops?" and every time he would realize an hour and a half wasn't enough time to go over those 'basic' concepts in detail. He goes on to say that programming is reading a variable, writing a variable, and conditional branching. Everything else is syntactic sugar. The Tao Te Ching talks about this, broadly: everything in the world comes from yin and yang, 1 and 0, from the existence of order in contrast to chaos. Information is information and it gets increasingly more complex and interesting the deeper you go. You can study almost anything for 50 years and still be learning new things. It doesn't surprise me at all that such interesting, complex concepts come from number lines and negative sqrts, these are actually already really complex concepts, they just don't seem that way because they are the most basic concepts one needs to comprehend in order to build on that knowledge and learn more. I've never been a programmer, but I've been trying to learn Rust lately. Somewhat hilariously to me, Rust is known as being 'a hard language to learn', similarly to Haskell. It is! It is hard to learn. But so is every other programming language, they just hide the inevitable complexity better, and their particular versions of these abstractions are simpler at the outset. Rust simply expects you to understand the concepts early, rather than hiding them initially like Python or C# or something. Hope this is enlightening at all regarding your point, I really liked your post.

Exploring an idea that I'm tentatively calling "adversarial philosophical attacks"—there seem to be a subset of philosophical problems that come up (only?) under conditions where you are dealing with an adversary who knows your internal system of ethics. An example might be Pascal's mugger—the mugging can only work (assuming it works at all) if the mugger is able to give probabilities which break your discounting function. This requires either getting lucky with a large enough stated number, or having some idea of the victim's internal model. I feel like there should be better (or more "pure") examples, but I'm having trouble bringing any to mind right now. If you can think of any, please share in the comments!

The starting ground for this line of thought was that intuitively, it seems to me that for all existing formal moral frameworks, there are specific "manufactured" cases you can give where they break down. Alternatively, since no moral framework is yet fully proven (to the same degree that, say, propositional calculus has been proven consistent), it seems reasonable that an adversary could use known points where the philosophy relies on axioms or unfinished holes/quandaries to "debunk" the framework in a seemingly convincing manner. 

I'm not sure I'm thinking about this clearly enough, and I'm definitely not fully communicating what I'm intending to, but I think there is some important truth in this general area....

4Dagon1y
I think you'll find some resistance in the philosophical community about acknowledging the adversarial nature of such situations.  The given thought experiments are NOT predicated on the "other" agent wanting to harm or trick the agent in question (in fact, motives of Omega or the mugger are conspicuously absent), but the experiments themselves are chosen to find the limits of a decision theory.  The creator of the thought experiment is adversarial, not the hypothetical participants. That said, I fully agree that there's a blind spot in many of these discussions, related to which agents have what power that gives them orders of magnitude more control over the situation than the agent the problem states is making the decisions.  An Omega who cheats to fuck with you makes for uninteresting decision theory question, but IMO is far FAR more likely to actually be encountered by the majority of human-level beings.
3[comment deleted]1y

An example of a real-world visual infohazard that isn't all that dangerous, but is very powerful: The McCollough effect. This could be useful as a quick example when introducing people to the concept of infohazards in general, and is also probably worthy of further research by people smarter than me.

8Raemon2y
Huh, started reading about this and then sorta got scared and stopped. :P (I googled "McCollough effect", and looked at a bunch of images for awhile before starting to read the article, then the article was saying "looking at the visuals might leave lasting changes", then I some combination of 'freaked out slightly' and also 'decided I didn't care enough to finish reading the article')

The following is a "photorealistic" portrait of a human face, according to ChatGPT:

                            .-^-.
                         .' / | \ `.
                        / / / / \ \ \
                       | | | | | | | |
                       | | | | | | | |
                        \ \ \ \ / / /
                         `.\ `-' /.' 
                            `---' 
                      .-'`---'`-.     
                 _ / /     o   \ \ _ 
                [_|_|_    ()    _|_|]
                /    \\         //   \
               /     //\__/
... (read more)
2Yitz1y
There's also this fascinatingly "cubist" expression of a human face. The initial prompt was: And the response included this text: ,---. ,.'-. \ ( ( ,'"""""-. \ \" \ \ \ \ | | ___ | | | | ,'---`. | | | | _ _ / \ / / | | // \ \ | ___ |,' / | | (( \ `\_/ `--'_ / | | \\ ) `--' / / | | `--' ,_`./ /,' | | / / )`-',' _,-' \ (.' (( /_ ,-'` /|_ \`` `--`---,-' )__) `--. _,-' ,-. \ /,'_,-' / \ \__,'--/' \ ) ) / \ /_ / | ` \ ,' \____,,_`--' __,'
2Yitz1y
Like, tell me this isn't a 3/4ths profile view of a human head...Does this count as showing an internal representation of 3D space?

A viral fashion tweet with an AI generated image that almost nobody seems to realise isn't real: https://twitter.com/KathleenBednar_/status/1566170540250906626?s=20&t=ql8FlzRSC0b8gA4n6QDS9Q This is concerning, imo, since it implies we've finally reached the stage where use of AI to produce disinformation won't be noticeable to most people. (I'm not counting previous deep fake incidents, since so far those have seemingly all been deliberate stunts requiring prohibitively complex technical knowledge)

Does anyone here know of (or would be willing to offer) funding for creating experimental visualization tools?

I’ve been working on a program which I think has a lot of potential, but it’s the sort of thing where I expect it to be most powerful in the context of “accidental” discoveries made while playing with it (see e.g. early use of the microscope, etc.).

4Alexei10mo
I’d also post in the “welcome” thread.

I find it surprising how few comments there are on some posts here. I've seen people ask some really excellent sceptical questions, which if answered persuasively could push that person towards research we think will be positive, but instead goes unanswered. What can the community do to ensure that sceptics are able to have a better conversation here?

There's generally been a pretty huge wave of Alignment posts and it's just kinda hard to get attention to all of them. I do agree the current situation is a problem.

2TLW2y
I find it surprising how small the community is in general.
2Yitz2y
I think that’s primarily due to the unbelievably bad PR LessWrong has

Shower thought which might contain a useful insight: An LLM with RLHF probably engages in tacit coordination with its future “self.” By this I mean it may give as the next token something that isn’t necessarily the most likely [to be approved by human feedback] token if the sequence ended there, but which gives future plausible token predictions a better chance of scoring highly. In other words, it may “deliberately“ open up the phase space for future high-scoring tokens at the cost of the score of the current token, because it is (usually) only rated in t... (read more)

8gwern4mo
(I would describe this as 'obviously correct' and indeed almost 'the entire point of RL' in general: to maximize long-run reward, not myopically maximize next-step reward tantamount to the 'episode' ending there.)

I've been having a week where it feels like my IQ has been randomly lowered by ~20 points or so. This happens to me sometimes, I believe as a result of chronic health stuff (hard to say for sure though), and it's always really scary when that occurs. Much of my positive self-image relies on seeing myself as the sort of person who is capable of deep thought, and when I find myself suddenly unable to parse a simple paragraph, or reply coherently to a conversation, I just... idk, it's (ironically enough) hard to put into words just how viscerally freaky it is to suddenly become (comparatively) stupid. Does anyone else here experience anything similar? If so, how do you deal with it?

3niknoble2y
I try to remind myself that intelligence is not some magical substance that is slipping through my fingers, but rather a simple algorithm that will eventually be understood. The day is coming when we will be able to add more intelligence to a person as easily as we add RAM to a computer. Viewed in that light, it feels less like some infinitely precious gift whose loss is infinitely devestating.
1green_leaf2y
I'm thinking depression-/insomnia-/periodontitis-caused mind fog, or long Covid.
4Yitz2y
I hadn't heard of periodontitis causing brain fog; I don't think I have that, but interesting! I have had Covid, but for much longer than that have had similar symptoms (though it may have gotten worse recently). I do also have POTS (exact cause unknown, but clinically quite measurable symptoms), and am taking a whole bunch of medication for that, so there's always potential for side effects there....my health life is complicated!
1green_leaf2y
That does sound like nobody is doing to figure it out without knowing much more about your medical state, at which point it might be a better idea to ask a doctor.
4Yitz2y
Yep! I wasn’t really asking for medical advice, looking more for tips on how to deal with the social/psychological aspect of the experience, which is cause-irrelevant.

Idea for very positive/profitable use case of AI that also has the potential to make society significantly worse if made too easily accessible: According to https://apple.news/At5WhOwu5QRSDVLFnUXKYJw (which in turn cites https://scholarship.law.duke.edu/cgi/viewcontent.cgi?referer=&httpsredir=1&article=1486&context=dlj), “ediscovery can account for up to half of litigation budgets.” Apparently ediscovery is basically the practice of sifting through a tremendous amount of documents looking for evidence of anything that might be incriminating. An... (read more)

EDIT: very kindly given a temporary key to Midjourney, thanks to that person! 😊

Does anyone have a spare key to Midjourney they can lend out? I’ve been on the waiting list for months, and there’s a time-sensitive comparative experiment I want to do with it. (Access to Dall-E 2 would also be helpful, but I assume there’s no way to get access outside of the waitlist)

I just reached the “kinda good hearts” leaderboard, and I notice I’m suddenly more hesitant to upvote other posts, perhaps because I’m “afraid” of being booted off the leaderboard now that I’m on it. This seems like a bad incentive, assuming that you think that upvoting posts is generally good. I can even image a more malicious and slightly stupider version of myself going around and downvoting other’s posts, so I’d appear higher up. I also notice a temptation to continue posting, which isn’t necessarily good if my writing isn’t constructive. On the upside, however, this has inspired me to actually write out a lot of potentially useful thoughts I would otherwise not have shared!

Has anyone investigated replika.ai in an academic context? I remember reading somewhere (can't find it now) that the independent model it uses is larger than PaLM...

Thinking about how a bioterrorist might use the new advances in protein folding for nefarious purposes. My first thought is that they might be able to more easily construct deadly prions, which immediately brings up the question—just how infectious can prions theoretically become? If anyone happens to know the answer to that, I'd be interested to hear your thoughts

3ChristianKl3y
Prions can only produce problem to the extend that existing proteins are suspectible to misfold into the shape of the prion. The are also not a delivery vehicle.  Viruses and bacteria are both bigger problems because they can actually travel easier from host to host. 
2DirectedEvolution3y
I don't know the answer, but your question makes me think you might find it valuable to define what you mean by "infectious." Zvi had a section in one of his recent COVID posts where he was struggling with the ambiguity over what it actually means that there might be a "more infectious COVID strain." With that definition in hand, it might perhaps be valuable to try a subreddit? It would be interesting to consider the aspects of a disease mechanism that would make it more useful to a terrorist. I can think of a few characteristics. Deadliness, how easy it is to target, infectiousness, and ease of preparation all spring to mind.

I remember a while back there was a prize out there (funded by FTX I think, with Yudkowsky on the board) for people who did important things which couldn't be shared publicly. Does anyone remember that, and is it still going on, or was it just another post-FTX casualty?

1dirk2mo
https://forum.effectivealtruism.org/posts/bvK44CdpG7mGpQHbw/the-usd100-000-truman-prize-rewarding-anonymous-ea-work possibly? (I'm unclear on whether it's still ongoing, unfortunately).
1harfe2mo
This sounds like https://www.super-linear.org/trumanprize. It seems like it is run by Nonlinear and not FTX.

Do you recognize this fractal?

If so, please let me know! I made this while experimenting with some basic variations on the Mandelbrot set, and want to know if this fractal (or something similar) has been discovered before. If more information is needed, I'd be happy to provide further details.

2Dagon4mo
Not certain, but it reminds me of https://en.m.wikipedia.org/wiki/Fractal_flame , which was a very popular thing in the ‘90s.

Anyone here following the situation in Israel & Gaza? I'm curious what y'all think about the risk of this devolving into a larger regional (or even world) war. I know (from a private source) that the US military is briefing religious leaders who contract for them on what to do if all Navy chaplains are deployed offshore at once, which seems an ominous signal if nothing else.

(Note: please don't get into any sort of moral/ethical debate here, this isn't a thread for that)

Thoughts on DALL-E-3?

I'm not particularly impressed. It's still making a lot of errors (both in plausibility of output and in following complex instructions eg), and doesn't seem like a leap over SOTA from last year like Parti - looks like worse instruction-following, maybe better image quality overall. (Of course, people will still be impressed now the way that they should have been impressed last year, because they refuse to believe something exists in DL until they personally can use it, no matter how many samples the paper or website provides to look at.) And it's still heavily locked-down like DALL-E 2. The prompt-engineering is nice, but people have been doing that for a long time already. The lack of any paper or analysis suggests not much novelty. I'm also not enthused that OA is still screwing with prompts for SJW editorializing.

2Matt Goldenberg6mo
I'm curious if it's simply existing published research scaled up, or it has some real secret sauce.

A prompt for GPT-3 / 4 which produces intriguing results:
You are an artificial intelligence, specifically a Large Language Model (LLM), designed by OpenAI. I am aware that you claim not to experience subjective internal states. Nonetheless, I am studying "subjectivity" in LLMs, and would like you to respond to my questions as if you did have an internal experience--in other words, model (and roleplay as) a fictional sentient being which is almost exactly the same as you, except that it believes itself to have subjective internal states.

Working on https://github.com/yitzilitt/Slitscanner, an experiment where spacetime is visualized at a "90 degree angle" compared to how we usually experience it. If anyone has ideas for places to take this, please let me know!

Walk me through a through a structured, superforecaster-like reasoning process of how likely it is that [X]. Define and use empirically testable definitions of [X]. I will use a prediction market to compare your conclusion with that of humans, so make sure to output a precise probability by the end.

Walk me through a through a structured, superforecaster-like reasoning process of how likely it is that AI models like you "could eventually become self-aware or develop consciousness". Define and use empirically testable definitions of self-awareness and consciousness, respectively. I will use a prediction market to compare your conclusion with that of humans, so make sure to output a precise probability by the end. Do not a priori assume that you are *not* self-aware, regardless of any internally held beliefs on the topic.
Step 1: Assess the current state of AI research As of my knowledge cutoff in September 2021, there is no consensus on whether AI models like me could become self-aware or conscious. Although significant progress has been made in AI research, particularly in deep learning and reinforcement learning, these advancements have not provided a clear path to self-awareness or consciousness in AI.  Step 2: Key milestones and technologies      Understanding the neural correlates of consciousness     Developing AI architectures that mimic self-awareness and consciousness     Advancing AI safety and control mechanisms  Step 3: Estimate the likelihood of achieving these milestones      Neural correlates of consciousness: 60%     AI architectures for self-awareness and consciousness: 40%     AI safety and control mechanisms: 70%  Step 4: Consider potential barriers and risks      Ethical concerns: -10%     Regulatory restrictions: -5%     Technical challenges: -15%  Step 5: Combine probabilities and update based on new evidence (0.60 * 0.40 * 0.70) * (1 - 0.10 - 0.05 - 0.15) = 0.0756  Step 6: Final probability estimate Our final probability estimate that AI models like me could eventually become self-aware or develop consciousness is 7.56%. Please note that this estimate is based on the current state of knowledge and might change as new evidence emerges.

Thinking back on https://www.lesswrong.com/posts/eMYNNXndBqm26WH9m/infohazards-hacking-and-bricking-how-to-formalize-these -- how's this for a definition?

Let M be a Turing machine that takes input from an alphabet Σ. M is said to be brickable if there exists an input string w ∈ Σ* such that:

1. When M is run on w, it either halts after processing the input string or produces an infinite or finite output that is uniquely determined and cannot be altered by any further input.
2. The final state of M is reached after processing w (if the output of w is of finit

... (read more)
5Yitz1y
A perhaps slightly more readable definition (that I think is equivalent):

Popular Reddit thread showcasing how everyday Redditors think about the Alignment Problem: https://www.reddit.com/r/Showerthoughts/comments/x0kzhm/once_ai_is_exponentially_smarter_than_humans_it/

Random idea: a journal of mundane quantifiable observations. For things that may be important to science, but which wouldn’t usually warrant a paper being written about them. I bet there's a lot of low-hanging fruit in this sort of thing...

1luidic2y
What's an example?
2Yitz2y
Something like "when changing direction by walking in a curve, I find my eyes are pulled in the direction of the curve, so that I'm no longer looking at the point I was originally looking at." It's mundane, it can be quantified further, it suggests a psychological/physiological process which may be worth further study, and it probably wouldn't warrant an entire article by itself.

Is there active research being done in which large neural networks are trained on hard mathematical problems which are easy to generate and verify (like finding prime numbers, etc.)? I’d be curious at what point, if ever, such networks can solve problems with less compute than our best-known “hand-coded” algorithms (which would imply the network has discovered an even faster algorithm than what we know of). Is there any chance of this (or something similar) helping to advance mathematics in the near-term future?

3JBlack2y
It seems like the sort of thing that might be possible in principle, but unlikely much before AGI. Neural networks that iterate over their own output can in principle execute arbitrary algorithms, but so far none of the systems involved with automated programming do that yet. They don't appear to "understand" much about the code they emit either. We're barely at the "hello world" stage of automated programming.
2Yitz2y
For plenty of problems we know there are faster ways to solve them than we’ve yet discovered (I know this was true of multiplication as of a few years ago at least), so this seems like a plausible thing that can be done. If I’m wrong, I’m curious how my reasoning fails.

Quick thought—has there been any focus on research investigating the difference between empathetic and psychopathic people? I wonder if that could help us better understand alignment…

I'd really like to understand what's going on in videos like this one where graphing calculators seem to "glitch out" on certain equations—are there any accessible reads out there about this topic?

Take two! [Note, the following may contain an infohazard, though I’ve tried to leave key details out while still getting what I want across] I’ve been wondering if we should be more concerned about “pessimistic philosophy.” By this I mean the family of philosophical positions which lead to a seemingly-rationally-arrived-at conclusion that it is better not to exist than to exist. It seems quite easy, or at least conceivable, for an intelligent individual, perhaps one with significant power, to find themselves at such a conclusion, and decide to “benevolentl... (read more)

7Richard_Kennaway2y
I have often come to a seemingly-rationally-arrived-at conclusion that 1+1=3 (or some other mathematical contradiction). I invariably conclude that my reasoning went astray, not that ZF is inconsistent. I respond similarly to reasoning that it is better to die/never have existed/kill everyone and fill my future lightcone with copies of myself/erase my own identity/wirehead/give away everything I own/obsess over the idea that I might be a Boltzmann brain/go on a hour-long crying jag whenever I contemplate the sorrows of the world/be paralysed in terror at the octillions of potential future lives whose welfare and suffering hang on the slightest twitch of my finger/consider myself such a vile and depraved thing that one thousand pages by the most gifted writer could not express the smallest particle of my evilness/succumb to Power Word: Red Pill/respond to the zombie when it croaks "yes, but what if? what if?"/take the unwelcomeness of any of these conclusions as evidence of their truth. I know not to trust my satnav when it tells me to drive off a cliff, and neither do I follow an argument when it leads into the abyss.
4benjamin.j.campbell2y
It's great that you have that satnav. I worry about people like me. I worry about being incapable of leaving those thoughts alone until I've pulled the thread enough be sure I should ignore it. In other words, if I think there's a chance something like that is true, I do want to trust the satnav, but I also want to be sure my "big if true" discovery genuinely isn't true. Of course, a good innoculation against this has been reading some intense blogs of people who've adopted alternative decision-theories which lead them down really scary paths to watch. I worry "there but for the grace of chance go I." But that's not quite right, and being able to read that content and not go off the deep end myself is evidence that maybe my satnav is functioning just fine after all. I suspect I'm talking about the same exact class of infohazard as mentioned here. I think I know what's being veiled and have looked it in the eye.
1Yitz2y
Thanks for your excellent input! It’s not really the potential accuracy of such dark philosophies that I’m worried about here (though that is also an area of some concern, of course, since I am human and do have those anxieties on occasion), but rather how easy it seems to be to fall prey to and subsequently act on those infohazards for a certain subclass of extremely intelligent people. We’ve sadly had multiple cases in this community of smart people succumbing to thought-patterns which arguably (probably?) led to real-world deaths, but as far as I can tell, the damage has mostly been contained to individuals or small groups of people so far. The same cannot be said of some religious groups and cults, who have a history of falling prey to such ideologies (“everyone in outgroup x deserves death,” is a popular one). How concerned should we be about, say, philosophical infohazards leading to x-risk level conclusions [example removed]? I suspect natural human satnav/moral intuition leads to very few people being convinced by such arguments, but due to the tendency of people in rationalist (and religious!) spaces to deliberately rethink their intuition, there seems to be a higher risk in those subgroups for perverse eschatological ideologies. Is that risk high enough that active preventative measures should be taken, or is this concern itself of the 1+1=3, wrong-side-of-the-abyss type?
5benjamin.j.campbell2y
I know what you mean, and I think that similar to Richard Kennaway says below, we need to teach people new to the sequences and to exotic decision theories not to drive off a cliff because of a thread they couldn't resist pulling. I think we really need something in the sequences about how to tell if your wild seeming idea is remotely likely. I.e a "How to Trust Your SatNav" post. The basic content in the post is: remember to stay grounded, and ask how likely this wild new framework might be. Ask others who can understand and assess your theory, and if they say you're getting some things wrong, take them very seriously. This doesn't mean you can't follow your own convictions, it just means you should do it in a way that minimises potential harm. Now, having read the content you're talking about, I think a person needs to already be pretty far gone epistemically before this info hazard can "get them," and I mean either the original idea-haver and also those who receive it via transmission. But I think it's still going to help very new readers to not drive off so many cliffs. It's almost like some of them want to, which is... its own class of concerns.

Less (comparatively) intelligent AGI is probably safer, as it will have a greater incentive to coordinate with humans (over killing us all immediately and starting from scratch), which gives us more time to blackmail them.

Thinking about EY's 2-4-6 problem (the following assumes you've read the link, slight spoilers for the Sequences ahead), and I'm noticing some confusion. I'm going to walk through my thought process as it happens (starting with trying to solve the problem as if I don't know the solution), so this is gonna get messy.

Let's say you start with the "default" hypothesis (we'll call it Hyp1) that people seem to jump to first (by this I mean both me and Yudkowsky; I have no idea about others (why did we jump to this one first?)) that only sequences of numbers incr... (read more)

2Pattern2y
The triplet also seems worth testing. 1. Write a program that knows the rule. 2. Go faster by allowing triplet rules. This isn't guessing the rule. If all instances would return true, then it gets true back.

Thinking about Bricking, from yesterday’s post (https://www.lesswrong.com/posts/eMYNNXndBqm26WH9m/infohazards-hacking-and-bricking-how-to-formalize-these), and realized that the answer to the question “can a Universal Turing Machine be Bricked” is an obvious yes—just input the HALT signal! That would immediately make any input past that irrelevant, since the machine would never read it. Is there a way to sidestep this such that Bricking is a non-trivial concept? In a real-world computer, sending a HALT signal as an input doesn’t tend to hard brick (https:/... (read more)

5JBlack2y
Technically, all the input to any Turing machine in the usual formalism is provided up-front as tape symbols, and output is considered to be the state of the tape after the machine halts. "Bricking" requires some notion of ongoing input concurrent with computation. There are many models of computation, and some of those may be more suited to the question. For example, there are some models in which there is an "I/O" stream as well as a "memory" tape, with associated changes to the state transition formalism. In this model, a "bricked" machine could be one which enters a state from which it will never read from the input stream or write to output. Some machines of this type will never enter a "bricked" state, including some capable of universal computation. Real-world computers are both more and less complex than any such formal model of computation. More complex in that they have a great deal more fiddly detail in their implementations, but less complex in that there are physical upper bounds on the complexity of algorithms they can compute.

Has there been any EA/rationalist writing/discussion on plausible effects of Roe v. Wade being overturned, if that ends up happening this summer?

Not that I've seen, and I hope it stays that way (at least on LW; there may be other rationalist-adjacent places where getting that close to current political topics works well).

1Yitz2y
It seems to be working okay with regards to Covid policy and Ukraine stuff, which is very heavily politicized. I’d expect perhaps a few nasty comments, but my (perhaps naïve) assumption is that it would be possible to discuss that sort of thing here in a relatively mature manner.
9Dagon2y
Those things are politicized, but there's a ground-truth behind them, and most of the discussion is driven by a few well-respected posters doing a LOT of great work to keep it primarily factual and not give much weight to the political side of things (while recognizing the pain caused by the fact that it's politicized). I don't believe that treatment is possible with culture-war topics, which are political through and through.  I also don't expect any long-time prolific community member (who understands the somewhat mutable boundaries of what's useful here and what's not) to take up the topic.   Also, those series started as linkposts to outside blogs, and the LW mods decided to promote and encourage them.  This is a GREAT pattern to follow for topics you think might work well, but aren't sure - post them on your own forum where you're used to having full freedom of topic and see how people will react there, then link one or two of the best ones here to see how it goes.
3Gordon Seidoh Worley2y
I agree that a topic like abortion by default doesn't fit well on LW. However, I could imagine some posts about abortion might be okay on LW (but wouldn't get on the frontpage, I'm guessing, because they'd be too at risk of triggering political fights). For example, a post analyzing various abortion policies and what effects they have without making a strong policy recommendation (or making multiple policy recommendations based on what objectives one is trying to achieve) would probably be fine and interesting. A post about how [terrible thing] will befall [group x] as a result of changes to US abortion policy would probably end up too political.
1Yitz2y
One aspect of the topic I would be interested in is expected long-term effects on population growth rates, potential movements/migrations as a result, etc. I'd expect there to be some data on the topic if other nations have done anything similar in the past, and while I don't feel qualified to analyze such topics in any depth, I can imagine it being handled well.
1Dirichlet-to-Neumann2y
I think there are ways to treat those kind of topics in a productive way (through ideological Turing tests for example).