suffering-focused-altruist/longtermist
my pgp public key, partly to prevent future LLM impersonation; you can always ask me to sign a dated message. [the private key is currently stored in plaintext within an encrypted drive, so is vulnerable to being read by local programs]:
-----BEGIN PGP PUBLIC KEY BLOCK-----
mDMEZiAcUhYJKwYBBAHaRw8BAQdADrjnsrbZiLKjArOg/K2Ev2uCE8pDiROWyTTO
mQv00sa0BXF1aWxhiJMEExYKADsWIQTuEKr6zx3RBsD/QW3DBzXQe0TUaQUCZiAc
UgIbAwULCQgHAgIiAgYVCgkICwIEFgIDAQIeBwIXgAAKCRDDBzXQe0TUabWCAP0Z
/ULuLWf2QaljxEL67w1b6R/uhP4bdGmEffiaaBjPLQD/cH7ufTuwOHKjlZTIxa+0
kVIMJVjMunONp088sbJBaQi4OARmIBxSEgorBgEEAZdVAQUBAQdAq5exGihogy7T
WVzVeKyamC0AK0CAZtH4NYfIocfpu3ADAQgHiHgEGBYKACAWIQTuEKr6zx3RBsD/
QW3DBzXQe0TUaQUCZiAcUgIbDAAKCRDDBzXQe0TUaUmTAQCnDsk9lK9te+EXepva
6oSddOtQ/9r9mASeQd7f93EqqwD/bZKu9ioleyL4c5leSQmwfDGlfVokD8MHmw+u
OSofxw0=
=rBQl
-----END PGP PUBLIC KEY BLOCK-----
I have signed no contracts or agreements whose existence I cannot mention.
The version of this that OpenAI did wasn't unexpected to me, but this one was, I wasn't sure about DeepMind. (Wait, now I'm remembering that "Google AI" is a separate group to "DeepMind"? Web says they were merged in 2023. I wonder which members on the inside this change reflects).
Oh, I've thought a lot about something similar that I call "background processing" - I think it happens during sleep, but also when awake. I think for me it works better when something is salient to my mind / my mind cares about it. According to this theory, if I was being forced to learn music theory but really wanted to think about video games, I'd get less new ideas about music theory from background processing, and maybe it'd be less entered into my long term memory from sleep.
I'm not sure how this effects more 'automatic' ('muscle memory') things (like playing the piano correctly in response to reading sheet music).
I'm unsure if you [practice piano] in the morning, you'll need to remind yourself about the experience before you fall asleep, also implying that you can only really consolidate 1 thing a day
I'm not sure about this either. It could also be formulated as there being some set amount of consolidation you do each night, and you can divide them between topics, but it's theoretically (disregarding other factors like motivation; not practical advice) most efficient if you do one area per day (because of stuff in the same topic having more potential to relate to each other and be efficiently compressed or generalized from or something. Alternatively, studying multiple different areas in a day could lead to creative generalization between them).
Pollywogs (the larval form of frogs, after eggs, and before growing legs) are an example where huge numbers of them are produced, and many die before they ever grow into frogs, but from their perspective, they probably have many many minutes of happy growth, having been born into a time and place where quick growth is easy: watery and full of food
Consider an alien species which requires oxygen, but for whom it was scarce during evolution, and so they were selected to use it very slowly and seek it ruthlessly, and feel happy when they manage to find some. It would be wrong for one of that species to conclude that a species on earth must be happy all the time because there's so much oxygen; because oxygen is abundant, we are neutral to it.
Most of the animal-life-moments I've seen in nature have exhibited a profound "OK-ness". Just chilling. Just being. Just finding some food. Just happy to exist.
Emotional status must be inferred. There are some cases we can easily infer happiness or suffering in nonhumans, by similarity to how we express emotions: a dog jumping around and excited to go outside, a pig squealing and struggling as they are lowered into a gas chamber. There are others we cannot easily know the experience of: a duck sitting in a lake, a caterpillar crawling across a leaf. To call them happy without an analysis from first-principles of what evolutionary pressures might lead to the experience or not of happiness in such moments is to project happiness onto them.
I think most of the rest of your comment is similarly projective.
I wonder if you have the opposite sort of bias to that which you say the OP might have; maybe you have a very happy outlook which causes you to project happiness-to-exist onto life forms you do not understand.
Certainly when "life itself" becomes questionable for animals in nature, they seem to do everything they can to keep their life... escaping danger, seeking safety, eating food (rather than engaging in a hunger strike or trying to forgo water and thirst to death to bring their own end on faster). Every indication we have is that animals are NOT suicidal, but rather they seem to "net value" their own lives. Presumably because these are good lives according to their own subjectivities.
A human with chronic depression will still flee if under attack. Evolution imbues beings with a drive to survive even when in great pain. It does not care whether we suffer, indeed it uses suffering to motivate us; the counter that we would choose to kill ourselves if we suffered enough does not hold, evolution gets around that by then also selecting for us to not kill ourselves.
(I note these things without implying a side on the question of what kinds of mental states are most common in wild animal lives, which looks difficult to speculate about with confidence.)
consider a system which is capable of self-modification and changing it's own goals, now the difference between an instrumental goal and a terminal goal erodes.
If an entity's terminal goal is to maximize paperclips, it would not self-modify into a stamp maximizer, because that would not satisfy the goal (except in contrived cases where doing that is the choice that maximizes paperclips). A terminal goal is a case of criteria according to which actions are chosen; "self-modify to change my terminal goal" is an action.
I sent the following offer (lightly edited) to discuss metaphysics to plex. I extend the same to anyone else.[1]
(Note: I don't do psychedelics or participate in the mentioned communities[2], and I'm deeply suspicious of intuitions/deep-human-default-assumptions. I notice unquestioned intuitions across views on this, and the primary thing I'd like to do in discussing is try to get the other to see theirs.)
Hi, I saw [this shortform] by you
Wondering if you'd want to discuss metaphysics itself / let me socratic-question[3] you. I lean towards the view that "metaphysics is extra-mathematical; it contains non-mathematical-objects, such as phenomenal qualia" but I'd like to mostly see if I can socratic-question you into noticing incoherence. I'm too uncertain about the true metaphysics to put forth a totalizing theory.
It might help if you've read this post first.
My starting questions:
Do you believe phenomenal qualia exist? If you do, are they mathematical objects? If you don't, can you watch this short video on a thought experiment?
(Commentary: someone in the comments of the linked post told me that video was what changed their position from 'camp 1' (~illusionist, e.g. "qualia is an intuition from evolution") to 'camp 2' ("qualia is metaphysically fundamental"))
- You mentioned 'matter' and 'math' as candidates for 'fundamental to reality'. My question depends on which is your view.
If you believe matter is fundamental, what do you mean by matter; how does it differ from math?
(Commentary: "matter" seems like an intuitive concept; one has an intuition that reality is "made of something", and calls that "thing" (undefined) "matter", and refers to "matter" without knowing "what it is" (also undefined); there being a word for this intuition prevents them from noticing their lack of knowledge, or the circular self-reference between "matter" and "what the world is made of")
If you believe math is fundamental, what distinguishes this particular mathematical universe from other ones; what qualifies this world as "real", if anything; what 'breathes fire into the equations and creates a world for them to describe'?
(Commentary: one self-consistent position answers "nothing" - that this world is just one of the infinitely many possible mathematical functions / programs. That 'real' secretly means 'the program(s?) we are part of'. Though I observe this position to be rare; most have a strong intuition that there is something which "makes reality real".)
- If you believe math is fundamental, what is math?[4]
Though I'll likely lose interest if it seems like we're talking past each other / won't resolve any cruxy disagreements.
(except arguably the qualia research institute's discord server, which might count because it has psychedelics users in it)
(Questioning with the goal of causing the questioned one to notice specific assumptions or intuitions to their beliefs, as a result of trying to generate a coherent answer)
From an unposted text:
The paradox of recursive explanation applies to metaphysics too
In Explain/Worship/Ignore (~500 words), Eliezer describes an apparent paradox: if you ask for some physical phenomena to be explained (shown to have a more fundamental cause), then ask the same of the explanation, and so on, the only conceivable outcomes seem paradoxical: infinite recurse, circular self-reference, or a 'special' first cause that does not itself need to be explained.
This is sometimes called the paradox of why. It's usually applied to physics; this text applies it to logic/math-structure and metaphysics/ontology too. In short, you can continually ask "but what is x" for any aspect of logic or ontology.
Here's a hypothetical Socratic dialogue.
Author: "What is reality?"
Interlocutor: "Reality is
<complete description of [very large mathematical structure that perfectly reflects the behavior of the universe]>
."Author: "Let's suppose that is true. But this 'math' you just used; what is it?"
Interlocutor: "Mathematics is a program which assigns, based on a few simple rules, 'T' or 'F' to inputs formatted in a certain way."
Author: "I think you just moved the hard part of the question to be one question away: What is a program?"
Interlocutor: "Hmm. I can't define a program in terms of math, because I just did the reverse. Wikipedia says a program is 'a sequence of instructions for a computer to execute'. If I just wrote that, Author would ask what it means for a computer to execute an instruction. What else could I write?
A computer is a part of physics arranged into a localized structure, and for this structure to 'execute an instruction' is for physical law to flow through (operate on, apply to) it, such that the structure's observed behavior matches that of a simpler-than-physics abstract system for transforming inputs to outputs. Unfortunately, I've already defined physics in terms of math, so defining a program in terms of physics would be circular. I think I'm at a dead end."
Author: "Do you want to revise one of your previous definitions?"
Interlocutor: "Maybe I could define math as some more fundamental thing instead of as a certain kind of program. But I just failed to find such a more fundamental thing for 'program'. Let's check Wikipedia again...
WP:Mathematics: 'Mathematics involves the description and manipulation of abstract objects that consist of either abstractions from nature or—in modern mathematics—purely abstract entities that are stipulated to have certain properties, called axioms'
WP:Mathematical_Object 'Typically, a mathematical object can be a value that can be assigned to a symbol, and therefore can be involved in formulas. Commonly encountered mathematical objects include numbers, expressions, shapes, functions, and sets.'
"I can't use the 'abstractions from nature' part, because I've already defined nature to be mathematical, so that would be circular. Saying math is made of numbers, expressions, symbols, etc, isn't helpful either, though; Author will just ask what those are, metaphysically. Okay, I concede for now, though maybe a commenter will propose such a more-fundamental-thing that I'm not aware of, though it would invite the same question."
[End of dialogue]
(It's not actually the case that I don't value their well-being; I don't want them to suffer and if they were tortured, imagining that would make me sad; I'd prefer beings who don't care about some subset of other beings / are value-orthogonal to just be prevented from hurting them. I just think Richard, in the current world, probably causes more tragedy, based on the comment, so yes I think the current world would be better if it did not have any such people.)
Agreed about the double standard part, that's something I was hoping to highlight.
A community that can accept "nazis" (in the vegan sense) cannot also accept "resistance fighters" (in the vegan sense). Either the "nazi" deserves to exist or he doesn't. But to test this dichotomy, somebody has to out themselves as a "nazi."
This doesn't seem true. It seems like it's saying that the directly opposing views on this cannot both exist in a "community" (to the extent LW is a community), but they evidently do both exist here (which is to be expected with enough users).
(quoting the comment by Richard_Kennaway that started this thread since it's 2 years old, and plausibly some will see my comment from the 'new' section and be confused by what I write next otherwise)
I eat meat and wear leather and wool. I do think that animals, the larger ones at least, can suffer. But I don’t much care. I don’t care about animal farming, nor the (non-human) animal suffering resulting from carnivores and parasites. I’d rather people not torture their pets, and I’d rather preserve the beauty and variety of nature, but that is the limit of my caring. If I found myself on the surface of a planet on which the evolution of life was just beginning, I would let it go ahead even though it mean all the suffering that the last billion years of this planet have seen.
Bring on the death threats.
[Even if one thinks, in a utilitarian sense, the world would be better without such a person in it], killing them would still be a waste of one's opportunity to effect the world, given there are much more effective ways to improve the future (e.g., donating $1k to an animal charity does more good IIUC; more ambitiously, helping solve alignment saves all the animals in one go, if we don't die to unaligned ASI first).
(I feel like this comment would be incomplete without also mentioning that I guess most but not all people stating they're indifferent to and cause non-human suffering now would reproach the view and behavior eventually, and that relative to future beings who have augmented their thinking ability and lived for thousands of years, all current beings are like children, some hurting others very badly in confusion.)
Which I took to mean that some they overlap in some instrumental goals. That is what you meant right?
No. I was trying to explain that: any agent that can be predicted by thinking of them as having two separate values for two different things, can also be predicted by thinking of them as maximizing some single value which internally references both things.
For example: "I value paperclips. I also value stamps, but one stamp is only half as valuable as a paperclip to me" → "I have the single value of maximizing this function over the world: {paperclip-amount×2 + stamp-amount}". (It's fine to think of it in either way)
can you be explicit (to be be honest, use layman's terms)
If you want, it would help me learn to write better, for you to list off all the words (or sentences) that confused you.
are you saying that superintelligent AGIs won't necessary converge in values because even a single superintelligent agent may have multiple terminal goals?
No, I was responding to your claim that I consider unrelated. Like I wrote at the top: "That [meaning your claim that humans have multiple terminal goals] is not relevant to whether there are convergent terminal values"[1]
some terminal goals may overlap in they share certain instrumental goals?
I don't know what this is asking / what 'overlap' means. That most terminal goals share instrumental subgoals is called instrumental convergence.
Which in other words means, even if it were true, "humans have multiple terminal goals" would not be a step of the argument for it
If there's no constraints on when they have to name a system "GPT-5", they can make the statement true by only naming a system "GPT-5" if it is smart enough. (cf. Not Technically a Lie)
Edit: though "and those two [...] I'd like to hear back from you in a little bit of time" implies a system named GPT-5 will be released 'in a little bit of time'