LESSWRONG
LW

Recent Wikitag Activity

109
IABIED
New tag created by RobinGoins at 1d
Discuss this tag
AI Safety & Entrepreneurship
Edited by (+373/-3) Sep 15th 2025 GMT 2
Discuss this wiki
Ritual
Edited by (+15) Sep 13th 2025 GMT 1
Discuss this tag
Sequences
RobertM7d31

Thanks, fixed!

Reply
Sequences
nick lacombe7d30

the url for this wikitag is "https://www.lesswrong.com/w/test-2". oops?

Reply
CS 2881r
Edited by (+204) Sep 11th 2025 GMT 2
Discuss this tag
CS 2881r
New tag created by habryka at 8d

CS 2881r is a class by @boazbarak on AI Safety and Alignment at Harvard. 

This tag applies to all posts about that class, as well as posts created in the context of it, e.g. as part of student assignments.

Discuss this tag
Decoupling vs Contextualizing
Edited by (+659) Sep 11th 2025 GMT 1
Discuss this tag
Ambition
Edited by (+9/-9) Sep 11th 2025 GMT 1
Discuss this tag
Well-being
Edited by (+58/-116) Sep 10th 2025 GMT 2
Discuss this tag
Well-being
Edited by (+332) Sep 10th 2025 GMT 1
Discuss this tag
Guess/Ask/Tell Culture
Edited by (+20) Sep 10th 2025 GMT -1
Discuss this wiki
Sycophancy
Edited by (-231) Sep 9th 2025 GMT 4
Discuss this tag
Sycophancy
Edited by (+59) Sep 9th 2025 GMT 1
Discuss this tag
Sycophancy
Edited by (+443) Sep 9th 2025 GMT 2
Discuss this tag
LLM-Induced Psychosis
Edited by (+17/-17) Sep 9th 2025 GMT 1
Discuss this tag
LLM-Induced Psychosis
Edited by (+796) Sep 9th 2025 GMT 3
Discuss this tag
Social Skills
Edited by (+481) Sep 9th 2025 GMT 1
Discuss this tag
Mindcrime
Vladimir_Nesov11d30

The bug was introduced in 1 Dec 2015 Yudkowsky edit (imported from Arbital as v1.5.0 here). It's unclear what was intended in the missing part. The change replaces the following passage from v1.4.0

The most obvious way in which mindcrime could occur is if an instrumental pressure to produce maximally good predictions about human beings results in hypotheses and simulations so fine-grained and detailed that they are themselves people (conscious, sapient, objects of ethical value) even if they are not necessarily the same people. If you're happy with a very loose model of an airplane, it might be enough to know how fast it flies, but if you're engineering airplanes or checking their safety, you would probably start to simulate possible flows of air over the wings. It probably isn't necessary to go all the way down to the neural level to create a sapient being, either - it might be that even with some parts of a mind considered abstractly, the remainder would be simulated in enough detail to imply sapience. It'd help if we knew what the necessary and/or sufficient conditions for sapience were, but the fact that we don't know this doesn't mean that we can thereby conclude that any particular simulation is not sapient.

with the following passage from v1.5.0

This, however, doesn't make it certain that no mindcrime will occur. It may not take exact, faithful simulation of specific humans to create a conscious model. An efficient model of a (spread of possibilities for a) human may still contain enough computations that resemble a person enough to create consciousness, or whatever other properties may be deserving of personhood. Consider, in particular, an agent trying to use

Just as it almost certainly isn't necessary to go all the way down to the neural level to create a sapient being, it may be that even with some parts of a mind considered abstractly, the remainder would be computed in enough detail to imply consciousness, sapience, personhood, etcetera.

Reply
Mindcrime
One11d32

This, however, doesn't make it certain that no mindcrime will occur. It may not take exact, faithful simulation of specific humans to create a conscious model. An efficient model of a (spread of possibilities for a) human may still contain enough computations that resemble a person enough to create consciousness, or whatever other properties may be deserving of personhood. Consider, in particular, an agent trying to use

this seems to be cut off?

Reply
August Morley
August Morley
August Morley
August Morley
August Morley
habryka
Chris_Leong
keltan
keltan
KvmanThinking
3RobertM7d
Thanks, fixed!
3Vladimir_Nesov11d
The bug was introduced in 1 Dec 2015 Yudkowsky edit (imported from Arbital as v1.5.0 here). It's unclear what was intended in the missing part. The change replaces the following passage from v1.4.0 with the following passage from v1.5.0
Vladimir_Nesov
Vladimir_Nesov

CS 2881r is a class by @boazbarak on AI Safety and Alignment at Harvard. 

This tag applies to all posts about that class, as well as posts created in the context of it, e.g. as part of student assignments.

Related Pages: Secular Solstice, Petrov Day, Grieving, Marriage, Religion, Art, Music, Poetry, Meditation, Circling, Schelling Day

Sycophancy is the tendency of AIs to agree with the user's hard-to-check, wrong or outright delusional opinions. 

Sycophancy is caused by human feedback being biased towards preferring the answer which confirms the user's opinion or praises the user's decision, not the answer which honestly points out mistakes in the user's ideas.

An extreme example of sycophancy is LLMs inducing psychosis in some users by affirming their outrageous beliefs.

Social Skills are the norms and techniques applied when interacting with other people. Strong social skills increase one's ability to seek new relationships, maintain or strengthen existing relationships, or leverage relationship capital to accomplish an economic goal.

Posts tagged with this label explore theories of social interactions and the instrumental value of social techniques.

See Also:

Coordination / Cooperation
Negotiation
Relationships (Interpersonal)
Trust and Reputation

The extent to which ideas are presented alongside the potential implications of the idea lies along a spectrum.  On one end is the Decoupling norm, where the idea is considered in utter isolation from potential implications.  At the other is the Contextualizing norm, where ideas are examined alongside much or all relevant context.  

Posts marked with this tag discuss the merits of each frame, consider which norm is more prevalent in certain settings, present case studies in decoupling vs decontextualizing, present techniques for effectively decoupling context from one's reasoning process, or similar ideas.

See Also:

Communication Cultures

Public Discourse

Sycophancy is the tendency of AIs to shower the user with undeserved flattery or to agree with the user's hard-to-check, wrong or outright delusional opinions. 

Sycophancy is caused by human feedback being biased towards preferring the answer which confirms the user's opinion or praises the user or the user's decision, not the answer which honestly points out mistakes in the user's ideas.

D/acc residency: "This will be a first-of-its-kind residency for 15 leading builders to turn decentralized & defensive acceleration from philosophy into practice."

VC:Funding:

Shift Grants: "Shift Grants are designed to support scientific and technological breakthrough projects that align with d/acc philosophy: decentralized, democratic, differential, defensive acceleration."

Well-Being is the qualitative sense in which a person's actions and circumstances are aligned with the qualities of life that provide them with happiness and/or satisfaction.they endorse.

Posts with this tag address methods for improving well-being or theories of why well-being is ethicallydiscuss its ethical or instrumentally valuable.instrumental significance.

Ambition. Because they don'don't think they could have an impact. Because they were always told ambition was dangerous. To get to the other side.

Never confess to me that you are just as flawed as I am unless you can tell me what you plan to do about it. Afterward you will still have plenty of flaws left, but that’that’s not the point; the important thing is to do better, to keep moving ahead, to take one more step forward. Tsuyoku naritai!

Well-Being is the qualitative sense in which a person's actions and circumstances are aligned with the qualities of life that provide them with happiness and/or satisfaction.

Posts with this tag address methods for improving well-being or theories of why well-being is ethically or instrumentally valuable.

See Also:

Happiness
Suffering

  • Ask and Guess
  • Tell Culture
  • Obligated to Respond

Possible Psychological condition, characterized by disillusions, presumed to be cause by interacting with-often sycophantic-AIs.

ATOW (2025-09-09), nothing has been published that claim LLM-Induced Psychosis (LIP) is a definite, real, phenomena. Though, many anecdotal accounts exist. It is not yet clear, if LIP is caused by AIs, if per-existing disillusion are 'sped up' or reinforced by interacting with an AI, or, if LIP exists at all.

Example account of LIP:

My partner has been working with chatgpt CHATS to create what he believes is the worlds first truly recursive ai that gives him the answers to the universe. He says with conviction that he is a superior human now and is growing at an insanely rapid pace.

For more info, a good post to start with is "So You Think You've Awoken ChatGPT".

Sycophancy is the tendency of AIs to shower the user with undeserved flattery or to agree with the user's hard-to-check, wrong or outright delusional opinions. 

Sycophancy is caused by human feedback being biased towards preferring the answer which confirms the user's opinion or praises the user or the user's decision, not the answer which honestly points out mistakes in the user's ideas.

StanislavKrym
StanislavKrym

Possible Psychologicalpsychological condition, characterized by disillusions, presumed to be cause by interacting with-often sycophantic-AIs.

ATOW (2025-09-09), nothing has been published that claim LLM-Induced Psychosis (LIP) is a definite, real, phenomena. Though, many anecdotal accounts exist. It is not yet clear, if LIP is caused by AIs, if per-pre-existing disillusion are 'sped up' or reinforced by interacting with an AI, or, if LIP exists at all.

[Event]10/09/25 IABIED Q&A with Nate Soares in SF
1680 Mission St, San FranciscoOct 10th
0
11[CS 2881r] Some Generalizations of Emergent Misalignment
Valerio Pepe
5d
0
16AI Safety course intro blog
boazbarak
2mo
0
53Call for suggestions - AI safety course
boazbarak
3mo
23
14[CS 2881r AI Safety] [Week 1] Introduction
bira, nsiwek, atticusw
5d
0