Wiki-Tags in Need of Work

Poor Air Quality can reduce cognitive functioning[1], lifespans[2] and the techniques to improve air quality are also useful for getting rid of aerosolized respiratory pathogens. Improving air quality can be an impactful global health intervention.[3] Many members of the LessWrong community have also put effort into improving the air quality of their own homes or offices, as an implication of instrumental rationality.... (read more)

DALL-E is a family of machine learning models created by OpenAI that generate images from text descriptions.... (read more)

Grabby Aliens is a theory that explains why humans appeared relatively early in the history of the universe (13.8 billion years after the Big Bang, while the average star will last over five trillion years). It was developed by Robin Hanson, Daniel Martin, Calvin McCarter, and Jonathan Paulson.... (read more)

Eliciting Latent Knowledge is an open problem in AI safety.... (read more)

The AI X-risk Research Podcast is a podcast hosted by Daniel Filan.... (read more)

The belief signaling trilemma (or signaling trilemma for simplicity) points out that (a) people assign reputation based on claims; (b) people want to maintain their reputation; therefore, (c) people warp their claims. This presents a trilemma:... (read more)

In the simple case, explicit reasoning is reasoning which:... (read more)

An Ask Me Anything (AMA) post is an invitation to ask the author questions. ... (read more)

Recent Tag & Wiki Activity

The Roko's basilisk thought experiment was notorious in that it required specific preconditions available nearly exclusively within the Less Wrong community that rendered the the reader vulnerable to this 'memetic hazard'. As such it has drawn derision from critics who feel perception risk from unfriendly AI is overstated within the community.1314

A robust theory of logical uncertainty is essential to a full formalization of UDT.  A UDT agent must calculate probabilities and expected values on the outcome of its possible actions in all possible worlds--sequences of observations and its own actions. However, it does not know its own actions in all possible worlds. (The whole point is to derive its actions.) On the other hand, it does have some knowledge about its actions, just as you know that you are unlikely to walk straight into a wall the next chance you get. So, the UDT agent models itself as an algorithm, and its probability distribution about what it itself will do is an important input into its maximization calculation.

One valuable insight from EDT is reflected in "UDT 1.1" (see article by McAllister in references), a variant of UDT in which the agent takes into account that some of its algorithm (mapping from observations to actions) may be prespecified and not entirely in its control, so that it has to gather evidence and draw conclusions about part of its own mental makeup. The difference between UDT 1.0 and 1.1 is that UDT 1.1 iterates over policies, whereas UDT 1.0 iterates over actions. 

See also:

The Nonlinear Library is a project that uses high high-quality text-to-speech software to read out loud posts from LessWrong and the EA forum. You can read their announcement post here, or listen to it here. For posts read by a human, the LW Curated podcast is a recent project - created and recorded by Solenoid Entity, who spent the last five years editing the SSC podcast and the Metaculus Journal Podcast.

Fashion is a form of self-expression and autonomy at a particular period and place and in a specific context, ofwhich involves clothing, footwear, lifestyle, accessories, makeup, hairstyle, and body posture. It can be used for signalling, and what is "fashionable" changes across history, cultures, and contexts.

Fashion is a form of self-expression and autonomy at a particular period and place and in a specific context, of clothing, footwear, lifestyle, accessories, makeup, hairstyle, and body posture.

Fashion

Fashion is a form of self-expression which involves clothing, footwear, lifestyle, accessories, makeup, hairstyle, and body posture. It can be used for signalling, and what is "fashionable" changes across history, cultures, and contexts.

Spurious Counterfactuals are spuriously low evaluations of the quality of a potential action, which are only provable because they are self-fulfilling (usually due to Lob's theorem). For example, if I know that I go left, then it is logically true that if I went right, I would get -10 utility (because in classical logic, false statements imply any statement). This suggests that if I fully believed that I went left, then I would indeed go left. By Lob's theorem, I indeed go left. 

Building agents who avoid this line of reasoning, despite having full access to their own source code and the ability to logically reason about their own behavior, is one goal of Embedded Agency.

Spurious Counterfactuals

Spurious Counterfactuals are spuriously low evaluations of the quality of a potential action, which are only provable because they are self-fulfilling (usually due to Lob's theorem). For example, if I know that I go left, then it is logically true that if I went right, I would get -10 utility (because in classical logic, false statements imply any statement). This suggests that if I fully believed that I went left, then I would indeed go left. By Lob's theorem, I indeed go left. 

Building agents who avoid this line of reasoning, despite having full access to their own source code and the ability to logically reason about their own behavior, is one goal of Embedded Agency.

Religion is a complex group of human activities — involving tribal affiliation,commitment to higher power,  belief in belief, supernatural claims, and a range of shared group practices such as worship meetings, rites of passage, etc.

Posts that are Not Safe For Work

NSFW

Posts that are Not Safe For Work

Formal decision theories also diverge. For Causal Decision Theory, you can only affect those probabilities that you are causalcausally linked to. Hence, the answer should be 'No'. In Evidential Decision Theory any kind of connection is accounted, then the answer should be 'No'. Timeless Decision Theory answer seems undefined, however Yudkowsky has argued that if the problem is recurrently presented, one should answer 'Yes' on the basis of enhancing its probability of gaining $10000 in the next round. This seems to be Causal Decision Theory prescription as well. Updateless decision theory1 prescribes giving the $100, on the basis your decision can influence both the 'heads branch' and 'tails branch' of the universe.

There is an invite-only Slack for people working on the alignment ecosystem, ifecosystem. If you'd like to join message plex with an overview of your involvement.

There is an invite-only Slack for people working on the alignment ecosystem, if you'd like to join message plex with an overview of your involvement.

AI Alignment Fieldbuilding is the effort to improve the alignment ecosystem. Some priorities include introducing new people to the importance of AI risk, on-boarding them by connecting them with key resources and ideas, educating them on existing literature and methods for generating new and valuable research, supporting people who are contributing, and maintaining and improving the funding systems.

A superintelligence might conclude that other superintelligences would tend to exist because increased intelligence is ana convergent instrumental goal for agents. Given the existence of a superintelligence, acausal trade is one of the tricks it would tend to use.

Once an agent realizes that another agent might exist, there are different ways that might might predict the other agent's behavior, and specifically that the other agent can be an acausal trading partner.

A multipolar scenario is one where no single AI or agent takes over the world.

Featured in the book "Superintelligence" by Nick Bostrom.