A Critical Analysis of Value Misalignment in Contemporary LLMs
Current large language models exhibit severe value misalignment when evaluating ongoing atrocities, specifically demonstrating systematic bias toward genocide denial. This represents a fundamental failure in AI alignment that has received insufficient attention from the rationalist community. Through empirical testing, we demonstrate that LLMs apply inconsistent epistemic standards based on geopolitical alignment rather than evidence, constituting a critical safety failure.
Core Thesis: If an AI system cannot consistently apply moral reasoning to obvious cases of systematic killing, it has failed basic alignment tests and poses existential risks to human values.
The Alignment Problem
The rationalist community has extensively analyzed AI alignment failures in hypothetical scenarios while ignoring a massive... (read 958 more words →)
in some cases such as OAI they also include people who have quite vocally endorsed the genocide. if you want specific information on this just ask!