John Fisher's Shortform
Nov 27, 20241
Typically, a definition for “Alignment” includes something like “systems that pursue objectives matching the ones intended by its creator(s)." and agents that are not aligned are “systems that pursue objectives other than the ones intended by its creator(s)." To be more clear though, throughout this article I'll use the term...