LLMs could be as conscious as human emulations, potentially — LessWrong