Lately, much has been discussed about PauseAI, or even stopping research completely, until further progress has been made in theory or technical approaches to alignment. After thinking about this for some time, I wondered if there was a way to formalize this reasoning in mathematical terms when I stumbled upon what might be an interesting, possibly novel approach to alignment:
What if we leveraged the nature of slower computing substrates to run AI at a slower pace than current digital computers?
By "slower substrate", I don't mean just diminishing CPU/GPU clock speeds, number of cores, or RAM/VRAM. I mean choosing fundamentally slower forms of computing that would be impossible to speed up past some... (read 812 more words →)
Regarding the second issue (the point that the LLM may not know how to play at different time ratios) - I wonder whether this may be true for current LLM and DRL systems where inference happens in a single step (and where intelligence is derived mainly from training), BUT potentially not true for the next crop of reasoners, where a significant portion of intelligence is derived from the reasoning step that happens at inference time and which we are hoping to target with this scheme. One can imagine that sufficiently advanced AI, even if not explicitly trained for a given task, will eventually succeed at that task, by reasoning about it and... (read more)