AI in a vat: Fundamental limits of efficient world modelling for safe agent sandboxing
Drawing inspiration from the ‘brain in a vat’ thought experiment, this blogpost investigates methods to simplify world models that remain agnostic to the agent under evaluation. This work was done together with Alec Boyd and Manuel Baltieri, with support from the UK ARIA Safeguarded AI programme and the PIBBSS Affiliateship...
The point is that a coarse-graining can turn a quasi-probability into a probability. This is significant because quasi-probabilities are limited in that they cannot be sampled etc.
Does that make sense?