The ability to complete sequences is equivalent to prediction. The way GPT-3 completes sequences is it that it predicts what the next token will be and then it outputs the prediction. You can use the same model on images. In general, the agent, based on all of its input data up to some point, tries to generate future data. If it can predict its own input data reliably that means it has a model of the world which is similar to reality. This is similar to Solomonoff induction. Once you have a good approximation of Solomonoff induction (which is uncomputable), you combine the approximation (somehow) with reinforcement learning and expected utility maximization and get an approximation of AIXI. Since I’m not an expert in reinforcement learning I’m not sure which part is harder, but intuition tells me the hard part of all of this would be approximating Solomonoff induction, and once you have a good world-model, it seems to me it’s relatively straightforward to maximize utility. I hope I’m wrong. (if you think I am please explain why)
See FAQ #4 on MIRI’s website below. Edit: It was written in 2013 so it is probably best viewed as a jumping off point from which you can make further updates based on what has happened in the world since then.
Comment
Yes but GPT-3 offers us new evidence we should try to update on. It’s debatable to say how many bits of evidence that provides, but we can also update based on this Discontinuous progress in history: an update:
Comment
I agree. I’m not sure how much to update on the things you mention or on other things that have happened since 2013, so I think my answer serves as more of jumping off point than something authoritative. I edited it to mention that.
Comment