Elementary Infra-Bayesianism

https://www.lesswrong.com/posts/9uj2Mto9CNdWZudyq/elementary-infra-bayesianism

Link post Contents

Why Infra-Bayesianism?

Engaging with the work of Vanessa Kosoy is a rite of passage in the AI Safety space. Why is that?

Uncertain updates

Imagine it’s late in the night, the lights are off, and you are trying to find your smartphone. You cannot turn on the lights, and you are having a bit of trouble seeing properly[2]. You have a vague sense about where your smartphone should be (your prior, panel a). Then you see a red blinking light from your smartphone (sensory evidence, panel b). Since your brain is really good at this type of thing, you integrate the sensory evidence with your prior optimally (despite your disinhibited state) to obtain an improved sense of where your smartphone might be (posterior, panel c). That’s just boring old Bayes, nothing to see here, move along.\mathbb{P}(S|E) = \frac{\mathbb{P}(E|S)\mathbb{P}(S)}{\mathbb{P}(E)}Now let’s say you are even more uncertain about where you put your smartphone.[3] It might be one end of the room or the other (bimodal prior, panel a). You see a blinking light further to the right (sensory evidence, panel b), so your overall belief shifts to the right (bimodal posterior, panel c). Importantly, by conserving probability mass, your belief that the phone might be on the left end of the room is reduced. The absence of evidence is evidence of absence. This is still only boring old Bayes. To go Infra, we have to go weird.# Fundamentally uncertain updates Let’s say you are really, fundamentally unsure about where you put your phone. If someone were to put a gun to your head threaten to sign you up for sweaters for kittens unless you give them your best guess, you could not.[4] This is the situation Vanessa Kosoy finds herself in[5].[6] With Infra-Bayesianism, she proposes a theoretical framework for thinking in situations where you can’t (or don’t want to) specify a prior on your hypotheses. Because she is a mathematician, she is using the proper terminology for this:

You are unsure about the location of your smartphone (and mortally afraid to get it wrong). You follow the red blinking light, but you never discard your alternative hypothesis that the smartphone might be at the other end of the room. At the slightest indication that something is off you’ll discard all the information you have collected and start the search from scratch. This is a very cautious strategy, and it might be appropriate when you’re in dangerous domains with the potential for catastrophic outliers, basically what Nassim Taleb calls Black Swan events. I’m not sure how productive this strategy is, though; noise might dramatically mess up your updates at some point.

Closing thoughts

This concludes the introduction to Elementary Infrabayesianism. I realize that I have only scratched the surface of what’s in the sequence, and there is more coming out every other month, but letting yourself get nerd-sniped is just about as important as being able to stop working on something and publish. I hope what I wrote here is helpful to some, in particular in conjunction with the other explanations on the topic (1 2 3) which go a bit further than I do in this post. I’m afraid at this point I’m obliged to add a hot take on what all of this means for AI Safety. I’m not sure. I can tell myself a story about how being very careful about how quickly you discard alternative hypotheses/​narrow down the hypothesis space is important. I can also see the outline of how this framework ties in with fancy decision theory. But I still feel like I only scratched the surface of what’s there. I’d really like to get a better grasp of that Nirvana trick, but timelines are short and there is a lot out there to explore.

Comment

https://www.lesswrong.com/posts/9uj2Mto9CNdWZudyq/elementary-infra-bayesianism?commentId=toDKAg9mbgw9cHG3A

I upvoted because distillations are important, but a first pass over your post left me much more confused than I was before. Another issue was the level of formality. Either use less, and stick to core intuitions, or use more. Personally, I would have liked to know the types of all the objects being mentioned, and a mention of what space X was (I’m assuming the space of possible events?), and some explanation of how P_{g}^{H} relates to P_+/P_- or even g. In fact, I can’t see how g relates to any of the objects. EDIT: I meant to say, you didn’t say how g relates to any of the distributions or update rules or whatever except via an unclear analogy. Though note, I’m pretty tired right now so take my confusion as not indicative of the average reader.

Comment

https://www.lesswrong.com/posts/9uj2Mto9CNdWZudyq/elementary-infra-bayesianism?commentId=tpqsA5wTHws3L8J2v

Thank you for your comment! You are right, these things are not clear from this post at all and I did not do a good job at clarifying that. I’m a bit low on time atm, but hopefully, I’ll be able to make some edits to the post to set the expectations for the reader more carefully. The short answer to your question is: Yep, X is the space of events. In Vanessa’s post it has to be compact and metric, I’m simplifying this to an interval in R. And P_+/P_- can be derived from P_g^H by plugging in g=0 and replacing the measure m(A) by the Lesbegue integral \int_A\mathrm{d}m. I have scattered notes where I derive the equations in this post. But it was clear to me that if I want to do this rigorously in the post, then I’d have to introduce an annoying amount of measure theory and the post would turn into a slog. So I decided to do things hand-wavy, but went a bit too hard in that direction.