Measure is unceasing

Memo on the grain of truth problem

Frameworks that try to predict reality can fail when they miss crucial details. Frameworks about AI have each failed in turn. For instance:

As a result, it doesn’t make sense to trust frameworks too much. From where I’m standing, it makes sense to bake in the ability to rapidly deal with surprises and to prospect possibilities from the bottom up by looking at events in the world.

Definition of the grain of truth problem

What if the correct hypothesis is not in your hypothesis space? Then you may not be able to identify dangers effectively, and you can’t process information in a straightforwardly Bayesian manner.

In judgmental forecasting

I’ve spent much of my professional life working within the judgmental forecasting and Bayesian traditions of knowledge, as presented by Phil Tetlock of E.T. Jaynes.

In the Tetlock tradition, forecasters aim to not be “hedgehogs”, who know one big thing, but rather “foxes”, who know many things. Aspiring forecasters come to know that a totalizing worldview is bad for accuracy. This lesson seems to have been lost somewhere along the way.

In the Bayesian tradition, one can update on the probability of A after learning B is:

$$ P(A | B) = \frac{P(A) \cdot P(B | A)}{P(B)} $$

however, if P(B) ~ 0, this rule breaks. If you didn’t think of A beforehand, you might be in that situation.

Why the grain of truth problem might be common

Some possible solutions and their limitations

Bound domain of applicability of models: When Nate Silver was predicting the chance of a Biden win, he was actually predicting the chance of a Biden win, conditional on a Biden/Trump race. He could have made that conditioning more apparent. Limitations: Good epistemic habit, but doesn’t solve the underlying problem.

Various mathy tricks: One can notice when the in-model probability is too low, pessimize over relevant hypothesis, set formalisms to allow considering hypothesis over only parts of the world, solve for a special case, etc. Limitation: Technical, may require well-specificed model, solutions often don’t translate well to judgmental forecasting practice.

Try hard to find a grain of truth. Limitations: Ennumerating many hypothesis exhaustively ex-ante tends to either become very messy or remain too abstract, and can fail if a real world event falls in a category boundary. Slicing the world into mutually exclusive options fails when you have imperfect categories and you get something on a fuzzy boundary1. Observing reality very closely to generate new hypotheses seems like a much better option, but is also costly.

Accept presence of unknowns, invest in rapid action. Good way to cope with the problem in practice, but also costly. Research communities may also have forgotten how to act in the real world.

Overall there doesn’t seem to be a silver bullet, only mitigations that are costly to different degrees.

Takeaways

The above was adapted from a talk I gave at Newspeak House. Thanks to attendees for discussion of various “Dewey Decimal Systems for reality”, and particularly for pointing me in the direction of metarationality.


  1. E.g., does hitting a plastics facility count as hitting Iran’s nuclear programme? Probably, but it’s unclear.