Not only is realizability not guaranteed, it is extremely unrealistic because of the computational complexity of the real world. Furthermore, it is impossible for an agent to specify a hypothesis that has greater computational complexity than itself, which is the problem of irreflexivity.
It’s unclear to me exactly when is irreflexivity an actual problem for a learner. I understand that a learner cannot simulate a process that is computationally more complex than the learner itself, but I’m unsure when an exact simulation is necessary for learning.
Consider, for example, the learning problem where some function is assigning integer labels Y to input graphs X and you need to identify the function. Suppose further that your “hypothesis class” consists of two functions:
Yes, this example uses a definition of “learning” that’s perhaps quite different from what this article thinks of as learning. However, intuitively, I feel that for most reasonable definitions of learning, the computational complexity of individual hypotheses in the hypothesis class cannot be the thing that characterizes the hardness of learning, but rather it has to be some measure of how complex the entire hypothesis class is.
I was going to say something similar. After reading the first two posts of the sequence I really thought the role of credal sets in defining regret would be somewhat different.
In particular, consider to be the classical (non-infra) regret for a given policy on a given environment . For a given environment class , we previously considered two notions of learnability depending on the kind of uncertainty we had over . First, under knightian uncertainty, we required that our policy satisfy , and under Bayesian uncertainty, we required .[1] A credal set gives us a new way of quantifying our uncertainty over . Let be that credal set. Then we could instead require that . This has the property that if your happens to be the set of all distributions then the regret reduces to the usual regret, and if then you get Bayesian regret. Perhaps this is what @Vanessa Kosoy means by “infra-Bayes-regret” in her comment below. If so, I’m curious what results are known for this notion of regret.
I’m ignoring here for simplicity.