I feel like more people should be concerned about how no interesting functions of the complexity prior normalise but maybe they are and I’m just ignoring them or maybe I’m missing something
“How many jelly beans are in the jar” well if your loss function is inverse-googological or harsher in the difference then for any number you can name you’d regret naming a bigger one because of your prior on very very large number of jelly beans