User:IssaRice/Scoring rule
how can we formalize the idea of a rule for scoring predictions?
first pass: statements and probabilities
we can start with a list of statements. each statement makes a yes/no prediction about the future, like "the die will show 3 when rolled". then we have a list of probabilities where is the probability someone assigns to being true. now, reality evaluates each statement, giving us a yes/no answer . our probabilities are scored against this response from reality. so a scoring rule S can be some function of . so the type can be .
if we are an ordinary statistician [1], we might pick a rule like . (this is actually almost the brier score)
second pass: events
in probability theory, we are used to dealing with events and random variables. in the previous section, we naively stated scoring rules in terms of statements and probabilities. but we might try now to phrase things in terms of events.
instead of statements , we could have a list of events . here, is an event expressing the fact that is true. then , where P is the probability measure which encodes our knowledge of what events are likely. is the outcome in some possible world, so . the idea here is we have some implicit sample space of all "possible worlds". then each is a possible world. but this is exactly the idea expressed by our reality function -- we could have had some other reality in which our same probabilities would perform differently.
so our second pass is that we can define a scoring rule as something that takes a list of events , a probability measure (which encodes the numbers , assuming we have access to the events), and a world . so
you might complain that in the previous section, we didn't tell the scoring function anything about what the statements were that were being predicted, whereas here we do (since we pass in the events). this is right, but it's just that here if we don't pass in the events, then just given P and we can't tell how to compute the probabilities.
third pass: non-binary predictions
instead of having all our predictions be about yes/no questions, we could allow more kinds of responses. for instance, instead of six predictions "the next roll of the die will be a i", yes=1/6, no=5/6, for i=1,...,6, we could instead have a single prediction like "the next roll will be a ..." 1: 1/6, 2: 1/6, ..., 6: 1/6.
given one prediction with k options (mutually exclusive and collectively exhaustive), and a second prediction with m options (again mutually exclusive and collectively exhaustive), we could roll them into a single prediction with km options. so in a way, we could work as if there was only one thing to be predicted.
so given probabilities of the k options for a single prediction, a scoring function could take those probabilities, plus information about which option was correct, so some number . so .
here we get to one of the confusions i have about the explanation of scoring rules in [2]. there, we score each probability separately, i.e. . and we assume that . i think it's better to write this as , where is 1 if j=i and 0 otherwise. this makes the dependence on all the probabilities clear.
how do we make sense of stuff like the invariance from Technical Explanation? we first need two partitions and of the sample space. these two partitions induce a third partition where each element of the new partition is the intersection of one element each from As and Bs, e.g. . call this new partition . then we want something like for every (need better notation?).
proving logarithmic scoring is proper
from [3], if is the actual beliefs of the forecaster, and is the submitted probabilities, then to be proper we want . to make sense of this, first just pay attention to the bare p's. we're looping through all possible p's we can submit to the scoring process, and we want to find the p that maximizes the score. but we can't just do this raw kind of maximization, because then we'd obviously get that the best probability is to assign 1 to something if it will happen, and 0 if it won't happen -- but we can't predict the future so well, so this is useless! so instead of maximizing the raw score, we want to maximize the expected score, given our current beliefs . so this maximization process spits out some probabilities that we should submit that will get us a high score, given our current beliefs. now, a score is proper if one way to do good is to just submit your actual beliefs.
for logarithmic scoring the expectation is . differentiating this, setting it to 0, and solving for p [is this even valid here? we need to check for convexity/increasing/whatever optimization stuff but idk how] gets us .