Why S-values (surprisals) Instead of P-values?
The other day, a colleague asked me an important question. When the hypothesis under consideration is not the null hypothesis of no effect (e.g., the difference between two mean values is 50 and not 0), a P-value equal to 1 indicates that—assuming chance is the only factor at play—we would obtain a "result" greater than or equal to the observed one 100% of the time in numerous future equivalent experiments. So, how can chance - assuming it operates alone - always generate a greater "result" than the observed experimental one? The apparent problem is easily solved by noting that the so-called "result" (or, better, the "statistical result") does not refer to the magnitude of the effect under examination but to the value of the test statistic. If we fix the hypothesis "difference = Δ = 50" and observe a difference of 50, then the test statistic (i.e., the statistical result) turns out to be 0! And what is the probability of obtaining a t...