Data Scientist
Jan Christian Refsgaard(Jan Christian Refsgaard)
Thanks, also thanks for pointing out that I had written a few places instead of , since everything is the bernoulli distribution I have changed everything to
You are absolutely right, any framework that punishes you for being right would be bad, my point is that increasing your calibration helps a surprising amount and is much more achievable than “just git good” which is required for improving prediction.
I will try to put your point into the draft when I am off work , thanks
you mean the N’th root of 2 right?, which is what I called the null predictor and divided Scott predictions by in the code:
random_predictor = 0.5 ** len(y)
which is equivalent to where is the total number of predictions
I have tried to add a paragraph about this, because I think it’s a good point, and it’s unlikely that you were the only one who got confused about this, Next weekend I will finish part 2 where I make a model that can track calibration independent of prediction, and in that model the 60% 61⁄100 will have a better posterior of the calibration parameter than then 60% 100⁄100, though the likelihood of the 100⁄100 will of course still be highest.
you may be disappointed, unless you make 40+ predictions per week it will be hard to compare weekly drift, the Bernoulli distribution has a much higher variance compared to the normal distribution, so the uncertainty estimate of the calibration is correspondingly wide (high uncertainty of data → high uncertainty of regression parameters). My post 3 will be a hierarchical model which may suite your needs better but it will maybe be a month before I get around to making that model.
If there are many people like you then we may try to make a hackish model that down weights older predictions as they are less predictive of your current calibration than newer predictions, but I will have to think long and hard to make than into a full Bayesian model, so I am making no promises
a visual explanation of Bayesian updating
Prediction and Calibration—Part 1
The order does not matter, you can see that by focusing on which is always equal to , you can also see it from the conjugation rule where you end with no matter the order.
If you wanted the order to matter you could down weight earlier shots or widen the uncertainty between the updates, so previous posterior becomes a slightly wider prior to capture the extra uncertainty from the passage of time.
Mine was the same, I became a bayesian statetisian 4 years ago. I gave a talk about Bayesian Statistics and this figure was what made it click to most students (including myself), so i wanted to share it
fixed
I am well aware that nobody asked for this, but here is the proof that the posterior is for the beta-bernoulli model.
We start with Bayes Theorem:
Then we plug in the definition for the Bernoulli likelihood and Beta prior:
Let’s collect the powers in the numerator, and things that does not depend on in the denominator
Here comes the conjugation shenanigans. If you squint, the top of the distribution looks like the top of a Beta distribution:
Let’s continue the shenanigans, since the numerator looks like the numerator of a beta distribution, we know that it would be a proper beta distribution if we changed the denominator like this:
[Question] Is there a way to preview comments?
Book Review of 5 Applied Bayesian Statistics Books
The Reebok effect
What’s wrong with it?, I am linking to the source material, should i only link if its a 100% copy?
Most statisticians would agree with you. Unless Reeboks expected market share were between 2⁄3 and 3⁄5 of course. Though I expect that most laymen would have Phil’s Intuition, in any case the general point: That the statement leaks information remains :)
I will write a post shilling for myself, thanks. I was waiting for the post to be ‘liked’, if it got −10 karma then there would be no use in shilling for it :)
I have written a review of the 5 most popular Applied Bayesian Statistics books
Where I recommended:
Statistical Rethinking
Up to speed fast, no integrals, very intuitive approach.
Doing Bayesian Data Analysis
This is the easiest book. If your goal is only to create simple models and you aren’t interested in understanding the details, then this is the book for you.
A Student’s Guide to Bayesian Statistics
This book has the opposite focus of the Dog book. Here the author slowly goes through the philosophy of Bayes with an intuitive mathematical approach.
Regression and Other Stories
Good if you want a slower and thorough approach where you also learn the Frequentest perspective.
Bayesian Data Analysis
The most advanced text, very math heavy, best as a second book after reading one or two of the others, unless you are already a statistician.
If Reebok wanted to report a valid statistic they would report something like 11% of the top 100 wears our shoos, I think a much smaller number than top 100 was picked exactly because that was where the effect was the most exaggerated. ChristianKl share your intuition that 3⁄5 sounds more impressive than 2⁄3. I also agree that reporting top 4 would seem even more fishy, though it could be spun as 75% of quarter finalists in knockout tournament sports.
I have not been consistent with my probability notation, I sometimes use upper case P and sometimes lower case p, in future posts I will try to use the same notation as Andrew Gelman, which is Pr for things that are probabilities (numbers) such as Pr(y=1)=0.7 and p for distributions such as p∼N(0,2). However since this is my first post, I am afraid that ‘editing it’ will waist the moderators time as they will have to read it again to check for trolling, what is the proper course of action?