Editing Talk:1132: Frequentists vs. Bayesians

Jump to: navigation, search
Ambox notice.png Please sign your posts with ~~~~

Warning: You are not logged in. Your IP address will be publicly visible if you make any edits. If you log in or create an account, your edits will be attributed to your username, along with other benefits.

The edit can be undone. Please check the comparison below to verify that this is what you want to do, and then save the changes below to finish undoing the edit.
Latest revision Your text
Line 1: Line 1:
I just sort of assumed he bet 50 dollars because if the sun had exploded, they'd be dead and therefore wouldn't need the machine. [[Special:Contributions/108.162.237.16|108.162.237.16]] 07:05, 21 August 2020 (UTC)
 
 
 
Something should be added about the prior probability of the sun going nova, as that is the primary substantive point. "The neutrino detector is evidence that the Sun has exploded. It's showing an observation which is 35 times more likely to appear if the Sun has exploded than if it hasn't (likelihood ratio of 35:1). The Bayesian just doesn't think that's strong enough evidence to overcome the prior odds, i.e., after multiplying the prior odds by 35 they still aren't very high." - http://lesswrong.com/r/discussion/lw/fe5/xkcd_frequentist_vs_bayesians/ [[Special:Contributions/209.65.52.92|209.65.52.92]] 23:51, 9 November 2012 (UTC)
 
Something should be added about the prior probability of the sun going nova, as that is the primary substantive point. "The neutrino detector is evidence that the Sun has exploded. It's showing an observation which is 35 times more likely to appear if the Sun has exploded than if it hasn't (likelihood ratio of 35:1). The Bayesian just doesn't think that's strong enough evidence to overcome the prior odds, i.e., after multiplying the prior odds by 35 they still aren't very high." - http://lesswrong.com/r/discussion/lw/fe5/xkcd_frequentist_vs_bayesians/ [[Special:Contributions/209.65.52.92|209.65.52.92]] 23:51, 9 November 2012 (UTC)
  
Line 20: Line 18:
  
 
Here's what I get for the application of Bayes' Theorem:
 
Here's what I get for the application of Bayes' Theorem:
: P(N|Y) = P(Y|N) * P(N) / P(Y): = P(Y|N) * P(N) / [P(Y|N) * P(N) + P(Y|~N) * P(~N)]
+
: P(N|Y) = P(Y|N) * P(N) / P(Y)
 +
: = P(Y|N) * P(N) / [P(Y|N) * P(N) + P(Y|~N) * P(~N)]
 
: = 35/36 * P(N) / [35/36 * P(N) + 1/36 * (1 - P(N))]
 
: = 35/36 * P(N) / [35/36 * P(N) + 1/36 * (1 - P(N))]
 
: = 35 * P(N) / [35 * P(N) - P(N) + 1]
 
: = 35 * P(N) / [35 * P(N) - P(N) + 1]
Line 36: Line 35:
 
:Yes, you would be able to ask. While neutrinos move almost at speed of light, the plasma of the explosion is significally slower, [http://en.wikipedia.org/wiki/Supernova 10% of speed of light tops]. You will have more that hour to ask. (Note that technically, sun can't go [http://en.wikipedia.org/wiki/Nova nova], because nova is white dwarf with external source of hydrogen. It can (and will), however, go supernova, which I assume is what Randall means.) -- [[User:Hkmaly|Hkmaly]] ([[User talk:Hkmaly|talk]]) 09:19, 12 November 2012 (UTC)
 
:Yes, you would be able to ask. While neutrinos move almost at speed of light, the plasma of the explosion is significally slower, [http://en.wikipedia.org/wiki/Supernova 10% of speed of light tops]. You will have more that hour to ask. (Note that technically, sun can't go [http://en.wikipedia.org/wiki/Nova nova], because nova is white dwarf with external source of hydrogen. It can (and will), however, go supernova, which I assume is what Randall means.) -- [[User:Hkmaly|Hkmaly]] ([[User talk:Hkmaly|talk]]) 09:19, 12 November 2012 (UTC)
  
:: Our sun will not go supernova, as it has insufficient mass.  It will slowly become hotter, rendering Earth uninhabitable in a few billion years.  In about 5 billion years it will puff up into a red giant, swallowing the inner planets.  After that, it will gradually blow off its lighter gasses, eventually leaving behind the core, a white dwarf. [[Special:Contributions/50.0.38.245|50.0.38.245]] 01:58, 15 November 2012 (UTC)
+
:: Our sun will not go supernova, as it has insufficient mass.  It will slowly become hotter, rendering Earth uninhabitable in a few billion years.  In about 5 billion years it will puff up into a red giant, swallowing the inner planets.  After that, it will gradually blow off its lighter gasses, eventually leaving behind the core, a white dwarf. [[User:dcmeserve|dcmeserve]]
 +
 
 +
:::I left your comment here so I can set you straight on something. '''''DO NOT EVER''''' edit any editor's comments on a discussion page. You can reply to their comment, but you do not edit another person's words. You do that again, you get the banhammer. [[User:Lcarsos|lcarsos]] ([[User talk:Lcarsos|talk]]) 17:38, 13 November 2012 (UTC)
 +
:::: Quote from the instructions at the bottom of the discussion edit page: "Please note that all contributions to explain xkcd '''may be edited, altered, or removed by other contributors'''. If you do not want '''your writing to be edited mercilessly''', then do not submit it here." But don't worry, I won't be contributing again, if it can be met with this kind of attitude.[[User:dcmeserve|dcmeserve]]
 +
 
 +
:::::If you'll note, those instructions are on the bottom of every edit page. Indeed, the notice is intended to be for article pages, where it is encouraged that an editor with an improvement, improve upon the words of another editor. However, on a talk page, discourse is meant to be conducted, by editors for the betterment of the article. For constructive discourse to occur, a person's words must be left in tact. The act of censorship hurts the common goal of betterment. Per [http://en.wikipedia.org/wiki/Wikipedia:Talk_page_guidelines#Editing_comments Wikipedia], the authoritative source on how a wiki works best: "you ''should not'' edit or delete the comments of other editors without their permission." I encourage, nay, ''implore'', you to assist in the improvement this wiki. But please, do it without editing other people's comments on a talk page, that's simply rude. [[User:Lcarsos|lcarsos]]<span title="I'm an admin. I can help.">_a</span> ([[User talk:Lcarsos|talk]]01:08, 15 November 2012 (UTC)
  
:::Please don't edit others' comments on talk pages; it's considered quite rude. On a talk page, discourse is meant to be conducted, by editors for the betterment of the article. For constructive discourse to occur, a person's words must be left in tact. The act of censorship hurts the common goal of betterment. Per [http://en.wikipedia.org/wiki/Wikipedia:Talk_page_guidelines#Editing_comments Wikipedia], the authoritative source on how a wiki works best: "you ''should not'' edit or delete the comments of other editors without their permission." [[User:Lcarsos|lcarsos]]<span title="I'm an admin. I can help.">_a</span> ([[User talk:Lcarsos|talk]])  17:38, 13 November 2012 (UTC) <small>Note: much of this conversation has been removed at the request of the authors.</small>
+
::::::This was my first time contributing to a wiki discussion page.  There was nothing written here to lead me to believe that it was different in this regard from any other wiki page; quite the opposite, as I mentioned. Perhaps you can find a way to describe this exception in the instructions at the bottom of the edit page. I'm sure you found out or helped form this bit of wiki etiquette long ago, but people who are new to contributing should not be verbally assaulted for not having heard of it already. And by a main editor, no less.[[User:dcmeserve|dcmeserve]]
  
 
I think the explanation is wrong or otherwise lacking in its explanation: The P-value is not the entire problem with the frequentist's viewpoint (or alternatively, the problem with the p-value hasn't been explained). The Frequentist has looked strictly at a two case scenario: Either the machine rolls 6-6 and is lying, or it doesn't rolls 6-6 and it is telling the truth. Therefore, there is a 35/36 probability (97.22%) that the machine is telling the truth and therefore the sun has exploded. The Bayesian is factoring in outside facts and information to improve the accuracy of the probability model. He says "Either the machine rolls 6-6 (a 1/36 probability, or 2.77%) or the sun has exploded (an aparently far less likely scenario). Given the comparison, the Bayesian believes it is MORE probable that the machine rolled 6-6 than the sun exploded, given the relative probabilities. If the latter is a 1 in a million chance (0.000001%), it is 2,777,777 times more likely that the machine rolled 6-6 than the sun exploded.
 
I think the explanation is wrong or otherwise lacking in its explanation: The P-value is not the entire problem with the frequentist's viewpoint (or alternatively, the problem with the p-value hasn't been explained). The Frequentist has looked strictly at a two case scenario: Either the machine rolls 6-6 and is lying, or it doesn't rolls 6-6 and it is telling the truth. Therefore, there is a 35/36 probability (97.22%) that the machine is telling the truth and therefore the sun has exploded. The Bayesian is factoring in outside facts and information to improve the accuracy of the probability model. He says "Either the machine rolls 6-6 (a 1/36 probability, or 2.77%) or the sun has exploded (an aparently far less likely scenario). Given the comparison, the Bayesian believes it is MORE probable that the machine rolled 6-6 than the sun exploded, given the relative probabilities. If the latter is a 1 in a million chance (0.000001%), it is 2,777,777 times more likely that the machine rolled 6-6 than the sun exploded.
Line 48: Line 52:
  
 
The Labyrinth reference reminds me of an old Doctor Who episode (Pyramid of Mars), where the Doctor is also faced with a truthful and untruthful set of guards. Summarized here: http://tardis.wikia.com/wiki/Pyramids_of_Mars_(TV_story) [[User:Fermax|Fermax]] ([[User talk:Fermax|talk]]) 04:49, 14 November 2012 (UTC)
 
The Labyrinth reference reminds me of an old Doctor Who episode (Pyramid of Mars), where the Doctor is also faced with a truthful and untruthful set of guards. Summarized here: http://tardis.wikia.com/wiki/Pyramids_of_Mars_(TV_story) [[User:Fermax|Fermax]] ([[User talk:Fermax|talk]]) 04:49, 14 November 2012 (UTC)
 
This is actually an example of the [https://en.wikipedia.org/wiki/Base_rate_fallacy Base rate fallacy]. --[[Special:Contributions/71.199.125.210|71.199.125.210]] 04:04, 19 November 2012 (UTC)
 
 
People have gone over this already, but just to be a bit more explicit:
 
Let NOVA be the event that there was a nova, and let YES be the event that the detector responds "Yes" to the question "Did the sun go nova?"
 
What we want is P(NOVA|YES)=P(YES|NOVA)*P(NOVA)/P(YES)
 
Suppose P(NOVA)=p is the prior probability of a nova.
 
Then P(YES|NOVA)=35/36, P(NOVA)=p, and P(YES)=p*35/36+(1-p)*1/36=1/36+34/36
 
So then P(NOVA|YES)=35p/(1+34p). If p is small, then P(NOVA|YES) is also small. In particular, the Bayesian statistician wins his bet at 1:1 odds if p<1/36, which is probably the case.
 
If the Bayesian statistician wants 95% confidence that he'll win his bet, then he needs p<1/666. =P
 
 
 
It's cute to attempt to connect this to the U.S. presidential election, but it's far likelier that it's a reference to Enrico Fermi taking bets at the Trinity test site as to whether or not the first atomic bomb would cause a chain reaction that would ignite the entire atmosphere and destroy the planet.  I'll bet you $50 it is.  [[Special:Contributions/71.229.88.206|71.229.88.206]] 21:29, 7 March 2013 (UTC)
 
 
I don't like the explanation at all. Some of the discussion posts give a good view on this. I'd like to share my thought about the last panel, though. The page reads as if the punch line is about the fact that you cannot spend the money if the sun was going to explode; but why does the bayesian propose this bet and not the frequentist - no reason for this. I think there is a better explanation for this panel: there are several proofs that bayesian probabilities result in "rational" behaviour: They state that if you act according to bayes' rule you cannot be cheated in betting. [[Special:Contributions/108.162.254.179|108.162.254.179]] 17:11, 6 March 2014 (UTC)
 
 
The last panel may refer to Nate Sliver's view expressed in his book {{w|The Signal and the Noise}} that if one believes one's prediction to be true one should be confident to bet on it. --[[User:Troy0|Troy0]] ([[User talk:Troy0|talk]]) 18:46, 6 July 2014 (UTC)
 
 
Please excuse my ignorance, but how is two sixes rolled on fair dice 31/32?  (In the explanation: "the detector is telling the truth (31 in 32)") --[[User:Pudder|Pudder]] ([[User talk:Pudder|talk]]) 17:06, 9 December 2014 (UTC)
 
:Just a missreading, not stupid. The detector is telling the truth when you dont role 2 sixes. roling 2 sixes is 1/6 * 1/6 or 1/36. So not roling is 35 in 36, wait oops 36 not 32, thanks. [[Special:Contributions/108.162.216.209|108.162.216.209]] 17:39, 9 December 2014 (UTC)
 
 
I have always thought that the suggested Bet is also a reference to the Dutch Book argument for judging and accounting for probabilities underlying Bayesian interpretations of probability theory. [[Special:Contributions/141.101.98.100|141.101.98.100]] 22:11, 12 August 2015 (UTC)
 
 
The likelyhood of a solar explosion may be wrong.  Since the detector I'd only used at night, the event is twice as likely to occur than listed.  That said, there's a 50% chance of the event never being detected, so I'm not sure.  Any one more knowledgeable than I care to comment? [[User:Mikemk|Mikemk]] ([[User talk:Mikemk|talk]]) 06:47, 5 April 2016 (UTC)
 
 
Huh. I thought that the last panel was pragmatism: "If the sun goes nova, $50 doesn't matter; I'll be dead. If the sun hasn't, I get $50!"
 
:Same, but sign your comments, [[User:Netherin5|Netherin5]] ([[User talk:Netherin5|talk]]) 14:09, 22 February 2019 (UTC)
 
 
This comic hurts my head. [[Special:Contributions/173.245.54.7|173.245.54.7]] 21:44, 12 August 2019 (UTC)
 
 
It is my feeling that sloppy or machiavellian academics have come to use the term "Bayesian" to mean something more like "we adjusted it to what we felt was most reasonable", which introduces so much bias that it actually leaves one unable to determine the scientific validity of the results. I was reading [https://academic.oup.com/beheco/article/23/3/619/224216 a publication], today, that made me think of that and look up this comic. —[[User:Kazvorpal|Kazvorpal]] ([[User talk:Kazvorpal|talk]]) 21:57, 11 November 2019 (UTC)
 
 
Is there a statistical angle I'm missing to the final part of the mouseover text 'did your brain fall out? [roll] yes...' Or is is purely linguistic between literal and figurative i.e. if his brain has fallen out as in he has made a careless error, then that's true. If it's literally did his brain fall out, is the 'yes' the 97% chance that it's talking about his mistake, or the ~3% chance that it's lying about the literal truth? [[Special:Contributions/172.69.69.244|172.69.69.244]] 14:46, 3 December 2019 (UTC)
 
 
As per Sagan, "Extraordinary claims require extraordinary evidence". [[Special:Contributions/108.162.229.54|108.162.229.54]] 10:42, 12 October 2020 (UTC)
 
 
When he Bayesian  says "Bet you $50 it hasn't.", he is saying that he will probably win the bet.  However, he isn't saying he knows whether the sun has exploded or whether the detector is lying.  What he is saying is roughly equivalent to "If we are playing Texas Holdem and I have a royal flush while you have nothing showing, I am probably going to win and might as well bet what I can.
 
 
If the eventist says "Detector! What would the Bayesian statistician say if I asked him whether I would say the sun had exploded", the Bayesian doesn't know what the detector would say.  (I am changing the wording slightly, but it doesn't make sense to me as stated.)  Therefore, the Bayesian can't give an answer.  The Bayesian's answer would therefore be "I am a neutrino detector (answers are sometimes true and sometimes false), not a labyrinth guard (answers are always true or  always false)".  He then predicts that the Bayesian would say "Seriously, did your brain fall out?"  After somebody hits the button, the detector answers truthfully (the likeliest option), and gives his opinion "YES". [[User:BradleyRoss|BradleyRoss]] ([[User talk:BradleyRoss|talk]]) 01:26, 23 February 2023 (UTC)
 
 
With regard to Bayesian having multiple meanings, this is probably similar to there being a Turing test, a Turing machine, and Turing Complete. [[User:BradleyRoss|BradleyRoss]] ([[User talk:BradleyRoss|talk]]) 01:32, 23 February 2023 (UTC)
 
 
No-one has discussed how to properly apply frequentists statistics to the problem.
 
1: Rerun the test several dozen times.
 
2: Find the 95% confidence interval of the generated data (A Poisson distribution is most appropriate for the modeling of event frequency).
 
3: If the 95% confidence interval includes the value of 1/36 then it supports the null hypothesis that there is no correlation is suggested between the sun and the positive detector results.
 
Bayesian and Frequentist Statistics should yield the same result if handled correctly, because they are basically algebraic rearrangements of each other. --[[Special:Contributions/172.71.154.215|172.71.154.215]] 21:33, 24 March 2023 (UTC)
 

Please note that all contributions to explain xkcd may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see explain xkcd:Copyrights for details). Do not submit copyrighted work without permission!

To protect the wiki against automated edit spam, we kindly ask you to solve the following CAPTCHA:

Cancel | Editing help (opens in new window)

Templates used on this page: