Talk:2701: Change in Slope

Explain xkcd: It's 'cause you're dumb.
Revision as of 04:54, 26 November 2022 by (talk) (Raw Data: 69 dots)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search

I am an occasional data scientist, and I can confirm this is why we have monitor stands that tilt. 16:33, 21 November 2022 (UTC)

The third e in "neeed" in the title text seems to be a typo Victor (talk) 16:41, 21 November 2022 (UTC)

I think Randall may have added it to represent that the speaker prolongs the "e" sound for emphasis, although that's usually done with 4-5 e's. Barmar (talk) 16:53, 21 November 2022 (UTC)
I had to double-check this, myself (presumed the 'Bot created the lage faithfully, but went straight to source to see if I needed to find a vandalism post to revert). May need a comment (to prevent hypercorrection, if not to note the implied emphisis) and certainly will if it turns out to be a typo and gets corrected (for which I'm sure a future checker will discover Randall's revisiting, but then worth a note to that effect). 17:42, 21 November 2022 (UTC)
I guess Randall fixed it, because I'm only seeing 2 'e's in the title text. Just updated it on the wiki. Zman350x (talk) 01:26, 22 November 2022 (UTC)

Bender Bot was one of the main characters in Futurama. Barmar (talk) 16:54, 21 November 2022 (UTC)

Just donning my unnecessary pedantry hat for a moment: his name is Bender Bending Rodriguez --192·168·0·1 (talk) 23:02, 21 November 2022 (UTC)

A couple(?) of authors used the word(s) "(point of) inflection", which is not really suitable for a join between two straight segments. Was tempted to talk about "discontinuity", but that really only applies to the meta-slope (derivatives, to one degree or other) where it suddenly jumps (at a point), or the derivative's derivative has jumps (as it enters and leaves the smoothly linking curve). Hope it works well enough how I left it, though. 21:28, 21 November 2022 (UTC)

For anyone curious, I used an image editor to turn the entire comic sideways and it actually does seem to work, to some degree anyway. SSM24 23:37, 21 November 2022 (UTC)

Added; thanks! 00:14, 22 November 2022 (UTC)
If you don't mind sharing: which program did you use? Did you tweak things like relative distance / camera FOV, to effectively select a specific point in the continuum that makes up the Dolly Zoom effect, and at the limit on one end results in orthographic projection? (Edit 10 minutes later: a better article to look at is Perspective distortion (photography)) Or did you just leave it at whatever the default is? Can you recreate the image with the two extremes, and share them? And lastly - can you upload the image (and potentially the new images) to the wiki directly, so they can be embedded in the page? Thanks! --NeatNit (talk) 17:21, 22 November 2022 (UTC)

This one shows the beauty of Explainxkcd: people reading the explanation are likely to learn accessible methods of substantial practical utility. 00:38, 22 November 2022 (UTC)

Hey, if it works for picking out lumber at Lowe’s, why not for graphs, too? - MadMarie

There was an old bit of explanation that related it to examining physical objects (for dent/bend-removal in metalwork, I think it was) that got wiped out by a later edit. Though I'm considering my own version, now generalised to cover your experience, as it seems quite relevant/analogous to me. 14:37, 22 November 2022 (UTC)

Whoever wrote the 1st explanation needs to go touch grass and learn how real people talk, pissed me off so much I just effectively rewrote the whole thing from scratch 06:34, 22 November 2022 (UTC)

Intrigued, looking at the first explanation (give or take that person's initial small errors/omissions) I personally find it more to the point than what it has become. Not to say the complete rewrite was wrong, but it got it not that much closer to the mythical perfection. IMO. 20:29, 22 November 2022 (UTC)

Going in a different direction than "this is silly" - if we ignore the "viewing point/parallax" issue, doing a change of basis like this is similar to linear methods like [SVD] & PCA, and considering the graph as a mappingg in a "higher dimension" is similar to the "kernel trick" popularized by Support Vector Machines 11:31, 22 November 2022 (UTC)

Raw Data[edit]

I love this cartoon. This is definitely something that was relevant in my work!

At my old job I had some commercial or public-domain software for extracting the raw data behind a scatter plot. If anyone has something like that handy, I would love to see someone extract the data behind the graph on the left, so that we can:

  1. Apply the affine transformation which generates the image on the right with the tilted paper.
  2. Apply the statistical tests which Randall Munroe is alluding to.
Knock yourself out:
104 points. 19:17, 22 November 2022 (UTC)
I only count 69 distinct dots, although a handful look like they might be merged pairs. What's up with that? 04:54, 26 November 2022 (UTC)
Can someone please check my work and tell me if I'm doing it right? I'm pretty sure I don't really know what I'm doing. I kind of cargo cult-coded the Savitzky-Golay filter stuff linked from the explanation and have zero understanding of what's actually going on. 21:58, 22 November 2022 (UTC)
Here's how Randall seems to be suggesting to do it, based on the light gray figures: [superceded] -- Can someone please help fix the residuals on the second plot? 01:18, 23 November 2022 (UTC)
I fixed the residuals and added an inset confidence interval comparisons for the two slopes, split by both their maximum difference and by the maximum sum of the r2 values: It's not clear from the gray text which method Randall is suggesting. 22:07, 24 November 2022 (UTC)
I added this to the end of the Colab notebook:
# Later in the Explainxkcd explanation, a "Significance of the Difference between Two Slopes Calculator"
# at is recommended, so ... we get:

# split by maximum slope difference:  (as shown in green and red)
#   t-Value:               5.52246856
#   Degrees of freedom:  100
#   Probability:           0.00000027 (significant as < 0.05)

# split by maximum sum of r²s:
#   t-Value:               6.25478825
#   Degrees of freedom:  100
#   Probability:           0.00000001 (also very significant)

# So, while the latter might technically be  about 27 times more likely, both represent undoutably
# different linear fits. Perhaps someone can ask Randall which he was suggesting, if indeed either?
What's the most reliable way to ask Randall this? Twitter? Email? Google Chat? 23:08, 24 November 2022 (UTC)
Why don't you generate a series of mildly noisy datasets of two slightly different but random lines each and see which method gets closest to the generating parameters? Also, please put more blank lines in your code, and consider right-aligning the comments. 01:56, 26 November 2022 (UTC)

What's funny is people are doing a lot of statistics and computer magic when you can just tilt your screen like the comic says and get the same effect :P (talk) 16:14, 25 November 2022 (please sign your comments with ~~~~)

(Ɔ┴∩) ᄅᄅ0ᄅ ɹǝqɯǝʌoN ϛᄅ 'ㄣϛ:ㄥƖ ᄅᄅᄅ˙ᄅ9Ɩ˙0ㄥ˙ᄅㄥƖ ¡ƃuoɹʍ ʇᴉq ɐ ʇuǝʍ ƃuᴉɥʇǝɯos ʇnq 'ʇɐɥʇ pǝᴉɹʇ I