Talk:1958: Self-Driving Issues

Explain xkcd: It's 'cause you're dumb.
Jump to: navigation, search


I assume the off-panel speaker is Megan, based on their positioning, but not sure what the ruling on the ambiguity is. PvOberstein (talk) 05:47, 21 February 2018 (UTC)

I made a note about the typo in the title text. Also, weird question, does the "created by a BOT" tag mean that the explanation was written by an AI? Or is it a joke I'm missing for some reason? Sorry, kind of a dumb question I guess. 09:04, 21 February 2018 (UTC)

Afaik the "created by a BOT" part is the default text when the bot which is crawling xkcd for a new comic inserts the comic here (and an empty explanation). In the past that part was often deleted when the first real edit was made. Some comics ago a habit evolved to actually change that line in relation to the comic at hand (e.g. "created by a SELF-DRIVING CAR" would be fitting here). Elektrizikekswerk (talk) 11:23, 21 February 2018 (UTC)
Oh, okay...19:06, 21 February 2018 (UTC)
It refers to the incomplete tag. The incomplete tag is created by a bot and just shows that there needs to be an explaination. Lupo (talk) 19:09, 25 February 2018 (UTC)

Could the firmware update be for the humans, because they are obviously malfunctioning in the scenario? Sebastian --162.158.90.30 09:46, 21 February 2018 (UTC)

Of course, that's the joke. It would be really impractical to install such a firmware update, there are about 7.5 billion people on earth - many of which we don't even have access to. I'd also suspect that most people would fight back if you tried shoving a USB flash drive into them. 162.158.93.9 11:24, 21 February 2018 (UTC)
Yeah, but how many of those 7.5 billion people are a safety risk for self-driving cars, though? LordHorst (talk) 12:59, 21 February 2018 (UTC)
My best guess is 12. 162.158.93.9 14:50, 21 February 2018 (UTC)
Huh, I never even realized this interpretation. :) Hawthorn (talk) 15:09, 21 February 2018 (UTC)
But you wouldn't use a flash drvie - it would be an OTA update.162.158.155.26 11:53, 22 February 2018 (UTC)
Funny, but now, as the explanation now states it is the cars that need to be updated to take this new behavior in to context. It is unclear if the cars should behave like humans, as Cueball mentions they already do, and if so should try to use the knowledge of human behavior to save life, or if they should behave like humans and try to take lives! :-D --Kynde (talk) 13:14, 21 February 2018 (UTC)
While this version occurred to me, I feel confident that the INTENDED meaning is to go within the context of this comic, that is, that the murderers are the non-drivers Cueball is afraid of, and both human and AI drivers alike still would prefer not to crash (especially seeing as the most available murder victims would be the people in the car, both passengers and possible driver). Still, the car-owners-want-to-kill version is amusing. LOL! NiceGuy1 (talk) 05:33, 23 February 2018 (UTC)
To me it's obvious, and the explanation overthinks it. The natural "voice" of the comic here is Cueball. He surmises that if people generally become murderers, they would need a firmware update "or something". Obviously he is taking "people" to be some sort of product, and the update to be the manufacturer's obligation. There is obvious religious precedent for this, whether the "firmware update" involves unleashing floods, burning cities with fire, sending messaiahs or revealing truths to prophets... and that is (imho) the joke: Cueball thinks an update is needed, then he generalises that to "or something" as he starts to realise (or as the reader starts to realise, despite Cueball's obliviousness) the enormity of the consequences of such a possible update. 108.162.229.166 03:42, 3 March 2018 (UTC)
Agree with the interpretation, but I think floods, fire, etc. would be more 'retiring hardware' than a 'firmware update' 162.158.155.26 09:28, 5 March 2018 (UTC)

I would say that it's easier to fool current AI than human ... except in "quick reaction needed" scenario. If you throw cardboard cutout of a pedestrian on road, AI will be fooled because it's not able to recognize it's not human and crash. Human will crash as well, because while he will eventually realize it's cutout, it would be too late. -- Hkmaly (talk) 23:35, 21 February 2018 (UTC)

A creative attacker could put up a sign with CAPTCHA-like text that would be readable by humans but not by an AI.

Except that both humans and AIs would disregard it as not being a real sign. In fact, this would be more likely to be successful as an attack against humans, who might at least be distracted by going "what's that?", and end up crashing as a result. The AI would just completely ignore it.162.158.155.26 11:56, 22 February 2018 (UTC)

One of the most ingrained features in humans is to always drive on the correct side of the road

Seems like rather a big claim there... 162.158.155.26 15:03, 22 February 2018 (UTC)

It still says "muderers"[sic] ... maybe it isn't just a misspelling? 162.158.126.106 22:59, 11 March 2018 (UTC)

One of my pet concerns regarding AI cars has been about those "repainted" lines often seen especially around construction areas. With a bit of light shining in the wrong direction, or some rain, as a human driver you need to be VERY aware about what is going on, and be able to disregard the appearance of those "ghost" lines being real. I am sure that any AI that has some quality does likewise, i.e., take multiple data points regarding what is really going on, not just the lines themselves. However, and to the point of these comments, I would suspect that, such lines being "so real", and therefore, an authentic risk, they probably would result in human drivers crashing. Does that happen? Are accident data searchable for this kind of things? Note that, if that were the case, there would be liability consequences, and the construction company involved should, probably, have to pay a bundle, following up that kind of chain of responsibility would be rather simple [citation needed], and, after this happens a few times, road construction companies would get the message, and be very careful about said ghost lines. But. Ghost lines still exists. Thus, either they pose no Real Life hazard, or showing liability is not so plain, or...
Some of our colleague commentards here seem to imply that AI would not be worse than humans, and that somehow humans are generally safe, and thus all is essentially OK?
With all due respect, humans are not safe drivers, not in any sense that would put a sane person's mind at ease regarding going out of the house, as accurately explained on [XKCD/1990]. What is going on, as has been pretty well established (see recent events, June 2020), is that we appear to have become inured to car accidents (inured: don't care). Some people have been pointing at this when trying to demonstrate that the covid-19 crisis is of no real import. Apparently (I personally don't know and only marginally care about any mathematical difference, in my mind both are terrible) the deaths from car accidents are significantly higher than those, at this date, by covid-19. Their argument being that we should not care so much about covid-19, because we don't (seem to) care about car accident deaths. My argument being, supported by the title text in XKCD/1990, and [this link in the comments there], that AI driven cars will win, not because they are that much safer, or the same as humans, but because courts and the general public will happily accept the resulting deaths as part of the way things are, as they do with human-caused road mayhem. Whatever. Anyway, as to the State of the Art in 2018, AI cars still get fiercely confused by rain and snow [see pictures and comment here], which would explain why so much testing has moved to Arizona... Yamaplos (talk) 13:55, 14 June 2020 (UTC)