Editing 2451: AI Methodology

Jump to: navigation, search

Warning: You are not logged in. Your IP address will be publicly visible if you make any edits. If you log in or create an account, your edits will be attributed to your username, along with other benefits.

The edit can be undone. Please check the comparison below to verify that this is what you want to do, and then save the changes below to finish undoing the edit.
Latest revision Your text
Line 8: Line 8:
  
 
==Explanation==
 
==Explanation==
 +
{{incomplete|Created by a BOT. TRAINED BY AN ADVERSARIAL AI. Please mention here why this explanation isn't complete. Do NOT delete this tag too soon.}}
  
The joke in this comic is that the people are using {{w|artificial intelligence}} (AI) without understanding how to, and that by doing this the research concerned is at best unreliable and possibly deliberately compromised. The researchers acknowledge that their approach is risky and requires extra verification, but repeatedly use equally or more unreliable AI-based solutions to these problems. Therefore, their problems are likely as bad as they ever were and any other team using one of their verification tools is likely to experience similar unreliability. For an introduction to machine learning, you can visit https://fast.ai/ .
+
The joke in this comic is that the classifier sucks.
  
===Original research===
+
This comic shows Cueball giving a presentation of some description. He is reassuring his audience of the validity of his research's methodology, which he says is "AI-based". There are many issues that can arise from an AI-based methodology, such as lingering influence from its training data or a bad algorithm reducing the quality of the investigation.
The first comment, that "some have questioned our AI-based methodology", refers to difficulty verifying the correctness of AI-based processing. A model (a program which solves a problem with AI-based statistical analysis) may appear reliable when it is instead insufficiently tested. Models are liable to experience issues due to lingering influences from its training data or a bad algorithm reducing the quality of the investigation. It is therefore necessary for research using such models to demonstrate that those models have been tested well enough that their results are likely to be useful. Frequently, additional tests are performed after training to confirm that the model can handle data collected in a different way to the data used to train it.
 
  
===Classifier of methodology quality===
+
Cueball seeks to reassure his audience by quantifying the quality of his methodology. He does this by creating yet another AI to rank methodologies. This would not actually improve the confidence of any audience member, as the AI would still have the same flaws as the methodology AI, due to being created by the same team.
Cueball seeks to reassure his audience by quantifying the quality of his methodology. He does this by creating yet another AI to rank methodologies. This approach is unlikely to instill confidence for a variety of reasons:
 
* The quality AI and original research AI were written by the same team. If the original research AI was ill-designed, the quality AI probably shares design problems with it.
 
* The specific kind of model created is unlikely to be the correct one. Cueball calls this a classifier, which is frequently a type of model which assigns an input into distinct mutually exclusive categories. For example, a classifier might be used to determine what language a chunk of text is, given that the chunk is in only one language. However, quality is a continuous aspect of the data. A classifier of methodologies is likely to sort them into "bad", "mediocre", and "good" categories, whereas an effective model should have the ability to give more precise grades. The choice of a classifier may indicate that Cueball doesn't know which types of models to use.
 
* The training data for this quality AI is not mentioned. If, for example, the team's previous research is used as examples of good methodologies, the AI is likely to judge all methodologies from them as good as well.
 
* A ''methodology section'' refers to quality of writing and is a specific section of a research paper. A good methodology section would accurately and clearly explain what he did, but does not mean the research methodology itself was valid. Cueball doesn't indicate whether he believes his model is trying to analyze the quality of the methodology described, but in any case this is nearly impossible for existing machine learning.
 
* An AI which attempts to judge a methodology section is receiving a great deal of input which is difficult to process. It would have to use {{w|natural language processing}} to understand the writing in the methodology section and would also require a lot of specialized knowledge about the subject matter to judge the quality. This would require artificial general intelligence (AGI), which has not yet been achieved. Since the AI does not have the ability to fully understand complex research, it will likely use unimportant details to judge the methodologies.
 
* The ranking AI heavily favors the methodology of Cueball's AI, and may be biased. It shows a normal distribution, with a singular outlier to the far right with an arrow above. It can be inferred (from the arrow) that this data-point represents the AI's methodology. It is a significant outlier, and as such it is probably not an accurate representation of Cueball's AI. Alternatively, this could be taken as AI 'nepotism', where Cueball's methodology AI is more likely to select AI-based approaches over others. This type of algorithmic bias is mentioned in [[2237: AI Hiring Algorithm]].
 
  
===Spacing AI (from title text)===
+
Furthermore, the AI heavily favours the methodology of Cueball's AI, and may be biased. It shows a normal distribution, with a singular outlier to the far right with an arrow above. It can be inferred this data-point represents the AI's methodology. It is a significant outlier, and as such it is probably not an accurate representation of Cueball's AI.
While there are many red flags in the original AI and quality AI, it is theoretically possible that they operate as Cueball claims. The title text's comments about spacing and diacritics prove that this is not the case and that the quality AI, at least, is completely broken. AI models are given input in various complex ways and determine based on statistical analysis which details are important. Such models can easily find details in the training data which correlate with correct answers but make the resulting model useless.
 
  
For example, a research team once created a model which was given medical information to determine how likely a patient was to have cancer. The model was trained on existing patient records and the team planned to use it on new patients. However, the original model did not use the medical information but instead simply checked the name of the hospital--a patient at a hospital with "cancer center" in the name was likely to have cancer. The model had identified a data point which correlated with the desired answer, but this correlation was not useful for the intended purpose. The model concerned was discarded and a new one created without the hospital name.
+
The title text is likely a continuation of Cueball's dialogue, saying that when the classifying AI was shown good research methodology descriptions, the AI identified weird spacing and diacritics as the indicators of a good methodology. Cueball then used his AI to figure out where to put these into his own methodology description to improve his research report. Adding weird symbols into a text doesn't improve the quality of the text {{Citation needed}} and hence Cueball may be doing something very similar to p-hacking, where data is manipulated to decrease the p-number, which represents the likelihood the data is a fluke. P-hacking is mentioned in [[882: Significant]]
  
In this case, the methodology sections are text written by humans, which can contain various artifacts of the writing process. These can include details like how the user chose to insert spaces, word usage, spelling, or diacritic marks which are optional in English (e.g. naive versus naïve). It appears that the training information identifies certain patterns which correlate with "good" methodologies. This indicates a few more problems for this research team:
+
==Transcript==
* Their AI is using pointless details to decide on the quality of methodology sections, so it is useless.
+
{{incomplete transcript|Do NOT delete this tag too soon.}}
* They haven't recognized that it's useless, so their other AI is probably fatally flawed.
 
* The spacing information is correlated strongly with good methodology, which implies that they probably don't have very many different sources for their training data. Their sample size is too small and the AI, even if it was improved to ignore this information, needs more data to have a chance at being useful.
 
  
==Transcript==
+
:[Cueball stands in front of a projection on a screen and points with a stick to a histogram with a bell curve to the left and one bar to the far right marked with an arrow]
:[Cueball is standing on a podium in front of a projection on a screen and points with a stick to a bar chart histogram with a bell curve to the left and a single bar to the far right marked with an arrow.]
 
 
:Cueball: Despite our great research results, some have questioned our AI-based methodology.
 
:Cueball: Despite our great research results, some have questioned our AI-based methodology.
 
:Cueball: But we trained a classifier on a collection of good and bad methodology sections, and it says ours is fine.
 
:Cueball: But we trained a classifier on a collection of good and bad methodology sections, and it says ours is fine.
  
 
{{comic discussion}}
 
{{comic discussion}}
 
+
[[Category:Artificial Intelligence]]
 
[[Category:Comics featuring Cueball]]
 
[[Category:Comics featuring Cueball]]
[[Category:Artificial Intelligence]]
 
[[Category:Public speaking]]
 
[[Category:Science]]
 

Please note that all contributions to explain xkcd may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see explain xkcd:Copyrights for details). Do not submit copyrighted work without permission!

To protect the wiki against automated edit spam, we kindly ask you to solve the following CAPTCHA:

Cancel | Editing help (opens in new window)