Understanding AI progress and its current capabilities
I often wonder why people do not spend more time to understand what AI can do. The best way I think, is to look at the tasks that is "solved", that works. This is both inspiring and greatly reduces the risks. These are problem "I" could solve given a competent enough team and budget. The best way i think is not to start looking at the basics or the math which is my main guess as for why people know so little.
So where do we start? Lets have a look at the tasks currently used to measure AI-models capability for language understanding. The Super General Language Understanding Evaluation Benchmark(SuperGLUE)
A few years ago a set of tasks was collected to evaluate progress on AI research and language understanding. It was called GLUE. Last year AI surpassed the human baselines.
The researchers however had a sense that these tasks would be beaten and gathered a set of new tasks, called SuperGLUE. It involves QA and other more difficult tasks. But its naming might be a bit misleading. It is not really Super general language understanding from a human perspective, but compared to where the field of AI was before the introduction of "true" deep learning models like Transformers and its continuations such as BERT and Electra it seems general.
Just before the start of 2021, SuperGlue was been beaten by a team from Microsoft. 12 Hours after Microsofts submission it was beaten again by a Google researcher. We need even more difficult tasks now. But how can you benefit from this progress?
Lets have a look through the SuperGLUE tasks with examples and how AIs results compares with humans.
This is a great task for question answering. Is the answer to the question yes or no given a source? Similar systems to these will be everywhere in a few years time. We are currently experimenting with various ways on how to make this incredibly easy for you to implement.
do iran and afghanistan speak the same language
Persian language -- Persian (/ˈpɜːrʒən, -ʃən/), also known by its endonym Farsi (فارسی fārsi(fɒːɾˈsiː) ( listen)), is one of the Western Iranian languages within the Indo-Iranian branch of the Indo-European language family. It is primarily spoken in Iran, Afghanistan (officially known as Dari since 1958), and Tajikistan(officially known as Tajiki since the Soviet era),and some other regions which historically were Persianate societies and considered part ofGreater Iran. It is written in the Persian alphabet, a modified variant of the Arabic script, which itself evolved from the Aramaic alphabet.
BoolQ (Boolean Questions, Clark et al., 2019a) is a QA task where each example consists of a short passage and a yes/no question about the passage. The questions are provided anonymously and unsolicited by users of the Google search engine, and afterwards paired with a paragraph from a Wikipedia article containing the answer. Following the original work, we evaluate with accuracy.
Human score: 89.0
AI score: 91.2
This is very similar to the tasks of GLUE. The main use for these are comparing sentences/texts and see similarities or entailments/contradictions. The progress made in this area will benefit you directly through Labelf where we make extensive use of these models for text classification.
It was a complex language. Not written down but handed down. One might say it was peeled down.
the language was peeled down
Entailment (Entailment, Neutral or Contradiction)
The CommitmentBank (De Marneffe et al., 2019) is a corpus of short texts in which at least one sentence contains an embedded clause. Each of these embedded clauses is annotated with the degree to which we expect that the person who wrote the text is committed to the truth of the clause. The resulting task framed as three-class textual entailment on examples that are drawn from the Wall Street Journal, fiction from the British National Corpus, and Switchboard. Each example consists of a premise containing an embedded clause and the corresponding hypothesis is the extraction of that clause. We use a subset of the data that had inter-annotator agreement above 0.85. The data is imbalanced (relatively fewer neutral examples), so we evaluate using accuracy and F1, where for multi-class F1 we compute the unweighted average of the F1 per class.
Train size: 250
Human score: 95.8/98.9
AI score: 95.8/97.6
3. The Choice Of Plausible Alternatives (COPA)
Does the model contain some sort of information of what might be the most plausible connection/reason. Does it "understand" anything about the world through questions about cause and effect?
My body cast a shadow over the grass.
The sun was rising
The grass was cut
Option 1 (The sun was rising)
The Choice Of Plausible Alternatives (COPA, Roemmele et al., 2011) dataset is a causal reasoning task in which a system is given a premise sentence and two possible alternatives. The system must choose the alternative which has the more plausible causal relationship with the premise. The method used for the construction of the alternatives ensures that the task requires causal reasoning to solve. Examples either deal with alternative possible causes or alternative possible effects of the premise sentence, accompanied by a simple question disambiguating between the two instance types for the model. All examples are handcrafted and focus on topics from online blogs and a photography-related encyclopedia. Following the recommendation of the authors, we evaluate using accuracy.
Train size: 400
Human score: 100
AI score: 98.4
4. The Choice Of Plausible Alternatives (COPA)
This task is similar to BoolQ but actually contains the answer to the question and can be viewed as a fact checker given a premise, a question and an answer.
While this process moved along, diplomacy continued its rounds. Direct pressure on theTaliban had proved unsuccessful. As one NSC staff note put it, "Under the Taliban, Afghanistan is not so much a state sponsor of terrorism as it is a state sponsored by terrorists." In early 2000, the United States began a high-level effort to persuade Pakistan to use its influence over the Taliban. In January 2000, Assistant Secretary of State Karl Inderfurth and the State Department's counterterrorism coordinator, Michael Sheehan, met with General Musharraf in Islamabad, dangling before him the possibility of a presidential visit in March as a reward for Pakistani cooperation. Such a visit was coveted by Musharraf, partly as a sign of his government's legitimacy. He told the two envoys that he would meet with Mullah Omar and press him on Bin Laden. They left, however, reporting to Washington that Pakistan was unlikely in fact to do anything, " given what it sees as the benefits of Taliban control of Afghanistan. "President Clinton was scheduled to travel to India. The State Department felt that he should not visit India without also visiting Pakistan. The Secret Service and the CIA, however, warned in the strongest terms that visiting Pakistan would risk the President's life. Counterterrorism officials also argued that Pakistan had not done enough to merit a presidential visit. But President Clinton insisted on including Pakistan in the itinerary for his trip to South Asia. His one-day stopover on March 25, 2000, was the first time a U.S. president had been there since 1969. At his meeting with Musharraf and others,President Clinton concentrated on tensions between Pakistan and India and the dangers of nuclear proliferation, but also discussed Bin Laden. President Clinton told us that when he pulled Musharraf aside for a brief, one-on-one meeting, he pleaded with the general for help regarding Bin Laden." I offered him the moon when I went to see him, in terms of better relations with the United States, if he'd help us get Bin Laden and deal with another issue or two." The U.S. effort continued.
What did the high-level effort to persuade Pakistan include?
Asking Pakistan to help the USA
The Multi-Sentence Reading Comprehension dataset (MultiRC, Khashabi et al., 2018) is a true/false question-answering task. Each example consists of a context paragraph, a question about that paragraph, and a list of possible answers to that question which must be labeled as true or false. Question-answering (QA) is a popular problem with many datasets. We use MultiRC because of a number of desirable properties: (i) each question can have multiple possible correct answers, so each question-answer pair must be evaluated independent of other pairs, (ii) the questions are designed such that answering each question requires drawing facts from multiple context sentences, and (iii) the question-answer pair format more closely matches the API of other SuperGLUE tasks than span-based extractive QA does. The paragraphs are drawn from seven domains including news, fiction, and historical text.
Train size: 27243
Human score: 81.8/51.9
AI score: 88.2/63.7
5. RECoRD (Reading Comprehension with Commonsense Reasoning Dataset)
Is the model Able to understand what entities in the text are being referred to?
The harrowing stories of women and children locked up for so-called 'moral crimes' in Afghanistan's notorious female prison have been revealed after cameras were allowed inside. Mariam has been in Badam Bagh prison for three months after she shot a man who just raped her at gunpoint and then turned the weapon on herself - but she has yet to been charged. Nuria has eight months left to serve of her sentence for trying to divorce her husband. She gave birth in prison to her son and they share a cell together. Scroll down for video Nuria was jailed for trying to divorce her husband. Her son is one of 62 children living at Badam Bagh prison @highlight Most of the 202 Badam Bagh inmates are jailed for so-called 'moral crimes' @highlight Crimes include leaving their husbands or refusing an arrange marriage @highlight 62 children live there and share cells with their mothers and five others
The baby she gave birth to is her husbands and he has even offered to have the courts set her free if she returns, but @placeholder has refused.
"Afghanistan", "Badam Bagh", "Mariam", "Nuria"
(Reading Comprehension with Commonsense Reasoning Dataset, Zhang et al., 2018) is a multiple-choice QA task. Each example consists of a news article and a Cloze-style question about the article in which one entity is masked out. The system must predict the masked out entity from a given list of possible entities in the provided passage, where the same entity may be expressed using multiple different surface forms, all of which are considered correct. Articles are drawn from CNN and Daily Mail. Following the original work, we evaluate with max (over all mentions) token-level F1 and exact match (EM).
Train size: 100730
Human score: 91.7/91.3
AI score: 94.5/94.1
6. Recognizing Textual Entailment (RTE)
Similar to CB, this task compares texts and see if they support each others claims or contradicts/ says nothing about each other.
No Weapons of Mass Destruction Found in Iraq Yet.
Weapons of Mass Destruction Found in Iraq.
The Recognizing Textual Entailment (RTE) datasets come from a series of annual competitions on textual entailment, the problem of predicting whether a given premise sentence entails a given hypothesis sentence (also known as natural language inference, NLI). RTE was previously included in GLUE, and we use the same data and format as before: We merge data from RTE1 (Dagan et al., 2006), RTE2 (Bar Haim et al., 2006), RTE3 (Giampiccolo et al., 2007), and RTE5 (Bentivogli et al., 2009). All datasets are combined and converted to two-class classification: entailment and not_entailment. Of all the GLUE tasks, RTE was among those that benefited from transfer learning the most, jumping from near random-chance performance (~56%) at the time of GLUE's launch to 85% accuracy (Liu et al., 2019c) at the time of writing. Given the eight point gap with respect to human performance, however, the task is not yet solved by machines, and we expect the remaining gap to be difficult to close.
Train size: 2490
Human score: 93.6
AI score: 93.2