Kyle Wiggers reports in Venture Beat:
It’s been hard to capture commonsense knowledge in a form that algorithms can make useful. It turns out that language models that read text and try to predict the next word and make sense of the future to autocomplete sentences capture commonsense knowledge. When explanations consisted only of justifications, the best accuracy the model could reach was 53%, in contrast to the 85% hit by models trained on open-ended explanations. Adding questions boosted performance to 70%, and to 90% when provided at inference time. “The idea behind explainable AI is to have a model generate explanations for their decisions so users can interact with them and understand them.”
Sufficiently sophisticated AI models are capable of performing incredible feats, from predicting which patients are likely to develop breast cancer and spotting early signs of glaucoma from eye scans to hallucinating fake landscapes that look indistinguishable from the real thing. But despite their versatility, there’s a shortcoming they universally share: a lack of commonsense reasoning. Try telling a machine learning algorithm to predict what’ll happen when you push a ball off a table, or when a person trips down the stairs. Unless it’s explicitly “taught” laws of physics through training on countless examples, it’ll struggle.
One solution is enumerating logic and applying it to a given AI model’s decision-making, but that’s a time-consuming and monotonous chore that doesn’t account for the many exceptions to probabilistic heuristics. That’s why scientists at Salesforce investigated an alternative approach, which they detail in a paper accepted into the 2019 Annual Meeting of the Association for Computational Linguistics: training a system on sequences of explanations for commonsense reasoning and highlighted annotations. They propose a new open source corpus — Common Sense Explanations (CoS-E) — for training and inference with a novel machine learning framework (Commonsense Auto-Generated Explanation, or CAGE), which they say improves performance on question-and-answer benchmarks by 10% over baselines and demonstrates an aptitude for reasoning in out-of-domain tasks.
“It turns out that, despite all the recent breakthroughs over the last decade, it’s been historically really hard to capture commonsense knowledge in a form that algorithms can actually make useful,” Salesforce chief scientist and coauthor on the paper Richard Socher told VentureBeat in a phone interview. “The reason I’m so excited for [the paper] here is that they have a first approach to capture commonsense knowledge, and it turns out that language models — simple models that read text and try to predict the next word and make sense of the future to autocomplete sentences — capture this commonsense knowledge.”
Compiling a data set
Devising the model was a multiset process.
To procure commonsense explanations for CoS-E, which is divided into two parts — a question token split and a random split — the team turned to Amazon’s Mechanical Turk and tasked human participants with explaining which of several answers was “most appropriate” given ground-truth answers. Annotators highlighted relevant words in questions that justified the ground truths, and then provided brief, open-ended explanations based on the highlighted justifications that served as the reasoning behind the questions.
For example, for the prompt “What could people do that involves talking?” the crowdworkers had to select from these answers: “confession,” “carnival,” or “state park.” Their explanation for “confession” might be “confession is the only vocal action,” and they might supply the reason “people talk to each other” or the rationale “people talk to people.”
Socher notes that CoS-E’s effectiveness isn’t constrained by the examples. CAGE achieves state-of-the-art results when trained on it, implying that even when drawing only on explanations that don’t have any word overlap with any of the answer choices, performance exceeds that of models which don’t use CoS-E.
“Usually, a lot of the tasks and data sets we look at have all the information [an AI model] needs to make a certain call,” explained Socher. “But [the model will] never be able to enumerate all the different possible types of reasoning to be able to do well on the test set, because the test set includes completely empty domains and things [the model has] never seen before.”
Devising a model
So how did CAGE come about? Well, Rajini and team drew examples from Common sense Question Answering (CQA), a corpus containing over multiple choice questions for developing common sense reasoning models. They paired them with corresponding CoS-E explanations from a natural language model conditioned on the question-and-answer choices. Next, they concatenated the explanations to the end of the original questions, answer choices, and outputs, and lastly fed them to a second commonsense reasoning model.
In this way, the team considerably extended the capabilities of CQA, which was designed to benchmark performance on tasks requiring proficiency in pronoun resolution. Whereas results from CQA tend to be somewhat ambiguous with respect to whether commonsense reasoning is actually being performed, the researchers assert that CoS-E’s explanations are explicit and can be used to study, analyze, and evaluate models’ reasoning capabilities.
The aforementioned language model was OpenAI’s GPT, a multilayer transformer decoder and the forebear of the highly capable GPT-2 model released last year. As with all deep neural networks, GPT contains neurons (mathematical functions loosely modeled after biological neurons) arranged in interconnected layers that transmit “signals” from input data and slowly adjust the synaptic strength — weights — of each connection. (That’s how the model extracts features and learns to make predictions.) Uniquely, however, it has attention: Every output element is connected to every input element, and the weightings between them are calculated dynamically.
For the commonsense reasoning model — a classification module that learned to perform predictions on the CQA task — the team chose Google’s BERT, which is unique in that it’s both bidirectional (allowing it to access context from both past and future directions) and unsupervised (meaning it can ingest data that’s neither classified nor labeled).
The team fine-tuned a pretrained GTP model on a combination of CQA and CoS-E data sets and experimented with language generation in two settings: “reasoning,” where the language model conditioned on questions, answer choices, and the human-generated explanation but not the actual predicted label, and “rationalization,” where the model conditioned on the predicted labels along with the input to generate rationalizations. The researchers found that reasoning outperformed the state-of-the-art on CQA by 10%, while rationalization bested the current top-ranking model by 6%.
The explanations in the rationalization setup can’t be considered commonsense reasoning, Rajani and colleagues note, because the model had access to the ground truth labels to input questions during training. Instead, they consider it an interpretability framework — a means of making the system’s decisions more transparent.
“The idea behind explainable AI is that you’d like to have an AI model to generate explanations for their decisions, and the most obvious reason for this is to gain users’ trust so that users can interact with them and they understand them,” Rajini told VentureBeat.
Surprising results
With models and data set in hand, the team moved onto the next experimental step: validation.
On CQA, they say that CAGE achieved accuracy of roughly 65%, which they claim is state-of-the-art. And during a test in which the commonsense question answering model was provided access to explanations that weren’t conditioned on the ground truth (during both training and validation), accuracy jumped nearly 10% from 64% to 72%.
Interestingly, the team found that when explanations consisted only of justifications, the best accuracy the model could reach was 53%, in contrast to the 85% hit by models trained on open-ended explanations. Adding questions to the mix boosted performance to 70%, and to 90% when provided at inference time.
The team separately carried out a test on two out-of-domain data sets: SWAG, a corpus with multiple choice questions about “a rich spectrum of grounded situations,” and Story Cloze, a collection of five-sentence “commonsense” stories. Model performance was slightly worse across the board, but the outputs exhibited surprisingly little in the way of grammatical or syntactical errors and contained information relevant to the scenarios at hand. In the case of the SWAG data set, where each question was a video caption with choices about what might happen next, generated explanations seemed to be grounded in given images even though the language model wasn’t trained on SWAG.
“It shows that it’s worthwhile for the [research] community to think about collecting explanations as they’re collecting new data sets,” said paper coauthor Bryan McCann. “[It turns out that] actually going to the trouble to having humans write a little sentence about why they [chose an answer to a question] will potentially be very useful … for accessibility, interpretability, and performance, as well.”
Work has already begun on CAGE frameworks with larger language models, which Socker predicts will boost accuracy even further.
“You can plug in any language model that’s pretrained and has weights available. Our hypothesis is that as you get larger and larger language models, you’ll capture more and more common sense,” he said. “Before, knowledge conglomeration used to be thought of as a human-in-the-loop endeavor … and the nice thing here is, we can allow this model to read text [and then] make sense from all the things that people are saying. It can read about the world … and really capture this common-sense reasoning ability.”
Rajani believes the work could lay the groundwork for more helpful, less frustrating AI assistants.
“For example, suppose that you’re interacting with a robot and you have a coffee mug and an empty glass in front of you, and you say ‘Pour me some water in a glass.’ If the robot had common sense, you wouldn’t have to be very specific — it’s not going to pour water in the coffee mug.”
23 comments:
Many people don't even know right now how to optimize a business. You can do this by unique organizational solutions. I did and I didn't regret it myself. I didn't. That is why I advise you to look and read the details about the same salesforce crm pricing , find out here details about the price, the cost and then decide for yourself whether you plan to use similar software. But I use myself, as productivity is substantially rewarded by business optimization.
Artificial intelligence makes the sales process smarter at every stage. Salesforce artificial intelligence technology automates data entry and predictive analytics to ensure every moment of sales is accounted for. It is important. It is also important to establish security in teams, together with https://alpacked.io/consulting/security-consulting this is easier to do.
Algoworks is a global outsourced IT Services company and our business is to make the offshore strategy work for ISVs and software-enabled companies.https://www.algoworks.com/salesforce/
https://www.algoworks.com/salesforce/
Thanks for sharing this valuable information
Algoworks is a technology company providing end-to-end mobile design and development services globally. From award-winning B2C native applications to robust cross-platform enterprise-grade mobile solutions, Algoworks innovates applications with the latest tech trends. Our customers include Silicon Valley start-ups to Fortune 500 companies.
https://www.algoworks.com/mobile-app-development-services/
Innovadel is a leading multi-channel e-commerce service provider company that helps businesses increase revenue, and conversion rate, and improve their business process. We are Salesforce Consulting Partner and expert in e-commerce development.
Since 2010 our team has won a lot of awards and has developed 50+ e-commerce websites for diverse companies. Our hardworking and passionate developers, designers, and support team deliver quality products. We follow a supple methodology so we can offer you our services with error-free results. Our experts take care of everything from designing, development, and implementation, to support.
Experience the ultimate in modern outdoor living with our Aluminium Lounge outdoor Sofa Set, designed to seamlessly transition between indoor and outdoor spaces. Crafted for weatherproof durability, this set features sleek black frames that beautifully contrast with the grey cushions, exuding contemporary elegance. The inclusion of a dining table adds versatility, making it perfect for both relaxing and entertaining. Enjoy comfort without compromise with the weatherproof cushions that accompany this set. Elevate your living spaces with the perfect combination of style and functionality through our Aluminium Lounge Sofa Set, available now.
Discover a whole new look for your smile with our Teeth Whitening Procedure!
At First Point Dental, we understand the importance of a bright and confident smile. Our teeth whitening procedures are designed to restore your smile's natural radiance and boost your confidence.
Factors such as the consumption of coloring substances and certain habits can cause tooth discoloration over time. Our professional teeth whitening program is your solution. Unlike common options, we tailor treatment plans to your specific needs to ensure the best possible outcome.
Our experienced dental team uses advanced technology to safely and effectively remove stains and restore a great smile. Our teeth whitening treatments will restore your teeth to a youthful glow and allow you to face the world with renewed confidence.
Experience the life-changing results of teeth whitening surgery at First Point Dental. Show me a brighter smile today!
Discover the impactful legal services offered by Refugees International. As a dedicated advocate for displaced individuals globally, we provide essential legal support to ensure their rights and well-being. Through our comprehensive legal services, we empower refugees by addressing their unique legal needs and challenges, striving to create a more just and secure future for them and their communities.
Can I Change My Flight on Expedia for Free
To find the cheapest day to book a JetBlue flight, aim for Tuesdays and Wednesdays. These days often offer lower fares compared to others. Booking during these mid-week days can help you secure better deals for your travels. What is the Cheapest Day to Book JetBlue Flight
To speak with American customer service, start by calling the company's customer service number. Follow the automated menu and choose the option that best fits your issue. Often, pressing "0" or saying "representative" will connect you to a live person. Be prepared with your account details and relevant information before making the call. How to Speak to American Customer Service
Can I correct my name on an American Airlines ticket
: American Airlines allows passengers to correct minor name errors on their tickets. This includes typos or small changes like correcting a misspelled name. Passengers should contact the airline's customer service to request the correction, and in some cases, they may need to provide documentation to verify the change.
What is the cancellation policy for American Airlines : American Airlines has a cancellation policy that varies based on the type of ticket purchased. Generally, non-refundable tickets can be canceled with a fee, while refundable tickets can be canceled without penalty. Additionally, changes or cancellations made within 24 hours of booking may be eligible for a full refund.
If you need to change the passenger name on a Delta flight, follow the guidelines outlined in our comprehensive guide. Additionally, if you're looking to speak with a Delta representative quickly, explore our tips to streamline the process and get the assistance you need promptly.
How to Change Passenger Name on Delta
Delta Representative Fast
If you're wondering whether you can cancel your American Airlines flight and receive a refund, the answer depends on various factors, including the type of ticket you purchased and the timing of your cancellation. American Airlines offers a 24-hour grace period for cancellations, allowing passengers to cancel their tickets within 24 hours of booking for a full refund, provided the booking was made at least two days before departure. Beyond this window, refundable tickets can be canceled for a full refund, while non-refundable tickets typically only allow for future travel credit, except in cases of significant schedule changes, death, or military orders. For detailed guidance and steps on how to navigate cancellations and refunds with American Airlines.
Can I cancel my American Airlines flight and get a refund
If you're looking to change your flight on Expedia, you might be wondering if you can do so for free. The ability to change your flight without incurring a fee depends on the airline's policies and the fare type you purchased. Many airlines have relaxed their change fee policies, especially for flexible fares or in response to the COVID-19 pandemic. However, changes to basic economy tickets usually still involve fees. It's crucial to review the specific terms of your ticket and the airline's policy. For comprehensive guidance and step-by-step instructions on changing your flight on Expedia.
Can I change my flight on Expedia for free
how do i speak to a person at expedia
To quickly talk with a live representative at Volaris Airlines, call their customer service phone number at + 1-844-987-7026.
How Can I Speak Directly to a Volaris Representative Fast?
Post a Comment