How AI Can Explain Its Thinking
Can we trust artificial intelligence?
New deep learning models for artificial intelligence have proved to be highly effective in detecting cancer and producing fully autonomous cars. Unfortunately, these new algorithms struggle to explain their decision-making processes to their human controllers. Recently, though, a team of international researchers taught AI to support its reasoning with textual and visual evidence, opening a window into the ‘block box’ of AI.
In their paper published on February 15, 2018, the team outlined their work on a potential model for artificial intelligence, which they have aptly named the “Pointing and Justification Explanation,” (PJ-X). Using this model, artificial intelligence gives textual and visual evidence for their decisions. For example, when given an image of a baseball game and asked what sport is depicted, the AI responds that it is baseball and highlights the areas of the image that it thinks are important. In the case of the baseball game, the AI will highlight the player’s bat, the ball, and the catcher’s mitt as justification for its analysis of the image.
The team then combined the visual justifications with textual explanations. When given an image of zebras and asked if the animals are in a zoo, the AI will provide a short fragment of text to explain its decision such as, “No because the zebras are standing in a green field.” Learning how to point to visual evidence helps the AI generate the their respective textual explanations. The effectiveness of their model demonstrates the value of developing a multimodal justification system.
Why a Machine Learning Investment?
Dating Wine Using Nuclear Signatures
On Black Holes: Gateway to Another Dimension, or Ghosts of Stars’ Pasts?
Shedding Light on Dark Matter: Using Machine Learning to Unravel Physics’ Hardest Questions
Machine Learning on Condensed Matter Physics & Quantum Computing
Interview with the Inventor of Amazon's Alexa
Interview with Astronaut Scott Kelly
Aquaponics: How Advanced Technology Grows Vegetables In The Desert
The World Cup Does Not Have a Lasting Positive Impact on Hosting Countries
Lockheed Martin Confirms the SR-72 – Son of Blackbird Will Reach Anywhere in the World in One Hour
The Implications of Machine Learning on Condensed Matter Physics & Quantum Computing
Sources:
Snow, Jackie. (2018, March 8). A new AI system can explain itself—twice. Retrieved March 6, 2019, from https://www.technologyreview.com/the-download/610447/a-new-ai-system-can-explain-itself-twi ce/
Park, D. H., Hendricks, L. A., Akata, Z., Schiele, B., Darrell, T., & Rohrbach, M. (2018, February 15). Multimodal Explanations: Justifying Decisions and Pointing to the Evidence. Retrieved March 6, 2019, from https://arxiv.org/abs/1802.08129
Written by Alex Sheen & Edited by Rachel Weissman & Alexander Fleiss