Developing Human-like Artificial Intelligence Has Risks

9739410452?profile=RESIZE_400xCreating human-like AI is about more than mimicking human behavior technology must also be able to process information, or ‘think’, like humans too if it is to be fully relied upon.  New research, published in the journal Patterns and led by the University of Glasgow’s School of Psychology and Neuroscience, uses 3D modeling to analyze the way Deep Neural Networks are part of the broader family of machine learning process information, to visualize how their information processing matches that of humans.

It is hoped this new work will pave the way for the creation of more dependable AI technology that will process information like humans and make errors that we can understand and predict.  Let’s just hope that it will not adopt less than positive human traits, such as lying, sleeping on the job, and showing up late for work.

One of the challenges still facing AI development is how to better understand the process of machine thinking, and whether it matches how humans process information, in order to ensure accuracy. Deep Neural Networks are often presented as the current best model of human decision-making behavior, achieving or even exceeding human performance in some tasks. However, even deceptively simple visual discrimination tasks can reveal clear inconsistencies and errors from the AI models, when compared to humans.[1]

Currently, Deep Neural Network technology is used in applications such a face recognition, and while it is very successful in these areas, scientists still do not fully understand how these networks process information, and therefore when errors may occur.  In this new study, the research team addressed this problem by modeling the visual stimulus that the Deep Neural Network was given, transforming it in multiple ways so they could demonstrate a similarity of recognition, via processing similar information between humans and the AI model.

Professor Philippe Schyns, the senior author of the study and Head of the University of Glasgow’s Institute of Neuroscience and Technology, said: “When building AI models that behave “like” humans, for instance, to recognize a person’s face whenever they see it as a human would do, we have to make sure that the AI model uses the same information from the face as another human would do to recognize it. If the AI doesn’t do this, we could have the illusion that the system works just like humans do, but then find it gets things wrong in some new or untested circumstances.”

The researchers used a series of modifiable 3D faces and asked humans to rate the similarity of these randomly generated faces to four familiar identities. They then used this information to test whether the Deep Neural Networks made the same ratings for the same reasons testing not only whether humans and AI made the same decisions, but also whether it was based on the same information. Importantly, with their approach, the researchers can visualize these results as the 3D faces that drive the behavior of humans and networks. For example, a network that correctly classified 2,000 identities was driven by a heavily caricatured face, showing it identified the faces processing very different face information than humans.  Researchers hope this work will pave the way for more dependable AI technology that behaves more like humans and makes fewer unpredictable errors.

Red Sky Alliance is a Cyber Threat Analysis and Intelligence Service organization. For questions, comments or assistance, please contact the office directly at 1-844-492-7225, or feedback@wapacklabs.com

Weekly Cyber Intelligence Briefings:

Weekly Cyber Intelligence Briefings:

REDSHORTS - Weekly Cyber Intelligence Briefings

https://attendee.gotowebinar.com/register/3702558539639477516

 

[1] https://scitechdaily.com/developing-artificial-intelligence-that-thinks-like-humans/

E-mail me when people leave their comments –

You need to be a member of Red Sky Alliance to add comments!