New method for comparing neural networks exposes how artificial intelligence works



The "black box" of artificial intelligence has been examined by a team at Los Alamos National Laboratory to create a novel method for comparing neural networks that will aid researchers in understanding neural network activity. In applications like virtual assistants, facial recognition technology, and self-driving cars, neural networks are used to identify patterns in datasets.

Haydn Jones, a researcher in the Advanced Research in Cyber Systems group at Los Alamos, stated that the artificial intelligence research community "doesn't necessarily have a thorough knowledge of what neural networks are doing; they give us good outcomes, but we don't know how or why." "Our new approach performs a better job of comparing neural networks, which is key toward better understanding the mathematics behind AI," says the author.

The study "If You've Trained One You've Trained Them All: Inter-Architecture Similarity Increases With Robustness," which was recently presented at the Conference on Uncertainty in Artificial Intelligence, was written by Jones as the lead author. The paper is an important step toward characterizing the behavior of robust neural networks in addition to studying network similarity.

High performance, but fragile, neural networks. For instance, autonomous vehicles employ neural networks to recognize signs. They are fairly adept at doing this in perfect circumstances. The slightest deviation, such as a sticker on a stop sign, can, however, cause the neural network to incorrectly detect the sign and never come to a stop.

Researchers are exploring for strategies to increase network resiliency in neural networks. One cutting-edge method involves "attacking" networks as they are being trained. The AI is trained to overlook abnormalities that researchers purposefully introduce. In essence, this procedure, known as adversarial training, makes it more difficult to trick the networks.

In a surprising discovery, Jones and his collaborators from Los Alamos, Jacob Springer and Garrett Kenyon, as well as Jones' mentor Juston Moore, applied their new network similarity metric to adversarially trained neural networks. They discovered that as the severity of the attack increases, adversarial training causes neural networks in the computer vision domain to converge to very similar data representations, regardless of network architecture.

According to Jones, "we discovered that neural networks start acting the same way when we teach them to be robust against adversarial attacks."

The Los Alamos team's findings show that the addition of adversarial training significantly reduces the search area for the "correct design" for neural networks, despite the fact that there has been much effort in both industry and academia to find it. With the knowledge that adversarial training causes various designs to converge to similar answers, the AI research community may not need to invest as much time in investigating different architectures.

"Our discovery that robust neural networks resemble one another makes it simpler to comprehend how robust AI may actually function. We may even be learning more about how humans and other animals perceive things "Jones exclaimed.

DOE/Los Alamos National Laboratory

Comments

Popular posts from this blog

Do You Sleep on Your Back or Side? Here's The Research on 'Optimal' Sleep Positions

The Science of Beards

Scientists Uncover a Surprising Connection Between Appetite and Sun Exposure