
As AI applications proliferate across industries and sectors, two key security questions arise:
- Are these AI applications cyber-secure?
- Can bad actors exploit them through attacks?
A recent article discusses these questions in relation to AI’s Achilles’ heel: adversarial machine learning (AdvML).
In “Lights Toward Adversarial Machine Learning: The Achilles’ Heel of Artificial Intelligence,” authors Luca Pajola and Mauro Conti take a cybersecurity practitioner’s viewpoint as they discuss the full range of AI application threats
- from the systems and libraries used to deploy an AI application,
- to threats arising in the AI application itself.
Here, we offer a quick overview of Pajola and Conti’s detailed look at AdvML and how it might best serve the needs of AI users today and in the future.
AdvML: Analyzing Adversaries and Entry Points
AI is increasingly deployed in high-risk applications—from “driving” autonomous taxis to directing armed drones to human targets—and having security assurances is far more than a “nice to have” concept. Or rather, it should be.
Enter AdvML, a research field that investigates cyberthreats and malicious actors aiming to manipulate or control AI applications. To do this, AdvML researchers build threat models based on two factors:
- Attacker knowledge: What do they know about the system that contains the AI application?
- Attacker capabilities: What type of operations they might logically engage in?
Based on a detailed examination of these factors and the literature, the authors distinguish two major categories of attack:
- AI-level cyberthreats, which exploit algorithm vulnerabilities in the AI application.
- System-level cyberthreats, which produce AI-level threats by exploiting vulnerabilities in the system that hosts the AI application.
The article explores these two attack categories in detail, including the different families of attacks at each level.
At the AI level, the most popular attack family is the model evasion, in which attackers alter the input with a perturbation that produces a misclassification; simple examples here include
- changing pixel values in a computer vision application, or
- inserting a typo in offensive language to evade detection by commercial tools.
System-level attacks occur by exploiting weaknesses at deeper levels in the AI lifecycle, from hardware to OS to libraries. These attacks can produce threats similar to those at the application level. The article offers two examples:
- OS-level attack. At this level, a backdoor attack can be executed after the AI is deployed by studying and flipping specific bits in the dynamic random-access memory.
- Library-level attack. Common AI libraries—such as Caffe, TensorFlow, and PyTorch—are susceptible to denial-of-service attacks; the consequences of these vary from application crash to model evasion.
AdvML: Changing Directions
As the authors point out, over the past decade, AdvML has focused on understanding potential AI application failures and generating families of adversarial threats. However, these studies are primarily conducted on testbeds that are far removed from our increasingly complex reality.
Moving forward, the authors argue that AdvML needs to shift its focus to filling the gap between research and industry, considering concrete threats to AI applications. Doing so requires deeper consideration of two key questions:
- Who are the AI consumers today?
- What might motivate attacks on these consumers?
Digging Deeper
To read on, see “Lights Toward Adversarial Machine Learning: The Achilles Heel of Artificial Intelligence” in the Sept/Oct issue of the IEEE Intelligent Systems magazine.
To dig even deeper, check out the following resources: