Differentiating adversarial machine learning attacks by predictive and generative AI systems, the report brings standardization to the emerging adversarial machine learning threat landscape.
“AI is useful but vulnerable to adversarial attacks. All models are vulnerable in all stages of their development, deployment, and use,” NIST’s Apostol Vassilev, a research team supervisor and one of the authors of the adversarial machine learning publication, told Nextgov/FCW. “At this stage with the existing technology paradigms, the number and power of attacks are greater than the available mitigation techniques.”
Some of the substantial changes in the final guidelines from the initial version released in January 2024 include an overview of generative AI models’ stages of learning, ongoing open problems in the field and an index on the classes of attacks on different AI systems.
The report lists three distinct threat types for each of the types of AI systems.
For predictive AI systems –– or programs that leverage data to offer forecasts and predictions –– the NIST guidelines review evasion attacks, data poisoning attacks and privacy attacks, all of which change the underlying data powering AI models.
For generative AI models, which are systems composed of algorithms that create unique and new outputs depending on a given input, the three listed attacks are: supply chain, direct prompting and indirect prompt injection.
Direct and indirect prompting attacks use different methods to insert harmful data into the model’s learning pool, potentially corrupting future output.
Support the originator by clicking the read the rest link below.