Breaking Machine Learning Models with Adversarial Attacks and its Variants

Authors

  • Pavan Reddy The George Washington University, Washington DC

DOI:

https://doi.org/10.32473/flairs.38.1.139042

Abstract

Machine learning models can be by adversarial attacks, subtle, imperceptible perturbations to inputs that cause the model to produce erroneous outputs. This tutorial introduces adversarial examples and its variants, explaining why even stateof-the-art models are vulnerable and how this impacts security in AI. It provides an overview of key concepts (such as black-box vs. white-box attack scenarios) and survey common attack techniques and defensive strategies. A hands-on component using Google Colab and the open-source Adversarial Lab toolkit allows attendees to craft adversarial examples and test model robustness in real time. Throughout, we emphasize both the practical skills and the ethical considerations needed to apply adversarial machine learning in a responsible
manner. Attendees will gain a comprehensive foundation
in adversarial attacks and insights into building more
robust, secure machine learning models.

Downloads

Published

14-05-2025

How to Cite

Reddy, P. (2025). Breaking Machine Learning Models with Adversarial Attacks and its Variants. The International FLAIRS Conference Proceedings, 38(1). https://doi.org/10.32473/flairs.38.1.139042