Breaking Machine Learning Models with Adversarial Attacks and its Variants
DOI:
https://doi.org/10.32473/flairs.38.1.139042Abstract
Machine learning models can be by adversarial attacks, subtle, imperceptible perturbations to inputs that cause the model to produce erroneous outputs. This tutorial introduces adversarial examples and its variants, explaining why even stateof-the-art models are vulnerable and how this impacts security in AI. It provides an overview of key concepts (such as black-box vs. white-box attack scenarios) and survey common attack techniques and defensive strategies. A hands-on component using Google Colab and the open-source Adversarial Lab toolkit allows attendees to craft adversarial examples and test model robustness in real time. Throughout, we emphasize both the practical skills and the ethical considerations needed to apply adversarial machine learning in a responsible
manner. Attendees will gain a comprehensive foundation
in adversarial attacks and insights into building more
robust, secure machine learning models.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 Pavan Reddy

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.