The diffusion of machine learning technologies and their applications is increasing nowadays, becoming more and more important in everyday life. One area of interest in this field is adversarial machine learning, where models are exposed to perturbed input data, known as "adversarial examples", to induce classification errors. This causes a serious threat to the reliability and security of models, especially when used in real-world applications. This thesis will investigate the generation of adversarial examples to attack deep neural networks used for image classification. The aim will be to minimise a black-box function using algorithms developed for structured optimisation. The formulation of the underlying mathematical model allows us to solve large-scale instances of the problem and find sparse solutions, generating adversarial examples that are very similar to the original images.

The diffusion of machine learning technologies and their applications is increasing nowadays, becoming more and more important in everyday life. One area of interest in this field is adversarial machine learning, where models are exposed to perturbed input data, known as "adversarial examples", to induce classification errors. This causes a serious threat to the reliability and security of models, especially when used in real-world applications. This thesis will investigate the generation of adversarial examples to attack deep neural networks used for image classification. The aim will be to minimise a black-box function using algorithms developed for structured optimisation. The formulation of the underlying mathematical model allows us to solve large-scale instances of the problem and find sparse solutions, generating adversarial examples that are very similar to the original images.

A Zeroth-Order Method for Adversarial Attacks

CORRADO, LORENZO
2022/2023

Abstract

The diffusion of machine learning technologies and their applications is increasing nowadays, becoming more and more important in everyday life. One area of interest in this field is adversarial machine learning, where models are exposed to perturbed input data, known as "adversarial examples", to induce classification errors. This causes a serious threat to the reliability and security of models, especially when used in real-world applications. This thesis will investigate the generation of adversarial examples to attack deep neural networks used for image classification. The aim will be to minimise a black-box function using algorithms developed for structured optimisation. The formulation of the underlying mathematical model allows us to solve large-scale instances of the problem and find sparse solutions, generating adversarial examples that are very similar to the original images.
2022
A Zeroth-Order Method for Adversarial Attacks
The diffusion of machine learning technologies and their applications is increasing nowadays, becoming more and more important in everyday life. One area of interest in this field is adversarial machine learning, where models are exposed to perturbed input data, known as "adversarial examples", to induce classification errors. This causes a serious threat to the reliability and security of models, especially when used in real-world applications. This thesis will investigate the generation of adversarial examples to attack deep neural networks used for image classification. The aim will be to minimise a black-box function using algorithms developed for structured optimisation. The formulation of the underlying mathematical model allows us to solve large-scale instances of the problem and find sparse solutions, generating adversarial examples that are very similar to the original images.
Optimization
Adversarial Attacks
Deep Learning
Zeroth-order Method
Security
File in questo prodotto:
File Dimensione Formato  
A_Zeroth_Order_Method_for_Adversarial_Attacks_Lorenzo_Corrado.pdf

accesso riservato

Dimensione 3.69 MB
Formato Adobe PDF
3.69 MB Adobe PDF

The text of this website © Università degli studi di Padova. Full Text are published under a non-exclusive license. Metadata are under a CC0 License

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.12608/46199