In a world where data collected by edge devices is of high interest, gathering it at a central point has become more challenging due to current data protection regulations. Decentralized paradigms, such as federated and gossip learning, have gained popularity as a solution to train models while avoiding to share raw data. In decentralized machine learning, edge devices train a local version of the model on their own private data. Moreover, local models are then iteratively shared and merged to produce an aggregated model. Federated learning has been proposed as an initial solution, which relies on a central server for the aggregation. In gossip learning, on the other hand, the model sharing and aggregation is performed in a peer-to-peer fashion, giving peers more control over which peers to collaborate with. Previous studies tested feasibility and limits of decentralized machine learning training models from scratch, which requires access to a large pool of data. In this thesis, we analyze a different aspect of decentralized training, exploring the case of decentralized fine-tuning of machine learning models with a limited amount of available data. Our research involves extensive experimentation with a number of model architectures and datasets, considering different data distributions and imbalances to simulate a real-life setup, demonstrating its applicability across different scenarios.

In a world where data collected by edge devices is of high interest, gathering it at a central point has become more challenging due to current data protection regulations. Decentralized paradigms, such as federated and gossip learning, have gained popularity as a solution to train models while avoiding to share raw data. In decentralized machine learning, edge devices train a local version of the model on their own private data. Moreover, local models are then iteratively shared and merged to produce an aggregated model. Federated learning has been proposed as an initial solution, which relies on a central server for the aggregation. In gossip learning, on the other hand, the model sharing and aggregation is performed in a peer-to-peer fashion, giving peers more control over which peers to collaborate with. Previous studies tested feasibility and limits of decentralized machine learning training models from scratch, which requires access to a large pool of data. In this thesis, we analyze a different aspect of decentralized training, exploring the case of decentralized fine-tuning of machine learning models with a limited amount of available data. Our research involves extensive experimentation with a number of model architectures and datasets, considering different data distributions and imbalances to simulate a real-life setup, demonstrating its applicability across different scenarios.

An Empirical Study of Decentralized Fine-tuning for Machine Learning Models

NABIL, YASSER
2023/2024

Abstract

In a world where data collected by edge devices is of high interest, gathering it at a central point has become more challenging due to current data protection regulations. Decentralized paradigms, such as federated and gossip learning, have gained popularity as a solution to train models while avoiding to share raw data. In decentralized machine learning, edge devices train a local version of the model on their own private data. Moreover, local models are then iteratively shared and merged to produce an aggregated model. Federated learning has been proposed as an initial solution, which relies on a central server for the aggregation. In gossip learning, on the other hand, the model sharing and aggregation is performed in a peer-to-peer fashion, giving peers more control over which peers to collaborate with. Previous studies tested feasibility and limits of decentralized machine learning training models from scratch, which requires access to a large pool of data. In this thesis, we analyze a different aspect of decentralized training, exploring the case of decentralized fine-tuning of machine learning models with a limited amount of available data. Our research involves extensive experimentation with a number of model architectures and datasets, considering different data distributions and imbalances to simulate a real-life setup, demonstrating its applicability across different scenarios.
2023
An Empirical Study of Decentralized Fine-tuning for Machine Learning Models
In a world where data collected by edge devices is of high interest, gathering it at a central point has become more challenging due to current data protection regulations. Decentralized paradigms, such as federated and gossip learning, have gained popularity as a solution to train models while avoiding to share raw data. In decentralized machine learning, edge devices train a local version of the model on their own private data. Moreover, local models are then iteratively shared and merged to produce an aggregated model. Federated learning has been proposed as an initial solution, which relies on a central server for the aggregation. In gossip learning, on the other hand, the model sharing and aggregation is performed in a peer-to-peer fashion, giving peers more control over which peers to collaborate with. Previous studies tested feasibility and limits of decentralized machine learning training models from scratch, which requires access to a large pool of data. In this thesis, we analyze a different aspect of decentralized training, exploring the case of decentralized fine-tuning of machine learning models with a limited amount of available data. Our research involves extensive experimentation with a number of model architectures and datasets, considering different data distributions and imbalances to simulate a real-life setup, demonstrating its applicability across different scenarios.
Gossip Learning
Federated Learning
Fine-tuning
Decentralized ML
File in questo prodotto:
File Dimensione Formato  
Nabil_Yasser.pdf

accesso aperto

Dimensione 1.62 MB
Formato Adobe PDF
1.62 MB Adobe PDF Visualizza/Apri

The text of this website © Università degli studi di Padova. Full Text are published under a non-exclusive license. Metadata are under a CC0 License

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.12608/68876