This thesis focuses on learning-based approaches for the Travelling Salesperson Problem (TSP). While these techniques achieve comparable performance to classical solvers on small sizes, they fail to generalize the learned strategies to larger instances. In particular, Graph Neural Network (GNN) architectures with non-autoregressive (NAR) decoding scheme show notably poor performance when applied to graph sizes beyond those seen during training. However, leveraging the efficiency of NAR models through the compositional paradigm has led to some of the top-performing learning-based methods in the literature. Even though such approaches had a lot of success by breaking down the original problem into subproblems, they ignore the potential to combine direct and compositional generalization. We address this by leveraging the direct size generalization improvements in the compositional setting. In particular, our model achieves competitive performance compared to previous methods, having 9 times fewer parameters than the current architecture used in the same framework. Moreover, during our exploration of direct NAR generalization, we also tested SizeShiftReg, a novel regularization method originally designed for the graph classification task. To the best of our knowledge, this is the first application of SizeShiftReg to learning-based methods in the Combinatorial Optimization (CO) domain. We obtain a 7.4% reduction in the average optimality gap on the test dataset, which indicates the potential of SizeShiftReg to improve the direct generalization of learning-based approaches for the TSP. However, further experimentation with significantly more computational resources is required to scale the application of SizeShiftReg to the amount of data used in state-of-the-art models.

This thesis focuses on learning-based approaches for the Travelling Salesperson Problem (TSP). While these techniques achieve comparable performance to classical solvers on small sizes, they fail to generalize the learned strategies to larger instances. In particular, Graph Neural Network (GNN) architectures with non-autoregressive (NAR) decoding scheme show notably poor performance when applied to graph sizes beyond those seen during training. However, leveraging the efficiency of NAR models through the compositional paradigm has led to some of the top-performing learning-based methods in the literature. Even though such approaches had a lot of success by breaking down the original problem into subproblems, they ignore the potential to combine direct and compositional generalization. We address this by leveraging the direct size generalization improvements in the compositional setting. In particular, our model achieves competitive performance compared to previous methods, having 9 times fewer parameters than the current architecture used in the same framework. Moreover, during our exploration of direct NAR generalization, we also tested SizeShiftReg, a novel regularization method originally designed for the graph classification task. To the best of our knowledge, this is the first application of SizeShiftReg to learning-based methods in the Combinatorial Optimization (CO) domain. We obtain a 7.4% reduction in the average optimality gap on the test dataset, which indicates the potential of SizeShiftReg to improve the direct generalization of learning-based approaches for the TSP. However, further experimentation with significantly more computational resources is required to scale the application of SizeShiftReg to the amount of data used in state-of-the-art models.

Size Generalization of Learning-Based Approaches for the Travelling Salesperson Problem

BULAT, NIKOLA
2023/2024

Abstract

This thesis focuses on learning-based approaches for the Travelling Salesperson Problem (TSP). While these techniques achieve comparable performance to classical solvers on small sizes, they fail to generalize the learned strategies to larger instances. In particular, Graph Neural Network (GNN) architectures with non-autoregressive (NAR) decoding scheme show notably poor performance when applied to graph sizes beyond those seen during training. However, leveraging the efficiency of NAR models through the compositional paradigm has led to some of the top-performing learning-based methods in the literature. Even though such approaches had a lot of success by breaking down the original problem into subproblems, they ignore the potential to combine direct and compositional generalization. We address this by leveraging the direct size generalization improvements in the compositional setting. In particular, our model achieves competitive performance compared to previous methods, having 9 times fewer parameters than the current architecture used in the same framework. Moreover, during our exploration of direct NAR generalization, we also tested SizeShiftReg, a novel regularization method originally designed for the graph classification task. To the best of our knowledge, this is the first application of SizeShiftReg to learning-based methods in the Combinatorial Optimization (CO) domain. We obtain a 7.4% reduction in the average optimality gap on the test dataset, which indicates the potential of SizeShiftReg to improve the direct generalization of learning-based approaches for the TSP. However, further experimentation with significantly more computational resources is required to scale the application of SizeShiftReg to the amount of data used in state-of-the-art models.
2023
Size Generalization of Learning-Based Approaches for the Travelling Salesperson Problem
This thesis focuses on learning-based approaches for the Travelling Salesperson Problem (TSP). While these techniques achieve comparable performance to classical solvers on small sizes, they fail to generalize the learned strategies to larger instances. In particular, Graph Neural Network (GNN) architectures with non-autoregressive (NAR) decoding scheme show notably poor performance when applied to graph sizes beyond those seen during training. However, leveraging the efficiency of NAR models through the compositional paradigm has led to some of the top-performing learning-based methods in the literature. Even though such approaches had a lot of success by breaking down the original problem into subproblems, they ignore the potential to combine direct and compositional generalization. We address this by leveraging the direct size generalization improvements in the compositional setting. In particular, our model achieves competitive performance compared to previous methods, having 9 times fewer parameters than the current architecture used in the same framework. Moreover, during our exploration of direct NAR generalization, we also tested SizeShiftReg, a novel regularization method originally designed for the graph classification task. To the best of our knowledge, this is the first application of SizeShiftReg to learning-based methods in the Combinatorial Optimization (CO) domain. We obtain a 7.4% reduction in the average optimality gap on the test dataset, which indicates the potential of SizeShiftReg to improve the direct generalization of learning-based approaches for the TSP. However, further experimentation with significantly more computational resources is required to scale the application of SizeShiftReg to the amount of data used in state-of-the-art models.
Graph Neural Network
Probability Heatmaps
MCTS
Regularization
Zero-Shot Learning
File in questo prodotto:
File Dimensione Formato  
Bulat_Nikola.pdf

accesso aperto

Dimensione 2.45 MB
Formato Adobe PDF
2.45 MB Adobe PDF Visualizza/Apri

The text of this website © Università degli studi di Padova. Full Text are published under a non-exclusive license. Metadata are under a CC0 License

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.12608/70902