Learning to Approximate Uniform Facility Location via Graph Neural Networks

2026-02-13Machine Learning

Machine LearningData Structures and AlgorithmsNeural and Evolutionary Computing
AI summary

The authors explore using message-passing neural networks (MPNNs) to solve a tough optimization problem called Uniform Facility Location (UniFL). Unlike past methods that need lots of training data or complex learning tricks, their model learns without supervision and can handle bigger problems than it was trained on. They combine ideas from classic approximation algorithms, which guarantee good solutions but lack flexibility, with neural networks that can adapt to real data. Their approach gives reliable solution quality and beats traditional approximation methods, getting closer to more exact but slow integer programming solutions.

Uniform Facility LocationMessage-Passing Neural NetworksCombinatorial OptimizationApproximation AlgorithmsSupervised LearningReinforcement LearningInteger Linear ProgrammingDifferentiable ModelsSize Generalization
Authors
Chendi Qian, Christopher Morris, Stefanie Jegelka, Christian Sohler
Abstract
There has been a growing interest in using neural networks, especially message-passing neural networks (MPNNs), to solve hard combinatorial optimization problems heuristically. However, existing learning-based approaches for hard combinatorial optimization tasks often rely on supervised training data, reinforcement learning, or gradient estimators, leading to significant computational overhead, unstable training, or a lack of provable performance guarantees. In contrast, classical approximation algorithms offer such performance guarantees under worst-case inputs but are non-differentiable and unable to adaptively exploit structural regularities in natural input distributions. We address this dichotomy with the fundamental example of Uniform Facility Location (UniFL), a variant of the combinatorial facility location problem with applications in clustering, data summarization, logistics, and supply chain design. We develop a fully differentiable MPNN model that embeds approximation-algorithmic principles while avoiding the need for solver supervision or discrete relaxations. Our approach admits provable approximation and size generalization guarantees to much larger instances than seen during training. Empirically, we show that our approach outperforms standard non-learned approximation algorithms in terms of solution quality, closing the gap with computationally intensive integer linear programming approaches. Overall, this work provides a step toward bridging learning-based methods and approximation algorithms for discrete optimization.