Jul 24, 2023
Řečník · 0 sledujících
Řečník · 0 sledujících
Řečník · 0 sledujících
Řečník · 0 sledujících
Řečník · 1 sledující
Řečník · 1 sledující
Řečník · 3 sledující
Optimization problems with nonlinear cost functions and combinatorial constraints appear in many real-world applications but remain challenging to solve efficiently. Fast optimization techniques like Mixed Integer Linear Programming can be fast in practice but cannot trivially optimize nonlinear cost functions. On the other hand, general nonlinear optimizers like gradient descent often do not handle complex combinatorial structures, may require many queries of the cost function, and are prone to local optima. To bridge this gap, we propose SurCo that learns Surrogate costs which can be used by existing Combinatorial solvers to output good solutions to the original nonlinear combinatorial optimization problem, combining the flexibility of gradient-based methods with the structure of linear combinatorial optimization. We learn these surrogate costs end-to-end with the nonlinear loss by differentiating through the linear surrogate solver.Three variants of SurCo are proposed: SurCo-zero operates on individual nonlinear problems, SurCo-prior trains a linear surrogate predictor on distributions of problems, and SurCo-hybrid uses a model trained offline to warm start online solving for SurCo-zero. We analyze our method theoretically and empirically, showing smooth convergence and improved performance. Experiments show that compared to state-of-the-art approaches and expert-designed heuristics, SurCo finds better solutions with comparable or faster solve time for three applications: embedding table sharding, inverse photonic design, and nonlinear route planning.Optimization problems with nonlinear cost functions and combinatorial constraints appear in many real-world applications but remain challenging to solve efficiently. Fast optimization techniques like Mixed Integer Linear Programming can be fast in practice but cannot trivially optimize nonlinear cost functions. On the other hand, general nonlinear optimizers like gradient descent often do not handle complex combinatorial structures, may require many queries of the cost function, and are prone to…
Professionelle Aufzeichnung und Livestreaming – weltweit.
Präsentationen, deren Thema, Kategorie oder Sprecher:in ähnlich sind
Matt Sampson, …
Lunjia Hu, …
Hamza Keurti, …