2022
DOI: 10.48550/arxiv.2205.14345
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Reinforcement Learning for Branch-and-Bound Optimisation using Retrospective Trajectories

Abstract: Combinatorial optimisation problems framed as mixed integer linear programmes (MILPs) are ubiquitous across a range of real-world applications. The canonical branch-and-bound (B&B) algorithm seeks to exactly solve MILPs by constructing a search tree of increasingly constrained sub-problems. In practice, its solving time performance is dependent on heuristics, such as the choice of the next variable to constrain ('branching'). Recently, machine learning (ML) has emerged as a promising paradigm for branching. Ho… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 22 publications
(28 reference statements)
0
2
0
Order By: Relevance
“…Consequently, practitioners rely on either approximate algorithms, which give restricted performance guarantees and poor scalability (Williamson and Shmoys, 2011), or heuristics, which have limited solution efficacy (Halim and Ismail, 2019). Since the first application of neural networks to CO by Hopfield and Tank (1985), the last decade has seen a resurgence in ML-for-CO (Bello* et al, 2017;Dai et al, 2017;Barrett et al, 2019;Gasse et al, 2019;Barrett et al, 2022;Parsonson et al, 2022b). The advantages of ML-for-CO over approximation algorithms and heuristics include handling complex problems at scale, learning either without external input and achieving super-human performance or imitating strong but computationally expensive solvers, and (after training) leveraging the fast inference time of a DNN forward pass to rapidly generate solutions.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Consequently, practitioners rely on either approximate algorithms, which give restricted performance guarantees and poor scalability (Williamson and Shmoys, 2011), or heuristics, which have limited solution efficacy (Halim and Ismail, 2019). Since the first application of neural networks to CO by Hopfield and Tank (1985), the last decade has seen a resurgence in ML-for-CO (Bello* et al, 2017;Dai et al, 2017;Barrett et al, 2019;Gasse et al, 2019;Barrett et al, 2022;Parsonson et al, 2022b). The advantages of ML-for-CO over approximation algorithms and heuristics include handling complex problems at scale, learning either without external input and achieving super-human performance or imitating strong but computationally expensive solvers, and (after training) leveraging the fast inference time of a DNN forward pass to rapidly generate solutions.…”
Section: Related Workmentioning
confidence: 99%
“…To select the algorithm hyperparameters, we conducted a Bayesian search across the search space summarised in Table 5, with simulations conducted in a light 32-worker RAMP environment with a maximum simulation run time of 2 × 10 5 seconds to speed up the search. We adopted similar search ranges to those used by Kurach et al (2019); Hoffman et al (2020); Parsonson et al (2022b). For each set of hyperparameters, we ran the algorithm for 100 learner steps (a.k.a.…”
Section: Reinforcement Learning Algorithmmentioning
confidence: 99%