2020
DOI: 10.1007/978-3-030-58942-4_12
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement Learning for Variable Selection in a Branch and Bound Algorithm

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 23 publications
(19 citation statements)
references
References 7 publications
0
17
0
Order By: Relevance
“…In this framework, RL methods are used to leverage the power of solvers or problemspecific solution heuristics by initializing values of some hyper-parameters. For example, RL can be utilized to select the branching variable in MIP solvers (Etheve et al 2020;Hottung et al 2020;Tang et al 2020). Some recent studies of Ma et al (2019); Deudon et al (2018); Chen and Tian (2019) show that optimization heuristics powered with RL methods outperform previous methods.…”
Section: Literature Reviewmentioning
confidence: 99%
“…In this framework, RL methods are used to leverage the power of solvers or problemspecific solution heuristics by initializing values of some hyper-parameters. For example, RL can be utilized to select the branching variable in MIP solvers (Etheve et al 2020;Hottung et al 2020;Tang et al 2020). Some recent studies of Ma et al (2019); Deudon et al (2018); Chen and Tian (2019) show that optimization heuristics powered with RL methods outperform previous methods.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Furthermore, SB is by no means an optimal branching policy, therefore methods which offer the potential to go beyond it, such as RL, are particularly appealing. Etheve et al 2020 proposed FMSTS which, to the best of our knowledge, is the only published work to apply RL to branching and is therefore the SOTA RL branching algorithm. By using a DFS node selection strategy, they used the deep Q-network (DQN) approach (Mnih et al, 2013) to approximate the Q-function of the B&B sub-tree size rooted at the current node; a local Q-function which, in their setting, was equivalent to the number of global tree nodes.…”
Section: Related Workmentioning
confidence: 99%
“…Since branching can be formulated as a Markov decision process (MDP) (He et al, 2014;Gasse et al, 2019;Etheve et al, 2020), reinforcement learning (RL) seems a natural approach to discovering novel branching heuristics with superior decision quality and no need for expensive data labelling. However, branching has thus far proved largely intractable for RL for reasons we summarise into three key challenges.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…However, about obtaining a higher dual value, the agent performs worst initially, but it finally obtains the best result, indicating its non-myopic policy. Different from [84], this work can transfer to larger instances, and the performance of the RL agent is significantly superior to FSB, RPB (reliability pseudocost branch), SVM, and GCN.…”
Section: Reinforcement Learning In Branchingmentioning
confidence: 99%