Learning Variable Ordering Heuristics with Multi-Armed Bandits and Restarts

被引:6
|
作者
Wattez, Hugues [1 ,2 ]
Koriche, Frederic [1 ,2 ]
Lecoutre, Christophe [1 ,2 ]
Paparrizou, Anastasia [1 ,2 ]
Tabary, Sebastien [1 ,2 ]
机构
[1] Univ Artois, CRIL, Arras, France
[2] CNRS, Paris, France
关键词
SOLVER; SEARCH;
D O I
10.3233/FAIA200115
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In constraint-based applications, the user is often required to be an expert as, for a given problem instance, many parameters of the used solver must be manually tuned to improve its efficiency. Clearly, this background knowledge burdens the spread of constraint programming technology to non-expert users. In order to alleviate this issue, the idea of "autonomous" constraint solving is to adjust the solver parameters and to efficiently handle any problem instance without manual tuning. Notably, the choice of the variable ordering heuristic can lead to drastically different performances. A key question arises then: how can we find the best variable ordering heuristic for a problem instance, given a set of available heuristics provided by the solver? To answer this question, we propose an algorithmic framework that combines multi-armed bandits and restarts. Each candidate heuristic is viewed as an arm, and the framework learns to estimate the best heuristic using a multi-armed bandit algorithm. The common mechanism of restarts is used to provide feedback for reinforcing the bandit algorithm. Based on a thorough experimental evaluation, we demonstrate that this framework is able to find the best heuristic for most problem instances; notably, it outperforms the state-of-the-art in terms of time and solved instances.
引用
收藏
页码:371 / 378
页数:8
相关论文
共 50 条
  • [41] Falcon: Fair Active Learning using Multi-armed Bandits
    Tae, Ki Hyun
    Zhang, Hantian
    Park, Jaeyoung
    Rong, Kexin
    Whang, Steven Euijong
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2024, 17 (05): : 952 - 965
  • [42] LEVY BANDITS: MULTI-ARMED BANDITS DRIVEN BY LEVY PROCESSES
    Kaspi, Haya
    Mandelbaum, Avi
    ANNALS OF APPLIED PROBABILITY, 1995, 5 (02): : 541 - 565
  • [43] Multi-player Multi-armed Bandits: Decentralized Learning with IID Rewards
    Kalathil, Dileep
    Nayyar, Naumaan
    Jain, Rahul
    2012 50TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2012, : 853 - 860
  • [44] Aggregation of Multi-Armed Bandits Learning Algorithms for Opportunistic Spectrum Access
    Besson, Lilian
    Kaufmann, Emilie
    Moy, Christophe
    2018 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2018,
  • [45] Learning by Repetition: Stochastic Multi-armed Bandits under Priming Effect
    Agrawal, Priyank
    Tulabandhula, Theja
    CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI 2020), 2020, 124 : 470 - 479
  • [46] Constructing Test Collections using Multi-armed Bandits and Active Learning
    Rahman, Md Mustafizur
    Kutlu, Mucahid
    Lease, Matthew
    WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, : 3158 - 3164
  • [47] Successive Reduction of Arms in Multi-Armed Bandits
    Gupta, Neha
    Granmo, Ole-Christoffer
    Agrawala, Ashok
    RESEARCH AND DEVELOPMENT IN INTELLIGENT SYSTEMS XXVIII: INCORPORATING APPLICATIONS AND INNOVATIONS IN INTELLIGENT SYSTEMS XIX, 2011, : 181 - +
  • [48] Quantum greedy algorithms for multi-armed bandits
    Hiroshi Ohno
    Quantum Information Processing, 22
  • [49] Online Multi-Armed Bandits with Adaptive Inference
    Dimakopoulou, Maria
    Ren, Zhimei
    Zhou, Zhengyuan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [50] Multi-Armed Bandits for Adaptive Constraint Propagation
    Balafrej, Amine
    Bessiere, Christian
    Paparrizou, Anastasia
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 290 - 296