<sc>Ner4Opt</sc>: named entity recognition for optimization modelling from natural language

被引:0
作者
Kadioglu, Serdar [1 ,2 ]
Dakle, Parag Pravin [1 ]
Uppuluri, Karthik [1 ]
Politi, Regina [2 ]
Raghavan, Preethi [1 ]
Rallabandi, Saikrishna [1 ]
Srinivasamurthy, Ravisutha
机构
[1] Fidel Investments, AI Ctr Excellence, Boston, MA 02109 USA
[2] Brown Univ, Dept Comp Sci, Providence, RI 02912 USA
关键词
Optimization modeling assistants; Named entity recognition; Large language models; Natural language processing; CONSTRAINTS;
D O I
10.1007/s10601-024-09376-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Solving combinatorial optimization problems involves a two-stage process that follows the model-and-run approach. First, a user is responsible for formulating the problem at hand as an optimization model, and then, given the model, a solver is responsible for finding the solution. While optimization technology has enjoyed tremendous theoretical and practical advances, the process has remained unchanged for decades. To date, transforming problem descriptions into optimization models remains a barrier to entry. To alleviate users from the cognitive task of modeling, we study named entity recognition to capture components of optimization models such as the objective, variables, and constraints from free-form natural language text, and coin this problem as Ner4Opt. We show how to solve Ner4Opt using classical techniques based on morphological and grammatical properties and modern methods leveraging pre-trained large language models and fine-tuning transformers architecture with optimization-specific corpora. For best performance, we present their hybridization combined with feature engineering and data augmentation to exploit the language of optimization problems. We improve over the state-of-the-art for annotated linear programming word problems. Large-language models (LLMs) are not yet versatile enough to turn text into optimization models or extract optimization entities. Still, when augmented with Ner4Opt annotations, the compilation accuracy of LLM-generated models improves significantly. We open-source our Ner4Opt library, release our training and fine-tuning procedures, and share our trained artifacts. We identify several next steps and discuss important open problems toward automated modeling.
引用
收藏
页码:261 / 299
页数:39
相关论文
共 109 条
[1]  
Achiam J, 2023, GPT-4 technical report., DOI DOI 10.48550/ARXIV.2303.08774
[2]  
AhmadiTeshnizi Ali, 2023, ARXIV
[3]  
Ahmed T., 2024, ARXIV
[4]  
[Anonymous], 2012, On the Difficulty of Training Recurrent Neural Networks, DOI DOI 10.48550/ARXIV.1211.5063
[5]  
Araci Dogu, 2019, Finbert: Financial sentiment analysis with pre-trained language models
[6]   A TREE-BASED STATISTICAL LANGUAGE MODEL FOR NATURAL-LANGUAGE SPEECH RECOGNITION [J].
BAHL, LR ;
BROWN, PF ;
DESOUZA, PV ;
MERCER, RL .
IEEE TRANSACTIONS ON ACOUSTICS SPEECH AND SIGNAL PROCESSING, 1989, 37 (07) :1001-1008
[7]  
Balcan M.-F., 2022, LEIBNIZ INT P INFOR, V235, P3
[8]  
Balcan Maria-Florina, 2021, ADV NEURAL INFORM PR, V34
[9]  
Beldiceanu Nicolas, 2012, Principles and Practice of Constraint Programming. Proceedings 18th International Conference, CP 2012, P141, DOI 10.1007/978-3-642-33558-7_13
[10]  
Beltagy I, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P3615