Learning Concept Lengths Accelerates Concept Learning in ALC

被引:7
作者
Kouagou, N'Dah Jean [1 ]
Heindorf, Stefan [1 ]
Demir, Caglar [1 ]
Ngomo, Axel-Cyrille Ngonga [1 ]
机构
[1] Paderborn Univ, Paderborn, Germany
来源
SEMANTIC WEB, ESWC 2022 | 2022年 / 13261卷
关键词
Concept learning; Concept length; Structured machine learning; Description logic; Learning from examples; Prediction of concept lengths; GENE ONTOLOGY; DL-FOIL;
D O I
10.1007/978-3-031-06981-9_14
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Concept learning approaches based on refinement operators explore partially ordered solution spaces to compute concepts, which are used as binary classification models for individuals. However, the number of concepts explored by these approaches can grow to the millions for complex learning problems. This often leads to impractical runtimes. We propose to alleviate this problem by predicting the length of target concepts before the exploration of the solution space. By these means, we can prune the search space during concept learning. To achieve this goal, we compare four neural architectures and evaluate them on four benchmarks. Our evaluation results suggest that recurrent neural network architectures perform best at concept length prediction with a macro F-measure ranging from 38% to 92%. We then extend the CELOE algorithm, which learns ALC concepts, with our concept length predictor. Our extension yields the algorithm CLIP. In our experiments, CLIP is at least 7.5 x faster than other state-of-the-art concept learning algorithms for ALC-including CELOE-and achieves significant improvements in the F-measure of the concepts learned on 3 out of 4 datasets. For reproducibility, we provide our implementation in the public GitHub repository at https://github.com/dice-group/LearnALCLengths.
引用
收藏
页码:236 / 252
页数:17
相关论文
共 38 条
  • [1] Gene Ontology: tool for the unification of biology
    Ashburner, M
    Ball, CA
    Blake, JA
    Botstein, D
    Butler, H
    Cherry, JM
    Davis, AP
    Dolinski, K
    Dwight, SS
    Eppig, JT
    Harris, MA
    Hill, DP
    Issel-Tarver, L
    Kasarskis, A
    Lewis, S
    Matese, JC
    Richardson, JE
    Ringwald, M
    Rubin, GM
    Sherlock, G
    [J]. NATURE GENETICS, 2000, 25 (01) : 25 - 29
  • [2] Baader F, 2003, DESCRIPTION LOGIC HANDBOOK: THEORY, IMPLEMENTATION AND APPLICATIONS, P43
  • [3] Badea L., 2000, Inductive Logic Programming. 10th International Conference, ILP 2000. Proceedings (Lecture Notes in Artificial Intelligence Vol.1866), P40
  • [4] Bergstra J, 2012, J MACH LEARN RES, V13, P281
  • [5] Towards SPARQL-Based Induction for Large-Scale RDF Data Sets
    Bin, Simon
    Buehmann, Lorenz
    Lehmann, Jens
    Ngomo, Axel-Cyrille Ngonga
    [J]. ECAI 2016: 22ND EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, 285 : 1551 - 1552
  • [6] A semantic matching energy function for learning with multi-relational data Application to word-sense disambiguation
    Bordes, Antoine
    Glorot, Xavier
    Weston, Jason
    Bengio, Yoshua
    [J]. MACHINE LEARNING, 2014, 94 (02) : 233 - 259
  • [7] DL-Learner-A framework for inductive learning on the Semantic Web
    Buehmann, Lorenz
    Lehmann, Jens
    Westphal, Patrick
    [J]. JOURNAL OF WEB SEMANTICS, 2016, 39 : 15 - 24
  • [8] Cho KYHY, 2014, Arxiv, DOI arXiv:1406.1078
  • [9] A Survey on Knowledge Graph Embedding: Approaches, Applications and Benchmarks
    Dai, Yuanfei
    Wang, Shiping
    Xiong, Neal N.
    Guo, Wenzhong
    [J]. ELECTRONICS, 2020, 9 (05)
  • [10] Demir C, 2021, Arxiv, DOI arXiv:2008.03130