Research article Special Issues

Span-based model for overlapping entity recognition and multi-relations classification in the food domain


  • Received: 17 December 2021 Revised: 24 February 2022 Accepted: 15 March 2022 Published: 18 March 2022
  • Information extraction (IE) is an important part of the entire knowledge graph lifecycle. In the food domain, extracting information such as ingredient and cooking method from Chinese recipes is crucial to safety risk analysis and identification of ingredient. In comparison with English, due to the complex structure, the richness of information in word combination, and lack of tense, Chinese IE is much more challenging. This dilemma is particularly prominent in the food domain with high-density knowledge, imprecise syntactic structure. However, existing IE methods focus only on the features of entities in a sentence, such as context and position, and ignore features of the entity itself and the influence of self attributes on prediction of inter entity relationship. To solve the problems of overlapping entity recognition and multi-relations classification in the food domain, we propose a span-based model known as SpIE for IE. The SpIE uses the span representation for each possible candidate entity to capture span-level features, which transforms named entity recognition (NER) into a classification mission. Besides, SpIE feeds extra information about the entity into the relation classification (RC) model by considering the effect of entity's attributes (both the entity mention and entity type) on the relationship between entity pairs. We apply SpIE on two datasets and observe that SpIE significantly outperforms the previous neural approaches due to capture the feature of overlapping entity and entity attributes, and it remains very competitive in general IE.

    Citation: Mengqi Zhang, Lei Ma, Yanzhao Ren, Ganggang Zhang, Xinliang Liu. Span-based model for overlapping entity recognition and multi-relations classification in the food domain[J]. Mathematical Biosciences and Engineering, 2022, 19(5): 5134-5152. doi: 10.3934/mbe.2022240

    Related Papers:

  • Information extraction (IE) is an important part of the entire knowledge graph lifecycle. In the food domain, extracting information such as ingredient and cooking method from Chinese recipes is crucial to safety risk analysis and identification of ingredient. In comparison with English, due to the complex structure, the richness of information in word combination, and lack of tense, Chinese IE is much more challenging. This dilemma is particularly prominent in the food domain with high-density knowledge, imprecise syntactic structure. However, existing IE methods focus only on the features of entities in a sentence, such as context and position, and ignore features of the entity itself and the influence of self attributes on prediction of inter entity relationship. To solve the problems of overlapping entity recognition and multi-relations classification in the food domain, we propose a span-based model known as SpIE for IE. The SpIE uses the span representation for each possible candidate entity to capture span-level features, which transforms named entity recognition (NER) into a classification mission. Besides, SpIE feeds extra information about the entity into the relation classification (RC) model by considering the effect of entity's attributes (both the entity mention and entity type) on the relationship between entity pairs. We apply SpIE on two datasets and observe that SpIE significantly outperforms the previous neural approaches due to capture the feature of overlapping entity and entity attributes, and it remains very competitive in general IE.



    加载中


    [1] Y. Lin, S. Shen, Z. Liu, H. Luan, M. Sun, Neural relation extraction with selective attention over instances, in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, 1 (2016), 2124–2133. https://doi.org/10.18653/v1/P16-1200
    [2] M. Nickel, K. Murphy, V. Tresp, E. Gabrilovich, A review of relational machine learning for knowledge graphs, Proc. IEEE, 104 (2016), 11–33. https://doi.org/10.1109/JPROC.2015.2483592 doi: 10.1109/JPROC.2015.2483592
    [3] D. Nadeau, S. Sekine, A survey of named entity recognition and classification, Lingvisticae Investigationes, 30 (2007), 3–26. https://doi.org/10.1075/li.30.1.03nad doi: 10.1075/li.30.1.03nad
    [4] N. Bach, S. Badaskar, A review of relation extraction, 2007. Available from: https://www.cs.cmu.edu/nbach/papers/A-survey-on-Relation-Extraction.pdf.
    [5] M. Miwa, Y. Sasaki, Modeling joint entity and relation extraction with table representation, in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), (2014), 1858–1869. https://doi.org/10.3115/v1/D14-1200
    [6] Z. Zhong, D. Chen, A frustratingly easy approach for entity and relation extraction, in Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, (2021), 50–61. https://doi.org/10.18653/v1/2021.naacl-main.5
    [7] R. D. Semmel, J. Mayfield, Automated query formulation using an entity-relationship conceptual schema, J. Intell. Inf. Syst., 8 (1997), 267–290. https://doi.org/10.1023/A:1008629825231 doi: 10.1023/A:1008629825231
    [8] J. A. P. Sacenti, R. Fileto, R. Willrich, Knowledge graph summarization impacts on movie recommendations, J. Intell. Inf. Syst., 58 (2022), 43–66. https://doi.org/10.1007/s10844-021-00650-z doi: 10.1007/s10844-021-00650-z
    [9] Y. Li, Y. Chen, I. Podlubny, Stability of fractional-order nonlinear dynamic systems: Lyapunov direct method and generalized Mittag–Leffler stability, Comput. Math. Appl., 59 (2010), 1810–1821. https://doi.org/10.1016/j.camwa.2009.08.019 doi: 10.1016/j.camwa.2009.08.019
    [10] H. Zhou, S. Ning, Z. Liu, C. Lang, Z. Liu, B. Lei, Knowledge-enhanced biomedical named entity recognition and normalization: application to proteins and genes, BMC Bioinf., 21 (2020). https://doi.org/10.1186/s12859-020-3375-3 doi: 10.1186/s12859-020-3375-3
    [11] Y. Huang, B. Niu, An improved method for web text affective cognition computing based on knowledge graph, Comput. Mater. Continua, 59 (2019), 1–14. https://doi.org/10.32604/cmc.2019.06032 doi: 10.32604/cmc.2019.06032
    [12] J. Tan, Q. Qiu, W. Guo, T. Li, Research on the construction of a knowledge graph and knowledge reasoning model in the field of urban traffic, Sustainability, 13 (2021), 3191. https://doi.org/10.3390/su13063191 doi: 10.3390/su13063191
    [13] M. Habibi, L. Weber, M. Neves, D. L. Wiegandt, U. Leser, Deep learning with word embeddings improves biomedical named entity recognition, Bioinformatics, 33 (2017), i37–i48. https://doi.org/10.1093/bioinformatics/btx228 doi: 10.1093/bioinformatics/btx228
    [14] C. Friedman, P. O. Alderson, J. H. M. Austin, J. J. Cimino, S. B. Johnson, A general natural-language text processor for clinical radiology, J. Am. Med. Inf. Assoc., 1 (1994), 161–174. https://doi.org/10.1136/jamia.1994.95236146 doi: 10.1136/jamia.1994.95236146
    [15] J. Li, A. Sun, J. Han, C. Li, A survey on deep learning for named entity recognition, IEEE Trans. Knowl. Data Eng., 34 (2022), 50–70. https://doi.org/10.1109/TKDE.2020.2981314 doi: 10.1109/TKDE.2020.2981314
    [16] Y. LeCun, Y. Bengio, G. Hinton, Deep learning, Nature, 521 (2015), 436–444. https://doi.org/10.1038/nature14539 doi: 10.1038/nature14539
    [17] P. Gupta, H. Schütze, B. Andrassy, Table filling multi-task recurrent neural network for joint entity and relation extraction, in Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, (2016), 2537–2547. Available from: https://aclanthology.org/C16-1239.pdf.
    [18] G. Bekoulis, J. Deleu, T. Demeester, C. Develder, Joint entity recognition and relation extraction as a multi-head selection problem, Expert Syst. Appl., 114 (2018), 34–45. https://doi.org/10.1016/j.eswa.2018.07.032 doi: 10.1016/j.eswa.2018.07.032
    [19] M. Miwa, M. Bansal, End-to-end relation extraction using LSTMs on sequences and tree structures, in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, 1 (2016), 1105–1116. https://doi.org/10.18653/v1/P16-1105
    [20] D. Q. Nguyen, K. Verspoor, End-to-end neural relation extraction using deep biaffine attention, in Advances in Information Retrieval, 11437 (2019), 729–738. https://doi.org/10.1007/978-3-030-15712-8_47
    [21] S. Zheng, F. Wang, H. Bao, Y. Hao, P. Zhou, B. Xu, Joint extraction of entities and relations based on a novel tagging scheme, in Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 1 (2017), 1227–1236. https://doi.org/10.18653/v1/P17-1113
    [22] X. Ma, E. Hovy, End-to-end sequence labeling via bi-directional LSTM-CNNs-CRF, in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, 1 (2016), 1064–1074. https://doi.org/10.18653/v1/P16-1101
    [23] M. Cao, Z. Yang, L. Ling, H. Lin, J. Wang, Joint drug entity and relations extraction based on neural network, J. Comput. Res. Dev., 56 (2019), 1432–1440. https://crad.ict.ac.cn/EN/Y2019/V56/I7/1432
    [24] W. Lv, Z. Liao, S. Liu, Y. Zhang, MEIM: A multi-source software knowledge entity extraction integration model, Comput. Mater. Continua, 66 (2021), 1027–1042. https://doi.org/10.32604/cmc.2020.012478 doi: 10.32604/cmc.2020.012478
    [25] P. Zhou, W. Shi, J. Tian, Z. Qi, B. Li, H. Hao, et al., Attention-based bidirectional long short-term memory networks for relation classification, in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, 2 (2016), 207–212. https://doi.org/10.18653/v1/P16-2034
    [26] Z. Li, N. Ding, Z. Liu, H. Zheng, Y. Shen, Chinese relation extraction with multi-grained information and external linguistic knowledge, in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, (2019), 4377–4386. https://doi.org/10.18653/v1/P19-1430
    [27] R. Alzaidy, C. Caragea, C. L. Giles, Bi-lstm-crf sequence labeling for keyphrase extraction from scholarly documents, in The World Wide Web Conference, New York, (2019), 2551–2557. https://doi.org/10.1145/3308558.3313642
    [28] K. Lee, L. He, M. Lewis, L. Zettlemoyer, End-to-end neural coreference resolution, in Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, (2017), 188–197. https://doi.org/10.18653/v1/D17-1018
    [29] Y. Luan, D. Wadden, L. He, A. Shah, M. Ostendorf, H. Hajishirzi, A general framework for information extraction using dynamic span graphs, in Proceedings of the 2019 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, 1 (2019), 3036–3046. https://doi.org/10.18653/v1/N19-1308
    [30] D. Wadden, U. Wennberg, Y. Luan, H. Hajishirzi, Entity, relation, and event extraction with contextualized span representations, in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), (2019), 5784–5789. https://doi.org/10.18653/v1/D19-1585
    [31] J. Devlin, M. Chang, K. Lee, K. Toutanova, BERT: Pre-training of deep bidirectional transformers for language understanding, in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 1 (2019), 4171–4186. https://doi.org/10.18653/v1/N19-1423
    [32] Y. Cui, W. Che, T. Liu, B. Qin, Z. Yang, Pre-training with whole word masking for Chinese BERT, IEEE/ACM Trans. Audio, 29 (2021), 3504–3514. https://doi.org/10.1109/TASLP.2021.3124365 doi: 10.1109/TASLP.2021.3124365
    [33] K. Dixit, Y. Al-Onaizan, Span-level model for relation extraction, in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, (2019), 5308–5314. https://doi.org/10.18653/v1/P19-1525
    [34] L. Baldini Soares, N. FitzGerald, J. Ling, T. Kwiatkowski, Matching the blanks: Distributional similarity for relation learning, in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, (2019), 2895–2905. https://doi.org/10.18653/v1/P19-1279
    [35] Z. Zhang, X. Han, Z. Liu, X. Jiang, M. Sun, Q. Liu, ERNIE: Enhanced language representation with informative entities, in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, (2019), 1441–1451. https://doi.org/10.18653/v1/P19-1139
    [36] H. Peng, T. Gao, X. Han, Y. Lin, P. Li, Z. Liu, et al., Learning from context or names? An empirical study on neural relation extraction, in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), (2020), 3661–3672. 10.18653/v1/2020.emnlp-main.298
    [37] Y. Gao, Y. Wang, P. Wang, L. Gu, Medical named entity extraction from chinese resident admit notes using character and word attention-enhanced neural network, Int. J. Environ. Res. Public Health, 17 (2020), 1614. https://doi.org/10.3390/ijerph17051614 doi: 10.3390/ijerph17051614
    [38] G. Bekoulis, J. Deleu, T. Demeester, C. Develder, Adversarial training for multi-context joint entity and relation extraction, in Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, (2018), 2830–2836. https://doi.org/10.18653/v1/D18-1307
    [39] J. Ge, S. Li, F. Yiqiu, Joint extraction method of Chinese entity relationship based on mixture of characters and words, Appl. Res. Comput., 38 (2021), 2619–2623. https://doi.org/10.19734/j.issn.1001-3695.2021.01.0006 doi: 10.19734/j.issn.1001-3695.2021.01.0006
  • Reader Comments
  • © 2022 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2514) PDF downloads(109) Cited by(5)

Article outline

Figures and Tables

Figures(8)  /  Tables(7)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog