Abstract
Traditional Named Entity Recognition (NER) research only deals with flat entities and ignores nested entities. The transition-based method maps a sentence to a designated forest to recognize nested entities by predicting an action sequence through a state transition system which includes transition actions and a state of structures. However, the subsequent transition actions are affected by the previous transition actions resulting in error propagation, and the method ignores the correlation between the structures. To tackle these issues, we propose a new transition-based attention model (NTAM) to recognize nested entities. First, the structures and transition actions of the state transition system are redefined to eliminate error propagation. The prediction of an action sequence is converted to the prediction of a series of states, which predict whether the words between the structures can form entities. Second, we introduce an attention mechanism that strengthens the association between the structures. Experiments on two public nested NER datasets outperform previous state-of-the-art models.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Yan, H., Deng, B., Li, X., Qiu, X.: TENER: adapting transformer encoder for named entity recognition. arXiv:1911.04474 (2019)
Zheng, C., Cai, Y., Xu, J., Leung, H.F., Xu, G.: A boundary-aware neural model for nested named entity recognition. In: EMNLP-IJCNLP (2019)
Wang, Y., Shindo, H., Matsumoto, Y., Watanabe, T.: Nested named entity recognition via explicitly excluding the influence of the best path. In: ACL (2021)
Lu, W., Roth, D.: Joint mention extraction and classification with mention hypergraphs. In: EMNLP (2015)
Katiyar, A., Cardie, C.: Nested named entity recognition revisited. In: NAACL (2018)
Wang, J., Shou, L., Chen, K., Chen, G.: Pyramid: a layered model for nested named entity recognition. In: ACL (2020)
Tan, C., Qiu, W., Chen, M., Wang, R., Huang, F.: Boundary enhanced neural span classification for nested named entity recognition. In: AAAI (2020)
Shen, Y., Ma, X., Tan, Z., Zhang, S., Wang, W., Lu, W.: Locate and label: a two-stage identifier for nested named entity recognition. In: ACL-IJCNLP (2021)
Lou, C., Yang, S., Tu, K.: Nested named entity recognition as latent lexicalized constituency parsing. In: ACL (2022)
Wan, J., Ru, D., Zhang, W., Yu, Y.: Nested named entity recognition with span-level graphs. In: ACL (2022)
Li, X., Feng, J., Meng, Y., Han, Q., Wu, F., Li, J.: A unified MRC framework for named entity recognition. In: ACL (2020)
Shen, Y., et al.: Parallel instance query network for named entity recognition. In: ACL (2022)
Wang, B., Lu, W., Wang, Y., Jin, H.: A neural transition-based model for nested mention recognition. In: EMNLP (2018)
Zhang, Y., Clark, S.: Transition-based parsing of the Chinese treebank using a global discriminative model. In: IWPT (2009)
Ji, Z., Xia, T., Han, M., Xiao, J.: A neural transition-based joint model for disease named entity recognition and normalization. In: ACL-IJCNLP (2021)
Ohta, T., Tateisi, Y., Kim, J.D.: The GENIA corpus: an annotated research abstract corpus in molecular biology domain. In: HLT (2002)
Guo, G., Hua, L., Xuemin, X., Pu, Z.: The initial statistic analysis on tagged corpus of people’s daily. JSCL (2005)
Lee, J., et al.: BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36, 1234–1240 (2019)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: ACL (2019)
Chiu, B., Crichton, G., Korhonen, A., Pyysalo, S.: How to train good word embeddings for biomedical NLP. In: BioNLP (2016)
Song, Y., Shi, S., Li, J., Zhang, H.: Directional skip-gram: explicitly distinguishing left and right context for word embeddings. In: NAACL (2018)
Li, Y., He, Y., Qian, L., Zhou, G.: Chinese nested named entity recognition corpus construction. J. Chin. Inf. Process. 32, 19–26 (2018)
Jin, Y., Xie, J., Wu, D.: Chinese nested named entity recognition based on hierarchical tagging. J. Shanghai Univ. (Nat. Sci. Edn.) 27, 1–9 (2022)
Yuan, Z., Zhang, H.: Improving named entity recognition of Chinese legal documents by lexical enhancement. In: ICAICA (2021)
Acknowledgments
The work described in this paper is supported by Zhejiang Provincial Natural Science Foundation of China (LGF22F020014), National Key Research and Development Program of China (2020YFB1707700), National Natural Science Foundation of China (62036009, U1909203).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Gao, N., Yang, B., Wang, Y., Chen, P. (2023). NTAM: A New Transition-Based Attention Model for Nested Named Entity Recognition. In: Liu, F., Duan, N., Xu, Q., Hong, Y. (eds) Natural Language Processing and Chinese Computing. NLPCC 2023. Lecture Notes in Computer Science(), vol 14303. Springer, Cham. https://doi.org/10.1007/978-3-031-44696-2_22
Download citation
DOI: https://doi.org/10.1007/978-3-031-44696-2_22
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-44695-5
Online ISBN: 978-3-031-44696-2
eBook Packages: Computer ScienceComputer Science (R0)