TY - JOUR
T1 - Context-Aware Attentive Multilevel Feature Fusion for Named Entity Recognition
AU - Yang, Zhiwei
AU - Ma, Jing
AU - Chen, Hechang
AU - Zhang, Jiawei
AU - Chang, Yi
N1 - Funding information:
This work was supported in part by the National Natural Science Foundation of China under Grant 61976102, Grant U19A2065, and Grant 61902145; in part by the National Science Foundation (NSF) under Grant IIS-1763365 and Grant IIS-2106972; and in part by the University of California at Davis and Hong Kong Baptist University (HKBU) One-Off Tier 2 Start-Up Grant RCOFSGT2/20-21/SCI/004. (Corresponding authors: Hechang Chen; Yi Chang.)
Publisher copyright:
© 2022 IEEE.
PY - 2024/1
Y1 - 2024/1
N2 - In the era of information explosion, named entity recognition (NER) has attracted widespread attention in the field of natural language processing, as it is fundamental to information extraction. Recently, methods of NER based on representation learning, e.g., character embedding and word embedding, have demonstrated promising recognition results. However, existing models only consider partial features derived from words or characters while failing to integrate semantic and syntactic information, e.g., capitalization, inter-word relations, keywords, and lexical phrases, from multilevel perspectives. Intuitively, multilevel features can be helpful when recognizing named entities from complex sentences. In this study, we propose a novel attentive multilevel feature fusion (AMFF) model for NER, which captures the multilevel features in the current context from various perspectives. It consists of four components to, respectively, capture the local character-level (CL), global character-level (CG), local word-level (WL), and global word-level (WG) features in the current context. In addition, we further define document-level features crafted from other sentences to enhance the representation learning of the current context. To this end, we introduce a novel context-aware attentive multilevel feature fusion (CAMFF) model based on AMFF, to fully leverage document-level features from all the previous inputs. The obtained multilevel features are then fused and fed into a bidirectional long short-term memory (BiLSTM)-conditional random field (CRF) network for the final sequence labeling. Extensive experiments on four benchmark datasets demonstrate that our proposed AMFF and CAMFF models outperform a set of state-of-the-art baseline methods and the features learned from multiple levels are complementary.
AB - In the era of information explosion, named entity recognition (NER) has attracted widespread attention in the field of natural language processing, as it is fundamental to information extraction. Recently, methods of NER based on representation learning, e.g., character embedding and word embedding, have demonstrated promising recognition results. However, existing models only consider partial features derived from words or characters while failing to integrate semantic and syntactic information, e.g., capitalization, inter-word relations, keywords, and lexical phrases, from multilevel perspectives. Intuitively, multilevel features can be helpful when recognizing named entities from complex sentences. In this study, we propose a novel attentive multilevel feature fusion (AMFF) model for NER, which captures the multilevel features in the current context from various perspectives. It consists of four components to, respectively, capture the local character-level (CL), global character-level (CG), local word-level (WL), and global word-level (WG) features in the current context. In addition, we further define document-level features crafted from other sentences to enhance the representation learning of the current context. To this end, we introduce a novel context-aware attentive multilevel feature fusion (CAMFF) model based on AMFF, to fully leverage document-level features from all the previous inputs. The obtained multilevel features are then fused and fed into a bidirectional long short-term memory (BiLSTM)-conditional random field (CRF) network for the final sequence labeling. Extensive experiments on four benchmark datasets demonstrate that our proposed AMFF and CAMFF models outperform a set of state-of-the-art baseline methods and the features learned from multiple levels are complementary.
KW - Attention mechanism
KW - multilevel feature extraction
KW - named entity recognition (NER)
KW - sequence labeling
UR - http://www.scopus.com/inward/record.url?scp=85131816813&partnerID=8YFLogxK
U2 - 10.1109/TNNLS.2022.3178522
DO - 10.1109/TNNLS.2022.3178522
M3 - Journal article
SN - 2162-237X
VL - 35
SP - 973
EP - 984
JO - IEEE Transactions on Neural Networks and Learning Systems
JF - IEEE Transactions on Neural Networks and Learning Systems
IS - 1
ER -