Abstract
Named entity recognition is a highly relevant research topic in the field of natural language processing owing to its widespread application, which mainly involves three types of nested entities, discontinuous entities and flat named entities. Notably, nested named entities are characterized by one entity containing multiple sub-entities, ambiguous boundary definitions, and flexible structural formats. These features give rise to challenges such as semantic ambiguity, slow decoding efficiency, error propagation, and information loss. Therefore, to effectively address these issues and enhance classification performance, it is critical to integrate information sources such as internal markers, neighbouring word pairs, first and last word pairs, labels, and related spans. This is achieved in the present study via a newly proposed nested named entity recognition model based on triple cross affine attention. The proposed model encodes the input text using the BERT model and Bi-LSTM before extracting relevant features from the input text by applying DCNN. The extracted feature sequences are used as input into the triple cross affine attention module which computes the scores, allowing the model to classify and predict the outputs using the MLP layer. The experimental results demonstrate that the precision, recall, and F1 value metrics of the proposed method outperform other existing benchmark algorithmic models when applied to the ACE2004, ACE2005, and GENIA standard datasets. Additionally, it exhibits superior recognition performance in nested named entity recognition.
Keywords
Get full access to this article
View all access options for this article.
