基于BERT的中文命名实体识别方法
Chinese Named Entity Recognition Method Based on BERT
DOI: 10.12677/AIRR.2021.103021, PDF,   
作者: 王希, 张传武, 刘东升:西南民族大学电子信息学院,四川 成都
关键词: 命名实体识别BERTBLSTM条件随机场Named Entity Recognition BERT BLSTM Conditional Random Fields
摘要: 由于中文与英文本身存在较大的差异,中文命名实体识别的研究存在一系列的挑战。目前来说,BLSTM-CRF模型使用最为广泛。该模型采用深度学习模型与统计模型相结合的方式进行中文命名实体识别,能够有效提取出文本中的上下文信息并考虑标签之间的关系。但由于中文存在多义字或词,存在一个句子中相同字词含义差别很大的情况,该模型在这种情况下实体识别的性能并不理想。为了更好地实现字表示既可以包含各种多样化的句法和语义表示,又可以对多义字进行建模,引入了BERT语言模型,此模型可以根据上下文信息计算出更高的全局性字词向量表示以及在句中的权重。BERT-BLSTM-CRF命名实体识别模型通过BERT预训练模型增强词向量的表示,BLSTM获取上下文语义标签序列,再使用CRF求得最优解。本文使用人民日报数据集对提出模型的进行实验测试,从实验结果可以发现,该模型的实体识别性能与传统模型相比有较大的提升。
Abstract: Due to the large differences between Chinese and English, there are a series of challenges in the re-search of Chinese named entity recognition. Currently, the BLSTM-CRF model is the most widely used. The model uses a combination of deep learning models and statistical models for Chinese named entity recognition, which can effectively extract contextual information in the text and con-sider the relationship between tags. However, due to the presence of polysemous characters or words in Chinese, the meaning of the same words in a sentence may be very different. In this case, the performance of the entity recognition of the model is not ideal. In order to better realize that the word representation can not only contain a variety of diversified syntax and semantic representa-tions, but also can model polysemous words, the BERT language model is introduced, which can calculate higher global word vector representations and weights in sentences based on context in-formation. The BERT-BLSTM-CRF named entity recognition model enhances the representation of word vectors through the BERT pre-training model, uses BLSTM to obtain contextual semantic label sequences, and then uses CRF to find the optimal solution. Using the People’s Daily data set to test the proposed model, it can be found that the entity recognition performance of the model is greatly improved compared with the traditional model.
文章引用:王希, 张传武, 刘东升. 基于BERT的中文命名实体识别方法[J]. 人工智能与机器人研究, 2021, 10(3): 215-223. https://doi.org/10.12677/AIRR.2021.103021

参考文献

[1] Lample, G., Ballesteros, M., Subramanian, S., et al. (2016) Neural Architectures for Named Entity Recognition. Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, 12-17 June 2016, 260-270. [Google Scholar] [CrossRef
[2] Yang, Z., Salakhutdinov, R. and Cohen, W. (2016) Multi-Task Cross-Lingual Sequence Tagging from Scratch.
[3] de Oliveira, D.M., Laender, A.H.F., Veloso, A., et al. (2013) FS-NER: A Lightweight Filter-Stream Approach to Named Entity Recognition on Twitter Data. Proceedings of the 22nd International Conference on World Wide Web, Rio de Janeiro, 13-17 May 2013, 597-604. [Google Scholar] [CrossRef
[4] 罗熹, 夏先运, 安莹, 陈先来. 结合多头自注意力机制与BiLSTM-CRF的中文临床实体识别[J]. 湖南大学学报(自然科学版), 2021, 48(4): 45-55.
[5] 王昊, 林克柔, 孟镇, 李心蕾. 文本表示及其特征生成对法律判决书中多类型实体识别的影响分析[J/OL]. 数据分析与知识发现, 1-26.
http://kns.cnki.net/kcms/detail/10.1478.G2.20210326.0959.002.html, 2021-04-25.
[6] 钟诗胜, 陈曦, 赵明航, 张永健. 引入词集级注意力机制的中文命名实体识别方法[J/OL]. 吉林大学学报(工学版), 1-7. 2021-04-26. [Google Scholar] [CrossRef
[7] 何玉洁, 杜方, 史英杰, 宋丽娟. 基于深度学习的命名实体识别研究综述[J/OL]. 计算机工程与应用, 1-17.
http://kns.cnki.net/kcms/detail/11.2127.TP.20210326.0937.002.html, 2021-04-21.
[8] 王玥. 基于深度学习的命名实体识别研究[D]: [硕士学位论文]. 昆明: 云南财经大学, 2019.
[9] 朱岩, 张利, 王煜. 基于RoBERTa-WWM的中文电子病历命名实体识别[J]. 计算机与现代化, 2021(2): 51-55.
[10] Collobert, R., Weston, J., Bottou, L., et al. (2011) Natural Language Processing (Almost) from Scratch. Journal of Machine Learning Research, 12, 2493-2537.
[11] Huang, Z., Xu, W. and Yu, K. (2015) Bidirectional LSTM-CRF Models for Sequence Tagging. arXiv preprint arXiv:1508.01991.
[12] Chiu, J.P.C. and Nichols, E. (2015) Named Entity Recognition with Bidirectional LSTM-CNNs. Transactions of the Association for Computational Linguistics, 4, 357-370. [Google Scholar] [CrossRef
[13] Ma, X. and Hovy, E. (2016) End-to-End Sequence Labeling via Bi-Directional LSTM-CNNs-CRF. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Berlin, 7-12 August 2016, 1064-1074. [Google Scholar] [CrossRef
[14] Rei, M., Crichton, G.K.O. and Pyysalo, S. (2016) Attending to Characters in Neural Sequence Labeling Models. Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, Osaka, 11-16 December 2016, 309-318.
[15] 王子牛, 姜猛, 高建瓴, 陈娅先. 基于BERT的中文命名实体识别方法[J]. 计算机科学, 2019, 46(S2): 138-142.
[16] Graves, A. and Schmidhuber, J. (2005) Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures. Neural Networks, 18, 602-610. [Google Scholar] [CrossRef] [PubMed]
[17] Vaswani, A., Shazeer, N., Parmar, N., et al. (2017) Attention Is All You Need. 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, 4-9 December 2017, 5998-6008.
[18] Peters, M.E., Neumann, M., Iyyer, M., et al. (2018) Deep Contextualized Word Representations. Proceedings of NAACL-HLT 2018, New Orleans, 1-6 June 2018, 2227-2237.
[19] Devlin, J., Chang, M.W., Lee, K., et al. (2018) BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. Proceedings of NAACL-HLT 2019, Minneapolis, 2-7 June 2019, 4171-4186.
[20] 袁培森, 李润隆, 王翀, 徐焕良. 基于BERT的水稻表型知识图谱中关系抽取研究[J/OL]. 农业机械学报, 1-10.
http://kns.cnki.net/kcms/detail/11.1964.S.20210315.1809.015.html, 2021-04-26.
[21] 谢腾, 杨俊安, 刘辉. 基于BERT-BiLSTM-CRF模型的中文实体识别[J]. 计算机系统应用, 2020, 29(7): 48-55.