Abstract
Named entity recognition (NER) is a basic knowledge extraction task. At present, many domains face a lack of labeled data, but current models for low-resource NER does not utilize the features of domain text. In this paper, we propose the MAST-NER model to improve the NER performance on domain-specific text. This model introduces multiple type pools based on entity triggers, and enhances sequence tagging through a multi-head attention mechanism, where the query matrix is jointly constructed by each type of triggers. MAST-NER can take full advantages of entity triggers on domain text with similar sentence patterns, and enable each type of entity recognition to be enhanced. The experimental results show that the model in this paper can achieve higher cost-effectiveness, especially for domain datasets (up to 3.33%). For general domain datasets, this model also has a certain performance improvement.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Huang, Z., Xu, W., Yu, W.: Bidirectional LSTM-CRF models for sequence tagging. arXiv preprint arXiv:1508.01991 (2015)
Lin, B.Y., et al.: Triggerner: learning with entity triggers as explanations for named entity recognition. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 8503–8511 (2020)
Li, X., Feng, J., Meng, Y., Han, Q., Wu, F., Li, J.: A unified MRC framework for named entity recognition. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 5849–5859 (2020)
Luo, Y., Zhao, H.: Bipartite flat-graph network for nested named entity recognition. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 6408–6418 (2020)
Qiu, J., Wang, Q., Zhou, Y., Ruan, T., Gao, J.: Fast and accurate recognition of Chinese clinical named entities with residual dilated convolutions. In: 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 935–942. IEEE (2018)
Zeng, X., Li, Y., Zhai, Y., Zhang, Y.: Counterfactual generator: a weakly-supervised method for named entity recognition. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 7270–7280 (2020)
Mengge, X., Yu, B., Liu, T., Zhang, Y., Meng, E., Wang, B.: Porous lattice transformer encoder for Chinese NER. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 3831–3841 (2020)
Liu, T., Yao, J.-G., Lin, C.-Y.: Towards improving neural named entity recognition with gazetteers. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 5301–5307 (2019)
Kruengkrai, C., Hai Nguyen, T., Mahani Aljunied, S., Bing, L.: Improving low-resource named entity recognition using joint sentence and token labeling. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 5898–5905 (2020)
Liu, L., Ding, B., Bing, L., Joty, S., Si, L., Mulda, C.M.: A multilingual data augmentation framework for low-resource cross-lingual NER. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, vol. 1: Long Papers, pp. 5834–5846 (2021)
Safranchik, E., Luo, S., Bach, S.: Weakly supervised sequence tagging from noisy rules. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 5570–5578 (2020)
Li, Y., Song, Y., Jia, L., Gao, S., Li, Q., Qiu, M.: Intelligent fault diagnosis by fusing domain adversarial training and maximum mean discrepancy via ensemble learning. IEEE Trans. Indus. Informat. 17(4), 2833–2841 (2020)
Luo, Y., Wang, X., Cao, W.: A novel dataset-specific feature extractor for zero-shot learning. Neurocomputing 391, 74–82 (2020)
Xie, Z., Cao, W., Ming, Z.: A further study on biologically inspired feature enhancement in zero-shot learning. Int. J. Mach. Lear. Cybernet. 12(1), 257–269 (2020). https://doi.org/10.1007/s13042-020-01170-y
Devlin, J., Chang Kenton, M.-W., Toutanova, L.K.: BERT: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of NAACL-HLT, pp. 4171–4186 (2019)
Lin, Z., Feng, M., Nogueira dos Santos, C., Yu, M., Xiang, B., Zhou, B., Bengio, Y.: A structured self-attentive sentence embedding. arXiv preprint arXiv:1703.03130 (2017)
Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)
Han, X., et al. Overview of the CCKS 2019 knowledge graph evaluation track: entity, relation, event and QA. arXiv preprint arXiv:2003.03875 (2020)
Sang, E.T.K., De Meulder, F.: Introduction to the conll-2003 shared task: Language-independent named entity recognition. In: Proceedings of the Seventh Conference on Natural Language Learning at HLT-NAACL, pp. 142–147 (2003)
Li, J., et al.: Biocreative V CDR task corpus: a resource for chemical disease relation extraction. Database 2016 (2016)
Acknowledgment
This paper is based on a research project supported by National Key Research and Development Project (Grant No. 2018YFB1703104) and National Natural Science Foundation of China (Grant No. 61671157).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Xu, J., Li, M. (2022). MAST-NER: A Low-Resource Named Entity Recognition Method Based on Trigger Pool. In: Memmi, G., Yang, B., Kong, L., Zhang, T., Qiu, M. (eds) Knowledge Science, Engineering and Management. KSEM 2022. Lecture Notes in Computer Science(), vol 13370. Springer, Cham. https://doi.org/10.1007/978-3-031-10989-8_6
Download citation
DOI: https://doi.org/10.1007/978-3-031-10989-8_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-10988-1
Online ISBN: 978-3-031-10989-8
eBook Packages: Computer ScienceComputer Science (R0)