z-logo
open-access-imgOpen Access
Unsupervised cross-lingual model transfer for named entity recognition with contextualized word representations
Author(s) -
Huijiong Yan,
Tao Qian,
Liang Xie,
Shanguang Chen
Publication year - 2021
Publication title -
plos one
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.99
H-Index - 332
ISSN - 1932-6203
DOI - 10.1371/journal.pone.0257230
Subject(s) - computer science , natural language processing , artificial intelligence , named entity recognition , adapter (computing) , transformer , transfer of learning , word (group theory) , benchmark (surveying) , language model , task (project management) , linguistics , philosophy , physics , management , geodesy , quantum mechanics , voltage , economics , geography , operating system
Named entity recognition (NER) is one fundamental task in the natural language processing (NLP) community. Supervised neural network models based on contextualized word representations can achieve highly-competitive performance, which requires a large-scale manually-annotated corpus for training. While for the resource-scarce languages, the construction of such as corpus is always expensive and time-consuming. Thus, unsupervised cross-lingual transfer is one good solution to address the problem. In this work, we investigate the unsupervised cross-lingual NER with model transfer based on contextualized word representations, which greatly advances the cross-lingual NER performance. We study several model transfer settings of the unsupervised cross-lingual NER, including (1) different types of the pretrained transformer-based language models as input, (2) the exploration strategies of the multilingual contextualized word representations, and (3) multi-source adaption. In particular, we propose an adapter-based word representation method combining with parameter generation network (PGN) better to capture the relationship between the source and target languages. We conduct experiments on a benchmark ConLL dataset involving four languages to simulate the cross-lingual setting. Results show that we can obtain highly-competitive performance by cross-lingual model transfer. In particular, our proposed adapter-based PGN model can lead to significant improvements for cross-lingual NER.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here