E-viri
Recenzirano
-
Wang, Yuke; Lu, Ling; Wu, Yang; Chen, Yinong
Expert systems with applications, 10/2022, Letnik: 203Journal Article
•Lexicon information has been proved to be very useful in Chinese Named Entity task.•The existing methods make insufficient use of lexicon information.•We propose a Polymorphic Graph Attention Network for fusion lexicon into character.•Our method can be easily combined with pre-trained or sequence encoding model.•The proposed method supports multi-head attention and has excellent inference speed. Fusing lexicon information into Chinese characters, which has normally a number of meanings, has been proven to be effective for Chinese Named Entity Recognition (NER). However, the existing approaches to incorporating a matched Chinese word into its composition characters only take the word as a whole (no subdivision or part), which failed to capture fine-grained correlation in word-character space and failed to make full use of lexicon information. Moreover, existing approaches use the fixed (static) weights between words and characters. This limits the performance of NER. Considering the fact that the same word-character pairs have different interactions in different contexts, the weights of matched word-character pairs should be dynamic rather than fixed. In this paper, we propose a Polymorphic Graph Attention Network (PGAT), aiming at capturing dynamic correlation between characters and matched words from multiple dimensions, to enhance the character representation. By obtaining matched words of characters from lexicon, we carefully map the word-character in four positions, which are B (begin), M (middle), E (end) and S (single word). The proposed semantic fusion unit based on Graph Attention Network (GAT) can dynamically modulate attention of matched words and characters in the four dimensions B, M, E, and S. Thus, it can explicitly capture fine-grained correlation between characters and matched words across each dimension. Experiments on four Chinese NER datasets show that PGAT outperforms the baseline models. It demonstrates the significance of the attention capture and fusion capabilities of the proposed polymorphic graph. Furthermore, PGAT is used in character representation layer, which makes it easier to be combined with pre-trained models like BERT and other sequence encoding models like CNN and Transformer.
![loading ... loading ...](themes/default/img/ajax-loading.gif)
Vnos na polico
Trajna povezava
- URL:
Faktor vpliva
Dostop do baze podatkov JCR je dovoljen samo uporabnikom iz Slovenije. Vaš trenutni IP-naslov ni na seznamu dovoljenih za dostop, zato je potrebna avtentikacija z ustreznim računom AAI.
Leto | Faktor vpliva | Izdaja | Kategorija | Razvrstitev | ||||
---|---|---|---|---|---|---|---|---|
JCR | SNIP | JCR | SNIP | JCR | SNIP | JCR | SNIP |
Baze podatkov, v katerih je revija indeksirana
Ime baze podatkov | Področje | Leto |
---|
Povezave do osebnih bibliografij avtorjev | Povezave do podatkov o raziskovalcih v sistemu SICRIS |
---|
Vir: Osebne bibliografije
in: SICRIS
To gradivo vam je dostopno v celotnem besedilu. Če kljub temu želite naročiti gradivo, kliknite gumb Nadaljuj.