Residential College | false |
Status | 已發表Published |
Context-aware self-attention networks | |
Yang,Baosong1; Li,Jian2; Wong,Derek F.1![]() ![]() | |
2019 | |
Conference Name | 33rd AAAI Conference on Artificial Intelligence, AAAI 2019, 31st Annual Conference on Innovative Applications of Artificial Intelligence, IAAI 2019 and the 9th AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019 |
Source Publication | 33rd AAAI Conference on Artificial Intelligence, AAAI 2019, 31st Innovative Applications of Artificial Intelligence Conference, IAAI 2019 and the 9th AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019
![]() |
Pages | 387-394 |
Conference Date | 27 January 2019through 1 February 2019 |
Conference Place | Honolulu |
Publisher | AAAI Press |
Abstract | Self-attention model has shown its flexibility in parallel computation and the effectiveness on modeling both long- and short-term dependencies. However, it calculates the dependencies between representations without considering the contextual information, which has proven useful for modeling dependencies among neural representations in various natural language tasks. In this work, we focus on improving self-attention networks through capturing the richness of context. To maintain the simplicity and flexibility of the self-attention networks, we propose to contextualize the transformations of the query and key layers, which are used to calculate the relevance between elements. Specifically, we leverage the internal representations that embed both global and deep contexts, thus avoid relying on external resources. Experimental results on WMT14 English-German and WMT17 Chinese-English translation tasks demonstrate the effectiveness and universality of the proposed methods. Furthermore, we conducted extensive analyses to quantify how the context vectors participate in the self-attention model. |
DOI | 10.1109/HPCC-DSS-SmartCity-DependSys57074.2022.00322 |
URL | View the original |
Indexed By | CPCI-S |
Language | 英語English |
WOS Research Area | Computer Science ; Engineering |
WOS Subject | Computer Science, Artificial Intelligence ; Computer Science, Theory & Methods ; Engineering, Electrical & Electronic |
WOS ID | WOS:000485292600048 |
Scopus ID | 2-s2.0-85090803656 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Corresponding Author | Tu,Zhaopeng |
Affiliation | 1.NLP2CT Lab,Department of Computer and Information Science,University of Macau,Macao 2.Chinese University of Hong Kong,Hong Kong 3.Tencent AI Lab, |
First Author Affilication | University of Macau |
Recommended Citation GB/T 7714 | Yang,Baosong,Li,Jian,Wong,Derek F.,et al. Context-aware self-attention networks[C]:AAAI Press, 2019, 387-394. |
APA | Yang,Baosong., Li,Jian., Wong,Derek F.., Chao,Lidia S.., Wang,Xing., & Tu,Zhaopeng (2019). Context-aware self-attention networks. 33rd AAAI Conference on Artificial Intelligence, AAAI 2019, 31st Innovative Applications of Artificial Intelligence Conference, IAAI 2019 and the 9th AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, 387-394. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment