Home >

news ヘルプ

論文・著書情報


タイトル
和文:TransformerにおけるToken-Mixingの探索 
英文:Exploring Token-Mixing Structure for Transformer 
著者
和文: 浅倉 拓也, 宇都 有昭, 篠田 浩一.  
英文: Takuya Asakura, Kuniaki Uto, Koichi Shinoda.  
言語 Japanese 
掲載誌/書名
和文:人工知能学会全国大会 (第36回)論文集 
英文:Proceedings of the Annual Conference of JSAI 
巻, 号, ページ        
出版年月 2022年6月 
出版者
和文:一般社団法人 人工知能学会 
英文:Japanese Society for Artificial Intelligence 
会議名称
和文:人工知能学会全国大会 (第36回) 
英文: 
開催地
和文:京都 
英文: 
ファイル
公式リンク https://www.ai-gakkai.or.jp/jsai2022/
 
DOI https://doi.org/10.11517/pjsai.JSAI2022.0_3J4OS3b04
アブストラクト The Transformer model, which applies Channel-Mixing and Token-Mixing alternately to input data, has been developed for time-series data such as text and speech. Recent studies have shown that this model can also perform well image. Various improved models of transformers have been proposed for image processing, many of which have improved the structure of the fully connected layer, especially for Token-Mixing. However, these structures should be designed manually, which requires advanced knowledge about the characteristics of the target data. In this paper, we propose a method to automatically acquire Token-Mixing structures by learning the relationships between Tokens. In our experiments on the image classification tasks, the structure obtained by the proposed method achieves higher accuracy while having fewer parameters than the other Token-Mixing methods. We also visualized the Token-Mixing structures obtained by the proposed method, and observed that the proposed method tends to focus on spatially close Tokens.

©2007 Institute of Science Tokyo All rights reserved.