Home >

news ヘルプ

論文・著書情報


タイトル
和文: 
英文:Cross-View Human Action Recognition from Depth Maps Using Spectral Graph Sequences 
著者
和文: Kerola Tommi Matias, 井上 中順, 篠田浩一.  
英文: Tommi Kerola, Nakamasa Inoue, Koichi Shinoda.  
言語 English 
掲載誌/書名
和文: 
英文:Elsevier Journal of Computer Vision and Image Understanding (CVIU) 
巻, 号, ページ vol. 154        pp. 108-126
出版年月 2017年1月1日 
出版者
和文: 
英文:ElsevierInc. 
会議名称
和文: 
英文: 
開催地
和文: 
英文: 
ファイル
公式リンク http://www.sciencedirect.com/science/article/pii/S1077314216301588
 
DOI https://doi.org/10.1016/j.cviu.2016.10.004
アブストラクト We present a method for view-invariant action recognition from depth cameras based on graph signal processing techniques. Our framework leverages a novel graph representation of an action as a tempo- ral sequence of graphs, onto which we apply a spectral graph wavelet transform for creating our fea- ture descriptor. We evaluate two view-invariant graph types: skeleton-based and keypoint-based. The skeleton-based descriptor captures the spatial pose of the subject, whereas the keypoint-based is able to capture complementary information about human-object interaction and the shape of the point cloud. We investigate the effectiveness of our method by experiments on five publicly available datasets. By the graph structure, our method captures the temporal interaction between depth map interest points and achieves a 19.8% increase in performance compared to state-of-the-art results for cross-view action recog- nition, and competing results for frontal-view action recognition and human-object interaction. Namely, our method results in 90.8% accuracy on the cross-view N-UCLA Multiview Action3D dataset and 91.4% accuracy on the challenging MSRAction3D dataset in the cross-subject setting. For human-object interac- tion, our method achieves 72.3% accuracy on the Online RGBD Action dataset. We also achieve 96.0% and 98.8% accuracy on the MSRActionPairs3D and UCF-Kinect datasets, respectively.

©2007 Tokyo Institute of Technology All rights reserved.