Abstract
Recently, human action recognition has made satisfactory progress on almost daily activities. The emergence of depth sensors makes it more convenient to obtain human skeleton information from videos. Many problems in action recognition can be solved by using joint points, such as light changes and background noise. However, the unfixed viewpoint is one of the difficulties that affect action recognition, which is still under-explored for most existing skeleton-based works. Besides, training a complex deep model increases hardware complexity and cost. The proposed model uses a view adaptive asymmetric convolutional network (VA-ACN) to extract skeleton features from the skeleton data and transform the original body perspective into a new observable viewpoint. The ResNet50-based backbone is improved for high-performance classification. Compared with recent state-of-the-art models, the proposed model improves feature extraction and classification performance without increasing extra computation cost. Hence, it saves training time while improving the performance of action recognition under certain conditions. Experimental results show that the proposed model outperforms state-of-the-art methods.
Original language | English |
---|---|
Title of host publication | 2021 27th International Conference on Mechatronics and Machine Vision in Practice, M2VIP 2021 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
Pages | 475-479 |
Number of pages | 5 |
ISBN (Electronic) | 9781665431538 |
ISBN (Print) | 9781665431545 |
DOIs | |
Publication status | Published - 7 Jan 2022 |
Event | 2021 27th International Conference on Mechatronics and Machine Vision in Practice, M2VIP 2021 - Shanghai, China Duration: 26 Nov 2021 → 28 Nov 2021 |
Conference
Conference | 2021 27th International Conference on Mechatronics and Machine Vision in Practice, M2VIP 2021 |
---|---|
Country/Territory | China |
City | Shanghai |
Period | 26/11/21 → 28/11/21 |