keyboard_arrow_up
Naive Bayesian Fusion for Action Recognition from Kinect

Authors

Amel Ben Mahjoub1, Mohamed Ibn Khedher2, Mohamed Atri1 and Mounim A. El Yacoubi2, 1Monastir University, Tunisia and 2University of Paris Saclay, France

Abstract

The recognition of human actions based on three-dimensional depth data has become a very active research field in computer vision. In this paper, we study the fusion at the feature and decision levels for depth data captured by a Kinect camera to improve action recognition. More precisely, from each depth video sequence, we compute Depth Motion Maps (DMM) from three projection views: front, side and top. Then shape and texture features are extracted from the obtained DMMs. These features are based essentially on Histogram of Oriented Gradients (HOG) and Local Binary Patterns (LBP) descriptors. We propose to use two fusion levels. The first is a feature fusion level and is based on the concatenation of HOG and LBP descriptors. The second, a score fusion level, based on the naive-Bayes combination approach, aggregates the scores of three classifiers: a collaborative representation classifier, a sparse representation classifier and a kernel based extreme learning machine classifier. The experimental results conducted on two public datasets, Kinect v2 and UTD-MHAD, show that our approach achieves a high recognition accuracy and outperforms several existing methods.

Keywords

Action recognition, Depth motion maps, Features fusion, Score fusion, Naive Bayesian fusion, RGB-D.

Full Text  Volume 7, Number 16