Proposed human action recognition (HAR) method for feature-level fusion of time domain signal attributes computed from wearable inertial sensors (accelerometer and gyrosc...
Abstract:
Automated recognition of human activities or actions has great significance as it incorporates wide-ranging applications, including surveillance, robotics, and personal h...Show MoreMetadata
Abstract:
Automated recognition of human activities or actions has great significance as it incorporates wide-ranging applications, including surveillance, robotics, and personal health monitoring. Over the past few years, many computer vision-based methods have been developed for recognizing human actions from RGB and depth camera videos. These methods include space-time trajectory, motion encoding, key poses extraction, space-time occupancy patterns, depth motion maps, and skeleton joints. However, these camera-based approaches are affected by background clutter and illumination changes and applicable to a limited field of view only. Wearable inertial sensors provide a viable solution to these challenges but are subject to several limitations such as location and orientation sensitivity. Due to the complementary trait of the data obtained from the camera and inertial sensors, the utilization of multiple sensing modalities for accurate recognition of human actions is gradually increasing. This paper presents a viable multimodal feature-level fusion approach for robust human action recognition, which utilizes data from multiple sensors, including RGB camera, depth sensor, and wearable inertial sensors. We extracted the computationally efficient features from the data obtained from RGB-D video camera and inertial body sensors. These features include densely extracted histogram of oriented gradient (HOG) features from RGB/depth videos and statistical signal attributes from wearable sensors data. The proposed human action recognition (HAR) framework is tested on a publicly available multimodal human action dataset UTD-MHAD consisting of 27 different human actions. K-nearest neighbor and support vector machine classifiers are used for training and testing the proposed fusion model for HAR. The experimental results indicate that the proposed scheme achieves better recognition results as compared to the state of the art. The feature-level fusion of RGB and inertial sensors provides...
Proposed human action recognition (HAR) method for feature-level fusion of time domain signal attributes computed from wearable inertial sensors (accelerometer and gyrosc...
Published in: IEEE Access ( Volume: 7)
Funding Agency:
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Human Activities ,
- Action Recognition ,
- Human Activity Recognition ,
- Robust Recognition ,
- Feature-level Fusion ,
- Support Vector Machine ,
- K-nearest Neighbor ,
- Support Vector Machine Classifier ,
- Recognition Accuracy ,
- Image Sensor ,
- Inertial Measurement Unit ,
- Wearable Sensors ,
- Multiple Sensors ,
- Depth Camera ,
- Recognition Results ,
- RGB Camera ,
- Background Clutter ,
- Histogram Of Oriented Gradients ,
- RGB Video ,
- Deep Learning ,
- Decision-level Fusion ,
- Convolutional Neural Network ,
- Inertial Data ,
- Gyroscope ,
- Long Short-term Memory ,
- Maximum Accuracy ,
- Depth Features ,
- Video Sequences ,
- Sensor Modalities ,
- Long Short-term Memory Network
- Author Keywords
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Human Activities ,
- Action Recognition ,
- Human Activity Recognition ,
- Robust Recognition ,
- Feature-level Fusion ,
- Support Vector Machine ,
- K-nearest Neighbor ,
- Support Vector Machine Classifier ,
- Recognition Accuracy ,
- Image Sensor ,
- Inertial Measurement Unit ,
- Wearable Sensors ,
- Multiple Sensors ,
- Depth Camera ,
- Recognition Results ,
- RGB Camera ,
- Background Clutter ,
- Histogram Of Oriented Gradients ,
- RGB Video ,
- Deep Learning ,
- Decision-level Fusion ,
- Convolutional Neural Network ,
- Inertial Data ,
- Gyroscope ,
- Long Short-term Memory ,
- Maximum Accuracy ,
- Depth Features ,
- Video Sequences ,
- Sensor Modalities ,
- Long Short-term Memory Network
- Author Keywords