• Corpus ID: 3656512

Real-Time Human-Robot Interaction for a Service Robot Based on 3D Human Activity Recognition and Human-like Decision Mechanism

  title={Real-Time Human-Robot Interaction for a Service Robot Based on 3D Human Activity Recognition and Human-like Decision Mechanism},
  author={Kang Li and Shiying Sun and Jinting Wu and Xiaoguang Zhao and Min Tan},
This paper describes the development of a real-time Human-Robot Interaction (HRI) system for a service robot based on 3D human activity recognition and human-like decision mechanism. The Human-Robot Interactive (HRI) system, which allows one person to interact with a service robot using natural body language, collects sequences of 3D skeleton joints comprising rich human movement information about the user via Microsoft Kinect. This information is used to train a three-layer Long-Short-Term… 

Figures and Tables from this paper


A Gesture Based Interface for Human-Robot Interaction
A gesture interface for the control of a mobile robot equipped with a manipulator uses a camera to track a person and recognize gestures involving arm motion and is combined with the Viterbi algorithm for the recognition of gestures defined through arm motion.
Body activity interaction for a service robot
This paper designs a body activity interaction system for a service robot using Microsoft Kinect to capture owner body movements and provide input signals of the action recognition module, which can analyze meaning ofOwner body movements effectively and send orders to service robot quickly.
Human–Robot Interaction
Objective: The current status of human–robot interaction (HRI) is reviewed, and key current research challenges for the human factors community are described. Background: Robots have evolved from
Gesture Spotting and Recognition for Human–Robot Interaction
The proposed method can be efficient and effective in HRI, for automatic recognition of whole-body key gestures from motion sequences, and a sophisticated method of designing a transition gesture model is proposed.
Real-Time Person Tracking and Pointing Gesture Recognition for Human-Robot Interaction
It is shown experimentally, that the gesture recognition performance can be improved significantly by using visually gained information about head orientation as an additional feature.
Sequential learning for multimodal 3D human activity recognition with Long-Short Term Memory
This paper uses three-layer Long Short Term Memory (LSTM) to model long-term contextual information of temporal skeleton sequences for human activities which are represented by the trajectories of skeleton joints, and adds dropout mechanism and L2 regularization to the output to avoid overfitting, and obtain better representation for feature modeling.
Gesture recognition based on arm tracking for human-robot interaction
A novel approach for hand gesture recognition that utilizes upper body part tracking in a 9-dimensional configuration space and two Multi-Layer Perceptron/Radial Basis Function (MLP/RBF) neural network classifiers, one for each arm.
Real-time human action recognition based on depth motion maps
A l2-regularized collaborative representation classifier with a distance-weighted Tikhonov matrix is employed for action recognition, shown to be computationally efficient allowing it to run in real-time.
Hierarchical recurrent neural network for skeleton based action recognition
  • Yong Du, Wei Wang, Liang Wang
  • Computer Science
    2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
  • 2015
This paper proposes an end-to-end hierarchical RNN for skeleton based action recognition, and demonstrates that the model achieves the state-of-the-art performance with high computational efficiency.
An End-to-End Spatio-Temporal Attention Model for Human Action Recognition from Skeleton Data
This work proposes an end-to-end spatial and temporal attention model for human action recognition from skeleton data on top of the Recurrent Neural Networks with Long Short-Term Memory (LSTM), which learns to selectively focus on discriminative joints of skeleton within each frame of the inputs and pays different levels of attention to the outputs of different frames.