1
|
Wang F, Lu J, Fan Z, Ren C, Geng X. Continuous motion estimation of lower limbs based on deep belief networks and random forest. THE REVIEW OF SCIENTIFIC INSTRUMENTS 2022; 93:044106. [PMID: 35489877 DOI: 10.1063/5.0057478] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/21/2021] [Accepted: 03/18/2022] [Indexed: 06/14/2023]
Abstract
Due to the lag problem of traditional sensor acquisition data, the following movement of exoskeleton robots can affect the comfort of the wearer and even the normal movement pattern of the wearer. In order to solve the problem of lag in exoskeleton motion control, this paper designs a continuous motion estimation method for lower limbs based on the human surface electromyographic (sEMG) signal and achieves the recognition of the motion intention of the wearer through a combination of the deep belief network (DBN) and random forest (RF) algorithm. First, the motion characteristics of human lower limbs are analyzed, and the hip-knee angle and sEMG signal related to lower limb motion are collected and extracted; then, the DBN is used in the dimensionality reduction of the sEMG signal feature values; finally, the motion intention of the wearer is predicted using the RF model optimized by the genetic algorithm. The experimental results show that the root mean square error of knee and hip prediction results of the combined algorithm proposed in this article improved by 0.2573° and 0.3375°, respectively, compared to the algorithm with dimensionality reduction by principal component analysis, and the single prediction time is 0.28 ms less than that before dimensionality reduction, provided that other conditions are exactly the same.
Collapse
Affiliation(s)
- Fei Wang
- Faculty of Robot Science and Engineering, Northeastern University, No. 195, Innovation Road, Hunnan District, Shenyang 110819, China
| | - Jian Lu
- Faculty of Robot Science and Engineering, Northeastern University, No. 195, Innovation Road, Hunnan District, Shenyang 110819, China
| | - Zhibo Fan
- College of Information Science and Engineering, Northeastern University, No. 3-11, Wenhua Road, Heping District, Shenyang 110169, China
| | - Chuanjian Ren
- College of Information Science and Engineering, Northeastern University, No. 3-11, Wenhua Road, Heping District, Shenyang 110169, China
| | - Xin Geng
- College of Information Science and Engineering, Northeastern University, No. 3-11, Wenhua Road, Heping District, Shenyang 110169, China
| |
Collapse
|
2
|
Supraja P, Tom RJ, Tiwari RS, Vijayakumar V, Liu Y. 3D convolution neural network-based person identification using gait cycles. EVOLVING SYSTEMS 2021. [DOI: 10.1007/s12530-021-09397-y] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
3
|
Chen X, Luo X, Weng J, Luo W, Li H, Tian Q. Multi-View Gait Image Generation for Cross-View Gait Recognition. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 30:3041-3055. [PMID: 33544673 DOI: 10.1109/tip.2021.3055936] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Gait recognition aims to recognize persons' identities by walking styles. Gait recognition has unique advantages due to its characteristics of non-contact and long-distance compared with face and fingerprint recognition. Cross-view gait recognition is a challenge task because view variance may produce large impact on gait silhouettes. The development of deep learning has promoted cross-view gait recognition performances to a higher level. However, performances of existing deep learning-based cross-view gait recognition methods are limited by lack of gait samples under different views. In this paper, we take a Multi-view Gait Generative Adversarial Network (MvGGAN) to generate fake gait samples to extend existing gait datasets, which provides adequate gait samples for deep learning-based cross-view gait recognition methods. The proposed MvGGAN method trains a single generator for all view pairs involved in single or multiple datasets. Moreover, we perform domain alignment based on projected maximum mean discrepancy to reduce the influence of distribution divergence caused by sample generation. The experimental results on CASIA-B and OUMVLP dataset demonstrate that fake gait samples generated by the proposed MvGGAN method can improve performances of existing state-of-the-art cross-view gait recognition methods obviously on both single-dataset and cross-dataset evaluation settings.
Collapse
|
4
|
Ensemble Learning for Skeleton-Based Body Mass Index Classification. APPLIED SCIENCES-BASEL 2020. [DOI: 10.3390/app10217812] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
In this study, we performed skeleton-based body mass index (BMI) classification by developing a unique ensemble learning method for human healthcare. Traditionally, anthropometric features, including the average length of each body part and average height, have been utilized for this kind of classification. Average values are generally calculated for all frames because the length of body parts and the subject height vary over time, as a result of the inaccuracy in pose estimation. Thus, traditionally, anthropometric features are measured over a long period. In contrast, we controlled the window used to measure anthropometric features over short/mid/long-term periods. This approach enables our proposed ensemble model to obtain robust and accurate BMI classification results. To produce final results, the proposed ensemble model utilizes multiple k-nearest neighbor classifiers trained using anthropometric features measured over several different time periods. To verify the effectiveness of the proposed model, we evaluated it using a public dataset. The simulation results demonstrate that the proposed model achieves state-of-the-art performance when compared with benchmark methods.
Collapse
|
5
|
Pham HH, Salmane H, Khoudour L, Crouzil A, Velastin SA, Zegers P. A Unified Deep Framework for Joint 3D Pose Estimation and Action Recognition from a Single RGB Camera. SENSORS (BASEL, SWITZERLAND) 2020; 20:E1825. [PMID: 32218350 PMCID: PMC7180926 DOI: 10.3390/s20071825] [Citation(s) in RCA: 20] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/27/2020] [Revised: 03/22/2020] [Accepted: 03/23/2020] [Indexed: 12/05/2022]
Abstract
We present a deep learning-based multitask framework for joint 3D human pose estimation and action recognition from RGB sensors using simple cameras. The approach proceeds along two stages. In the first, a real-time 2D pose detector is run to determine the precise pixel location of important keypoints of the human body. A two-stream deep neural network is then designed and trained to map detected 2D keypoints into 3D poses. In the second stage, the Efficient Neural Architecture Search (ENAS) algorithm is deployed to find an optimal network architecture that is used for modeling the spatio-temporal evolution of the estimated 3D poses via an image-based intermediate representation and performing action recognition. Experiments on Human3.6M, MSR Action3D and SBU Kinect Interaction datasets verify the effectiveness of the proposed method on the targeted tasks. Moreover, we show that the method requires a low computational budget for training and inference. In particular, the experimental results show that by using a monocular RGB sensor, we can develop a 3D pose estimation and human action recognition approach that reaches the performance of RGB-depth sensors. This opens up many opportunities for leveraging RGB cameras (which are much cheaper than depth cameras and extensively deployed in private and public places) to build intelligent recognition systems.
Collapse
Affiliation(s)
- Huy Hieu Pham
- Cerema Research Center, 31400 Toulouse, France; (H.H.P.); (L.K.)
- Informatics Research Institute of Toulouse (IRIT), Université de Toulouse, CNRS, 31062 Toulouse, France;
- Vingroup Big Data Institute (VinBDI), Hanoi 10000, Vietnam
| | | | - Louahdi Khoudour
- Cerema Research Center, 31400 Toulouse, France; (H.H.P.); (L.K.)
| | - Alain Crouzil
- Informatics Research Institute of Toulouse (IRIT), Université de Toulouse, CNRS, 31062 Toulouse, France;
| | - Sergio A. Velastin
- School of Electronic Engineering and Computer Science, Queen Mary University of London, London E1 4NS, UK
- Zebra Technologies Corp., London SE1 9LQ, UK
- Department of Computer Science and Engineering, University Carlos III de Madrid, 28270 Colmenarejo, Spain
| | | |
Collapse
|
6
|
Gait Recognition and Understanding Based on Hierarchical Temporal Memory Using 3D Gait Semantic Folding. SENSORS 2020; 20:s20061646. [PMID: 32188067 PMCID: PMC7146167 DOI: 10.3390/s20061646] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/22/2020] [Revised: 03/11/2020] [Accepted: 03/13/2020] [Indexed: 12/02/2022]
Abstract
Gait recognition and understanding systems have shown a wide-ranging application prospect. However, their use of unstructured data from image and video has affected their performance, e.g., they are easily influenced by multi-views, occlusion, clothes, and object carrying conditions. This paper addresses these problems using a realistic 3-dimensional (3D) human structural data and sequential pattern learning framework with top-down attention modulating mechanism based on Hierarchical Temporal Memory (HTM). First, an accurate 2-dimensional (2D) to 3D human body pose and shape semantic parameters estimation method is proposed, which exploits the advantages of an instance-level body parsing model and a virtual dressing method. Second, by using gait semantic folding, the estimated body parameters are encoded using a sparse 2D matrix to construct the structural gait semantic image. In order to achieve time-based gait recognition, an HTM Network is constructed to obtain the sequence-level gait sparse distribution representations (SL-GSDRs). A top-down attention mechanism is introduced to deal with various conditions including multi-views by refining the SL-GSDRs, according to prior knowledge. The proposed gait learning model not only aids gait recognition tasks to overcome the difficulties in real application scenarios but also provides the structured gait semantic images for visual cognition. Experimental analyses on CMU MoBo, CASIA B, TUM-IITKGP, and KY4D datasets show a significant performance gain in terms of accuracy and robustness.
Collapse
|
7
|
Pham HH, Salmane H, Khoudour L, Crouzil A, Zegers P, Velastin SA. Spatio⁻Temporal Image Representation of 3D Skeletal Movements for View-Invariant Action Recognition with Deep Convolutional Neural Networks. SENSORS 2019; 19:s19081932. [PMID: 31022945 PMCID: PMC6514994 DOI: 10.3390/s19081932] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/06/2019] [Revised: 04/10/2019] [Accepted: 04/17/2019] [Indexed: 11/22/2022]
Abstract
Designing motion representations for 3D human action recognition from skeleton sequences is an important yet challenging task. An effective representation should be robust to noise, invariant to viewpoint changes and result in a good performance with low-computational demand. Two main challenges in this task include how to efficiently represent spatio–temporal patterns of skeletal movements and how to learn their discriminative features for classification tasks. This paper presents a novel skeleton-based representation and a deep learning framework for 3D action recognition using RGB-D sensors. We propose to build an action map called SPMF (Skeleton Posture-Motion Feature), which is a compact image representation built from skeleton poses and their motions. An Adaptive Histogram Equalization (AHE) algorithm is then applied on the SPMF to enhance their local patterns and form an enhanced action map, namely Enhanced-SPMF. For learning and classification tasks, we exploit Deep Convolutional Neural Networks based on the DenseNet architecture to learn directly an end-to-end mapping between input skeleton sequences and their action labels via the Enhanced-SPMFs. The proposed method is evaluated on four challenging benchmark datasets, including both individual actions, interactions, multiview and large-scale datasets. The experimental results demonstrate that the proposed method outperforms previous state-of-the-art approaches on all benchmark tasks, whilst requiring low computational time for training and inference.
Collapse
Affiliation(s)
- Huy Hieu Pham
- Cerema, Project team STI, 1 avenue du Colonel Roche, F-31400 Toulouse, France.
- Informatics Research Institute of Toulouse (IRIT), Paul Sabatier University, Toulouse 31062, France.
| | - Houssam Salmane
- Cerema, Project team STI, 1 avenue du Colonel Roche, F-31400 Toulouse, France.
| | - Louahdi Khoudour
- Cerema, Project team STI, 1 avenue du Colonel Roche, F-31400 Toulouse, France.
| | - Alain Crouzil
- Informatics Research Institute of Toulouse (IRIT), Paul Sabatier University, Toulouse 31062, France.
| | - Pablo Zegers
- Aparnix, La Gioconda 4355, 10B, Las Condes, Santiago 7550076, Chile.
| | - Sergio A Velastin
- Cortexica Vision Systems Ltd., London SE1 9LQ, UK.
- School of Electronic Engineering and Computer Science, Queen Mary University of London, London E1 4NS, UK.
- Department of Computer Science, University Carlos III of Madrid, 28903 Leganés, Spain.
| |
Collapse
|
8
|
|
9
|
Khokhlova M, Migniot C, Morozov A, Sushkova O, Dipanda A. Normal and pathological gait classification LSTM model. Artif Intell Med 2019; 94:54-66. [PMID: 30871683 DOI: 10.1016/j.artmed.2018.12.007] [Citation(s) in RCA: 27] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2018] [Accepted: 12/27/2018] [Indexed: 10/27/2022]
Abstract
Computer vision-based clinical gait analysis is the subject of permanent research. However, there are very few datasets publicly available; hence the comparison of existing methods between each other is not straightforward. Even if the test data are in an open access, existing databases contain very few test subjects and single modality measurements, which limit their usage. The contributions of this paper are three-fold. First, we propose a new open-access multi-modal database acquired with the Kinect v.2 camera for the task of gait analysis. Second, we adapt to use the skeleton joint orientation data to calculate kinematic gait parameters to match golden-standard MOCAP systems. We propose a new set of features based on 3D low-limbs flexion dynamics to analyze the symmetry of a gait. Third, we design a Long-Short Term Memory (LSTM) ensemble model to create an unsupervised gait classification tool. The results show that joint orientation data provided by Kinect can be successfully used in an inexpensive clinical gait monitoring system, with the results moderately better than reported state-of-the-art for three normal/pathological gait classes.
Collapse
Affiliation(s)
| | | | - Alexey Morozov
- Kotelnikov Institute of Radio Engineering and Electronics of RAS, Moscow 125009, Russia
| | - Olga Sushkova
- Kotelnikov Institute of Radio Engineering and Electronics of RAS, Moscow 125009, Russia
| | - Albert Dipanda
- Le2i, FRE CNRS 2005, Univ. Bourgogne Franche-Comté, France.
| |
Collapse
|
10
|
El-Alfy H, Mitsugami I, Yagi Y. Gait Recognition Based on Normal Distance Maps. IEEE TRANSACTIONS ON CYBERNETICS 2018; 48:1526-1539. [PMID: 28600269 DOI: 10.1109/tcyb.2017.2705799] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Gait is a commonly used biometric for human recognition. Its main advantage relies on its ability to identify people at distances at which other biometrics fail. In this paper, we develop a new approach for gait recognition that combines the distance transform with curvatures of local contours. We call our gait feature template the normal distance map. Our method encodes both body shapes and boundary curvatures into a novel feature descriptor that is more robust than existing gait representations. We evaluate our approach on the widely used and challenging USF and CASIA-B datasets. Furthermore, we evaluate it on the OU-ISIR gait dataset, the largest one available in the literature, to obtain statistically reliable results. We verify our approach is significantly superior to the current state-of-the-art under most conditions.
Collapse
|
11
|
Zou Q, Ni L, Wang Q, Li Q, Wang S. Robust Gait Recognition by Integrating Inertial and RGBD Sensors. IEEE TRANSACTIONS ON CYBERNETICS 2018; 48:1136-1150. [PMID: 28368842 DOI: 10.1109/tcyb.2017.2682280] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Gait has been considered as a promising and unique biometric for person identification. Traditionally, gait data are collected using either color sensors, such as a CCD camera, depth sensors, such as a Microsoft Kinect, or inertial sensors, such as an accelerometer. However, a single type of sensors may only capture part of the dynamic gait features and make the gait recognition sensitive to complex covariate conditions, leading to fragile gait-based person identification systems. In this paper, we propose to combine all three types of sensors for gait data collection and gait recognition, which can be used for important identification applications, such as identity recognition to access a restricted building or area. We propose two new algorithms, namely EigenGait and TrajGait, to extract gait features from the inertial data and the RGBD (color and depth) data, respectively. Specifically, EigenGait extracts general gait dynamics from the accelerometer readings in the eigenspace and TrajGait extracts more detailed subdynamics by analyzing 3-D dense trajectories. Finally, both extracted features are fed into a supervised classifier for gait recognition and person identification. Experiments on 50 subjects, with comparisons to several other state-of-the-art gait-recognition approaches, show that the proposed approach can achieve higher recognition accuracy and robustness.
Collapse
|
12
|
Switonski A, Krzeszowski T, Josinski H, Kwolek B, Wojciechowski K. Gait recognition on the basis of markerless motion tracking and DTW transform. IET BIOMETRICS 2018. [DOI: 10.1049/iet-bmt.2017.0134] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022] Open
Affiliation(s)
- Adam Switonski
- Faculty of Automatic Control, Electronics and Computer ScienceSilesian University of Technologyul. Akademicka 1644‐100GliwicePoland
| | - Tomasz Krzeszowski
- Faculty of Electrical and Computer EngineeringRzeszow University of Technologyul. Wincentego Pola 235‐959RzeszowPoland
| | - Henryk Josinski
- Faculty of Automatic Control, Electronics and Computer ScienceSilesian University of Technologyul. Akademicka 1644‐100GliwicePoland
| | - Bogdan Kwolek
- Faculty of Computer Science, Electronics and TelecommunicationsAGH University of Science and Technology30 Mickiewicza Av.30‐059KrakowPoland
| | - Konrad Wojciechowski
- Research and Development Center of Polish‐Japanese Academy of Information TechnologyAleja Legionow 241‐902BytomPoland
| |
Collapse
|
13
|
Li W, Kuo CCJ, Peng J. Gait recognition via GEI subspace projections and collaborative representation classification. Neurocomputing 2018. [DOI: 10.1016/j.neucom.2017.10.049] [Citation(s) in RCA: 18] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
14
|
Das Choudhury S, Tjahjadi T. Clothing and carrying condition invariant gait recognition based on rotation forest. Pattern Recognit Lett 2016. [DOI: 10.1016/j.patrec.2016.05.009] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
15
|
Deng M, Wang C, Chen Q. Human gait recognition based on deterministic learning through multiple views fusion. Pattern Recognit Lett 2016. [DOI: 10.1016/j.patrec.2016.04.004] [Citation(s) in RCA: 30] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
16
|
|
17
|
Spatio-temporal feature extraction and representation for RGB-D human action recognition. Pattern Recognit Lett 2014. [DOI: 10.1016/j.patrec.2014.03.024] [Citation(s) in RCA: 48] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
18
|
Gait recognition based on joint distribution of motion angles. JOURNAL OF VISUAL LANGUAGES AND COMPUTING 2014. [DOI: 10.1016/j.jvlc.2014.10.004] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
|
19
|
Rogez G, Rihan J, Guerrero JJ, Orrite C. Monocular 3-D gait tracking in surveillance scenes. IEEE TRANSACTIONS ON CYBERNETICS 2014; 44:894-909. [PMID: 23955796 DOI: 10.1109/tcyb.2013.2275731] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
Gait recognition can potentially provide a noninvasive and effective biometric authentication from a distance. However, the performance of gait recognition systems will suffer in real surveillance scenarios with multiple interacting individuals and where the camera is usually placed at a significant angle and distance from the floor. We present a methodology for view-invariant monocular 3-D human pose tracking in man-made environments in which we assume that observed people move on a known ground plane. First, we model 3-D body poses and camera viewpoints with a low dimensional manifold and learn a generative model of the silhouette from this manifold to a reduced set of training views. During the online stage, 3-D body poses are tracked using recursive Bayesian sampling conducted jointly over the scene's ground plane and the pose-viewpoint manifold. For each sample, the homography that relates the corresponding training plane to the image points is calculated using the dominant 3-D directions of the scene, the sampled location on the ground plane and the sampled camera view. Each regressed silhouette shape is projected using this homographic transformation and is matched in the image to estimate its likelihood. Our framework is able to track 3-D human walking poses in a 3-D environment exploring only a 4-D state space with success. In our experimental evaluation, we demonstrate the significant improvements of the homographic alignment over a commonly used similarity transformation and provide quantitative pose tracking results for the monocular sequences with a high perspective effect from the CAVIAR dataset.
Collapse
|
20
|
2.5D multi-view gait recognition based on point cloud registration. SENSORS 2014; 14:6124-43. [PMID: 24686727 PMCID: PMC4029689 DOI: 10.3390/s140406124] [Citation(s) in RCA: 25] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/14/2014] [Revised: 03/24/2014] [Accepted: 03/24/2014] [Indexed: 11/24/2022]
Abstract
This paper presents a method for modeling a 2.5-dimensional (2.5D) human body and extracting the gait features for identifying the human subject. To achieve view-invariant gait recognition, a multi-view synthesizing method based on point cloud registration (MVSM) to generate multi-view training galleries is proposed. The concept of a density and curvature-based Color Gait Curvature Image is introduced to map 2.5D data onto a 2D space to enable data dimension reduction by discrete cosine transform and 2D principle component analysis. Gait recognition is achieved via a 2.5D view-invariant gait recognition method based on point cloud registration. Experimental results on the in-house database captured by a Microsoft Kinect camera show a significant performance gain when using MVSM.
Collapse
|
21
|
Krzeszowski T, Switonski A, Kwolek B, Josinski H, Wojciechowski K. DTW-Based Gait Recognition from Recovered 3-D Joint Angles and Inter-ankle Distance. COMPUTER VISION AND GRAPHICS 2014. [DOI: 10.1007/978-3-319-11331-9_43] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/02/2022]
|
22
|
On the improvement of human action recognition from depth map sequences using Space–Time Occupancy Patterns. Pattern Recognit Lett 2014. [DOI: 10.1016/j.patrec.2013.07.011] [Citation(s) in RCA: 47] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022]
|
23
|
Hak S, Mansard N, Stasse O, Laumond JP. Reverse control for humanoid robot task recognition. IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS. PART B, CYBERNETICS : A PUBLICATION OF THE IEEE SYSTEMS, MAN, AND CYBERNETICS SOCIETY 2012; 42:1524-37. [PMID: 22552575 DOI: 10.1109/tsmcb.2012.2193614] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Efficient methods to perform motion recognition have been developed using statistical tools. Those methods rely on primitive learning in a suitable space, for example, the latent space of the joint angle and/or adequate task spaces. Learned primitives are often sequential: A motion is segmented according to the time axis. When working with a humanoid robot, a motion can be decomposed into parallel subtasks. For example, in a waiter scenario, the robot has to keep some plates horizontal with one of its arms while placing a plate on the table with its free hand. Recognition can thus not be limited to one task per consecutive segment of time. The method presented in this paper takes advantage of the knowledge of what tasks the robot is able to do and how the motion is generated from this set of known controllers, to perform a reverse engineering of an observed motion. This analysis is intended to recognize parallel tasks that have been used to generate a motion. The method relies on the task-function formalism and the projection operation into the null space of a task to decouple the controllers. The approach is successfully applied on a real robot to disambiguate motion in different scenarios where two motions look similar but have different purposes.
Collapse
Affiliation(s)
- Sovannara Hak
- Institut des Systèmes Intelligents et de Robotique, Paris VI University, 75005 Paris, France.
| | | | | | | |
Collapse
|
24
|
Affiliation(s)
- Wei Bian
- Centre for Quantum Computation and Intelligent Systems, Faculty of Engineering and Information Technology, University of Technology, Sydney, Sydney, N.S.W. 2007, Australia.
| | | | | |
Collapse
|
25
|
Vakanski A, Mantegh I, Irish A, Janabi-Sharifi F. Trajectory Learning for Robot Programming by Demonstration Using Hidden Markov Model and Dynamic Time Warping. ACTA ACUST UNITED AC 2012; 42:1039-52. [PMID: 22411023 DOI: 10.1109/tsmcb.2012.2185694] [Citation(s) in RCA: 100] [Impact Index Per Article: 7.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
The main objective of this paper is to develop an efficient method for learning and reproduction of complex trajectories for robot programming by demonstration. Encoding of the demonstrated trajectories is performed with hidden Markov model, and generation of a generalized trajectory is achieved by using the concept of key points. Identification of the key points is based on significant changes in position and velocity in the demonstrated trajectories. The resulting sequences of trajectory key points are temporally aligned using the multidimensional dynamic time warping algorithm, and a generalized trajectory is obtained by smoothing spline interpolation of the clustered key points. The principal advantage of our proposed approach is utilization of the trajectory key points from all demonstrations for generation of a generalized trajectory. In addition, variability of the key points' clusters across the demonstrated set is employed for assigning weighting coefficients, resulting in a generalization procedure which accounts for the relevance of reproduction of different parts of the trajectories. The approach is verified experimentally for trajectories with two different levels of complexity.
Collapse
|
26
|
Hu M, Wang Y, Zhang Z. Maximisation of mutual information for gait-based soft biometric classification using Gabor features. IET BIOMETRICS 2012. [DOI: 10.1049/iet-bmt.2011.0004] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022] Open
|
27
|
STOP: Space-Time Occupancy Patterns for 3D Action Recognition from Depth Map Sequences. PROGRESS IN PATTERN RECOGNITION, IMAGE ANALYSIS, COMPUTER VISION, AND APPLICATIONS 2012. [DOI: 10.1007/978-3-642-33275-3_31] [Citation(s) in RCA: 138] [Impact Index Per Article: 10.6] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/11/2022]
|
28
|
Person Identification Using Full-Body Motion and Anthropometric Biometrics from Kinect Videos. COMPUTER VISION – ECCV 2012. WORKSHOPS AND DEMONSTRATIONS 2012. [DOI: 10.1007/978-3-642-33885-4_10] [Citation(s) in RCA: 33] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/02/2023]
|
29
|
Zhang Z, Hu M, Wang Y. A Survey of Advances in Biometric Gait Recognition. BIOMETRIC RECOGNITION 2011. [DOI: 10.1007/978-3-642-25449-9_19] [Citation(s) in RCA: 34] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
|