This paper presents a view-invariant approach to gait recognition in multi-camera scenarios exploiting a joint spatio-temporal data representation and analysis. First,
multi-view information is employed to generate a 3D voxel reconstruction of the scene under study. The analyzed subject
is tracked and its centroid and orientation allow recentering and aligning the volume associated to it, thus obtaining a representation invariant to translation, rotation and scaling. Temporal periodicity of the walking cycle is extracted to align the input data in the time domain. Finally,
Hyperspherical Radon Transform is presented as an efficient tool to obtain features from spatio-temporal gait templates for classification purposes. Experimental results
prove the validity and robustness of the proposed method for gait recognition tasks with several covariates.
CitationCanton-Ferrer, C.; Casas, J.; Pardas, M. Spatio-temporal alignment and hyperspherical radon transform for 3D gait recognition in multi-view environments. A: IEEE Computer-Society Conference on Computer Vision and Pattern Recognition Workshops. "2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops". San Francisco: 2010, p. 116-121.
All rights reserved. This work is protected by the corresponding intellectual and industrial property rights. Without prejudice to any existing legal exemptions, reproduction, distribution, public communication or transformation of this work are prohibited without permission of the copyright holder. If you wish to make any use of the work not provided for in the law, please contact: firstname.lastname@example.org