Loughborough University
Leicestershire, UK
LE11 3TU
+44 (0)1509 263171
Loughborough University

Loughborough University Institutional Repository

Please use this identifier to cite or link to this item: https://dspace.lboro.ac.uk/2134/20273

Title: Tracking object poses in the context of robust body pose estimates
Authors: Darby, John
Li, Baihua
Costen, Nicholas
Issue Date: 2014
Publisher: © Elsevier
Citation: DARBY, J., LI, B. and COSTEN, N., 2014. Tracking object poses in the context of robust body pose estimates. Computer Vision and Image Understanding, 127, pp.57-72
Abstract: This work focuses on tracking objects being used by humans. These objects are often small, fast moving and heavily occluded by the user. Attempting to recover their 3D position and orientation over time is a challenging research problem. To make progress we appeal to the fact that these objects are often used in a consistent way. The body poses of different people using the same object tend to have similarities, and, when considered relative to those body poses, so do the respective object poses. Our intuition is that, in the context of recent advances in body-pose tracking from RGB-D data, robust object-pose tracking during human-object interactions should also be possible. We propose a combined generative and discriminative tracking framework able to follow gradual changes in object-pose over time but also able to re-initialise object-pose upon recognising distinctive body-poses. The framework is able to predict object-pose relative to a set of independent coordinate systems, each one centred upon a different part of the body. We conduct a quantitative investigation into which body parts serve as the best predictors of object-pose over the course of different interactions. We find that while object-translation should be predicted from nearby body parts, object-rotation can be more robustly predicted by using a much wider range of body parts. Our main contribution is to provide the first object-tracking system able to estimate 3D translation and orientation from RGB-D observations of human-object interactions. By tracking precise changes in object-pose, our method opens up the possibility of more detailed computational reasoning about human-object interactions and their outcomes. For example, in assistive living systems that go beyond just recognising the actions and objects involved in everyday tasks such as sweeping or drinking, to reasoning that a person has missed sweeping under the chair or not drunk enough water today. © 2014 Elsevier B.V. All rights reserved.
Version: Accepted for publication
DOI: 10.1016/j.cviu.2014.06.009
URI: https://dspace.lboro.ac.uk/2134/20273
Publisher Link: http://dx.doi.org/10.1016/j.cviu.2014.06.009
ISSN: 1077-3142
Appears in Collections:Published Articles (Computer Science)

Files associated with this item:

File Description SizeFormat
John-CVIU+2014-accepted.pdfAccepted version4.71 MBAdobe PDFView/Open


SFX Query

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.