Head pose estimation for real-time low-resolution video

D. Pol, van der, R.H. Cuijpers, J.F. Juola

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

5 Citations (Scopus)

Abstract

Gaze direction is an important communicative cue. In order to use this cue for human-robot interaction, software needs to be developed that enables the estimation of head pose. We began by designing an application that is be able to make a good estimate of the head pose, and, contrary to earlier neural network approaches, that works for non-optimal lighting conditions. Initial results show that the approach using multiple networks trained with differing datasets, gives a good estimate of head pose, and it works well in poor lighting conditions. The solution is not optimal yet. Smart selection rules taking into account different lighting conditions would enable us to select the neural networks trained with images with similar lighting conditions. This research will allow us to use head orientation cues in Human-Robot interaction with low-resolution cameras and in poor lighting conditions. The software allows the robot to give a timely reaction to the dynamical communicative cues used by humans.
Original languageEnglish
Title of host publicationProceedings of the 28th Annual European Conference on Cognitive Ergonomics (ECCE 2010), 25-27 August 2010, Delft, the Netherlands
EditorsM. Neerincx, W.P. Brinkman
Place of PublicationNew York
PublisherAssociation for Computing Machinery, Inc
Pages353-354
ISBN (Print)978-1-60558-946-6
DOIs
Publication statusPublished - 2010
Event28th European Conference on Cognitive Ergonomics (ECCE 2010) - Delft, Netherlands
Duration: 25 Aug 201027 Aug 2010
Conference number: 28
http://ii.tudelft.nl/ecce2010/

Conference

Conference28th European Conference on Cognitive Ergonomics (ECCE 2010)
Abbreviated titleECCE 2010
Country/TerritoryNetherlands
CityDelft
Period25/08/1027/08/10
Internet address

Fingerprint

Dive into the research topics of 'Head pose estimation for real-time low-resolution video'. Together they form a unique fingerprint.

Cite this