A robust and accurate 3D hand posture estimation method for interactive systems

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

In this paper, a new 3D hand posture estimation system using a single camera and 3 interactive systems are introduced. Existing hand gesture recognition systems estimate hand's 3D models based on image features such as contour or skin texture. However, it was difficult to estimate the wrist rotation because the contour and the texture data do not have enough information to distinguish hand's sides. To solve this problem, we propose a new 3D hand posture estimation system that uses data of nail positions. Nail positions are an important factor to recognize hand's sides. Using nail positions, it becomes possible to detect whether the camera is facing palm or dorsum. In addition, nail areas can be robustly extracted from a skin area by a simple image processing technique. Our Proposed system uses a database consists of data-sets of the hand's contour, the nail positions, and finger joint angles. To estimate the hand posture, the system first extracts the hand's contour and the nail positions from the captured image, and searches for a similar data-set from the database. The system then outputs the finger joint angles of the searched data-set. Our experimental results show high accuracy in the hand posture estimation with the wrist rotation.

Original languageEnglish
Title of host publicationTEI'10 - Proceedings of the 4th International Conference on Tangible, Embedded, and Embodied Interaction
Pages321-322
Number of pages2
DOIs
Publication statusPublished - 2010
Externally publishedYes
Event4th International Conference on Tangible, Embedded, and Embodied Interaction, TEI'10 - Cambridge, MA
Duration: 2010 Jan 252010 Jan 27

Other

Other4th International Conference on Tangible, Embedded, and Embodied Interaction, TEI'10
CityCambridge, MA
Period10/1/2510/1/27

Fingerprint

Nails
Skin
Textures
Cameras
Gesture recognition
Image processing

Keywords

  • Hand gesture
  • Interaction device
  • Robot
  • Tactile feedback

ASJC Scopus subject areas

  • Computer Graphics and Computer-Aided Design
  • Human-Computer Interaction
  • Software

Cite this

Tamaki, E. (2010). A robust and accurate 3D hand posture estimation method for interactive systems. In TEI'10 - Proceedings of the 4th International Conference on Tangible, Embedded, and Embodied Interaction (pp. 321-322) https://doi.org/10.1145/1709886.1709963

A robust and accurate 3D hand posture estimation method for interactive systems. / Tamaki, Emi.

TEI'10 - Proceedings of the 4th International Conference on Tangible, Embedded, and Embodied Interaction. 2010. p. 321-322.

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Tamaki, E 2010, A robust and accurate 3D hand posture estimation method for interactive systems. in TEI'10 - Proceedings of the 4th International Conference on Tangible, Embedded, and Embodied Interaction. pp. 321-322, 4th International Conference on Tangible, Embedded, and Embodied Interaction, TEI'10, Cambridge, MA, 10/1/25. https://doi.org/10.1145/1709886.1709963
Tamaki E. A robust and accurate 3D hand posture estimation method for interactive systems. In TEI'10 - Proceedings of the 4th International Conference on Tangible, Embedded, and Embodied Interaction. 2010. p. 321-322 https://doi.org/10.1145/1709886.1709963
Tamaki, Emi. / A robust and accurate 3D hand posture estimation method for interactive systems. TEI'10 - Proceedings of the 4th International Conference on Tangible, Embedded, and Embodied Interaction. 2010. pp. 321-322
@inproceedings{8eec7ea17ab142c19a09faf5a7e6892d,
title = "A robust and accurate 3D hand posture estimation method for interactive systems",
abstract = "In this paper, a new 3D hand posture estimation system using a single camera and 3 interactive systems are introduced. Existing hand gesture recognition systems estimate hand's 3D models based on image features such as contour or skin texture. However, it was difficult to estimate the wrist rotation because the contour and the texture data do not have enough information to distinguish hand's sides. To solve this problem, we propose a new 3D hand posture estimation system that uses data of nail positions. Nail positions are an important factor to recognize hand's sides. Using nail positions, it becomes possible to detect whether the camera is facing palm or dorsum. In addition, nail areas can be robustly extracted from a skin area by a simple image processing technique. Our Proposed system uses a database consists of data-sets of the hand's contour, the nail positions, and finger joint angles. To estimate the hand posture, the system first extracts the hand's contour and the nail positions from the captured image, and searches for a similar data-set from the database. The system then outputs the finger joint angles of the searched data-set. Our experimental results show high accuracy in the hand posture estimation with the wrist rotation.",
keywords = "Hand gesture, Interaction device, Robot, Tactile feedback",
author = "Emi Tamaki",
year = "2010",
doi = "10.1145/1709886.1709963",
language = "English",
isbn = "9781605588414",
pages = "321--322",
booktitle = "TEI'10 - Proceedings of the 4th International Conference on Tangible, Embedded, and Embodied Interaction",

}

TY - GEN

T1 - A robust and accurate 3D hand posture estimation method for interactive systems

AU - Tamaki, Emi

PY - 2010

Y1 - 2010

N2 - In this paper, a new 3D hand posture estimation system using a single camera and 3 interactive systems are introduced. Existing hand gesture recognition systems estimate hand's 3D models based on image features such as contour or skin texture. However, it was difficult to estimate the wrist rotation because the contour and the texture data do not have enough information to distinguish hand's sides. To solve this problem, we propose a new 3D hand posture estimation system that uses data of nail positions. Nail positions are an important factor to recognize hand's sides. Using nail positions, it becomes possible to detect whether the camera is facing palm or dorsum. In addition, nail areas can be robustly extracted from a skin area by a simple image processing technique. Our Proposed system uses a database consists of data-sets of the hand's contour, the nail positions, and finger joint angles. To estimate the hand posture, the system first extracts the hand's contour and the nail positions from the captured image, and searches for a similar data-set from the database. The system then outputs the finger joint angles of the searched data-set. Our experimental results show high accuracy in the hand posture estimation with the wrist rotation.

AB - In this paper, a new 3D hand posture estimation system using a single camera and 3 interactive systems are introduced. Existing hand gesture recognition systems estimate hand's 3D models based on image features such as contour or skin texture. However, it was difficult to estimate the wrist rotation because the contour and the texture data do not have enough information to distinguish hand's sides. To solve this problem, we propose a new 3D hand posture estimation system that uses data of nail positions. Nail positions are an important factor to recognize hand's sides. Using nail positions, it becomes possible to detect whether the camera is facing palm or dorsum. In addition, nail areas can be robustly extracted from a skin area by a simple image processing technique. Our Proposed system uses a database consists of data-sets of the hand's contour, the nail positions, and finger joint angles. To estimate the hand posture, the system first extracts the hand's contour and the nail positions from the captured image, and searches for a similar data-set from the database. The system then outputs the finger joint angles of the searched data-set. Our experimental results show high accuracy in the hand posture estimation with the wrist rotation.

KW - Hand gesture

KW - Interaction device

KW - Robot

KW - Tactile feedback

UR - http://www.scopus.com/inward/record.url?scp=77950806352&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=77950806352&partnerID=8YFLogxK

U2 - 10.1145/1709886.1709963

DO - 10.1145/1709886.1709963

M3 - Conference contribution

SN - 9781605588414

SP - 321

EP - 322

BT - TEI'10 - Proceedings of the 4th International Conference on Tangible, Embedded, and Embodied Interaction

ER -