A Multi-View Hand Gesture RGB-D Dataset for Human-Robot Interaction Scenarios


Shukla D., ERKENT Ö., Piater J.

25th IEEE International Symposium on Robot and Human Interactive Communication (IEEE RO-MAN), New York, Amerika Birleşik Devletleri, 26 - 31 Ağustos 2016, ss.1084-1091 identifier identifier

  • Yayın Türü: Bildiri / Tam Metin Bildiri
  • Cilt numarası:
  • Doi Numarası: 10.1109/roman.2016.7745243
  • Basıldığı Şehir: New York
  • Basıldığı Ülke: Amerika Birleşik Devletleri
  • Sayfa Sayıları: ss.1084-1091
  • Hacettepe Üniversitesi Adresli: Hayır

Özet

Understanding semantic meaning from hand gestures is a challenging but essential task in human-robot interaction scenarios. In this paper we present a baseline evaluation of the Innsbruck Multi-View Hand Gesture (IMHG) dataset [1] recorded with two RGB-D cameras (Kinect). As a baseline, we adopt a probabilistic appearance-based framework [2] to detect a hand gesture and estimate its pose using two cameras. The dataset consists of two types of deictic gestures with the ground truth location of the target, two symbolic gestures, two manipulative gestures, and two interactional gestures. We discuss the effect of parallax due to the offset between head and hand while performing deictic gestures. Furthermore, we evaluate the proposed framework to estimate the potential referents on the Innsbruck Pointing at Objects (IPO) dataset [2].