Abstract
In order to realize a vision system for an autonomous mobile robot used in a human living environment, it is necessary to observe human behaviors and to react to those actions. In this paper, we propose a real-time human tracking method based on vision system for an autonomous mobile robot. First, the system detects body parts as moving areas in the scene, and a face region or region specific to human is extracted in the detected area using color. Next, the facial gesture and head gesture are recognized. We implement the vision system on a mobile robot, and experimentally show the system can detect and track a human and his face in real-time.
Original language | English |
---|---|
Title of host publication | Proceedings - IEEE International Workshop on Robot and Human Interactive Communication |
Pages | 442-449 |
Number of pages | 8 |
DOIs | |
Publication status | Published - 2001 |
Event | 10th IEEE International Workshop on Robot and Human Interactive Communication, ROMAN 2001 - Bordeaux and Paris Duration: 2001 Sep 18 → 2001 Sep 21 |
Other
Other | 10th IEEE International Workshop on Robot and Human Interactive Communication, ROMAN 2001 |
---|---|
City | Bordeaux and Paris |
Period | 01/9/18 → 01/9/21 |
Fingerprint
ASJC Scopus subject areas
- Software
- Artificial Intelligence
- Human-Computer Interaction
Cite this
Real-time vision system for autonomous mobile robot. / Doi, Masataka; Nakakita, Manabu; Aoki, Yoshimitsu; Hashimoto, Shuji.
Proceedings - IEEE International Workshop on Robot and Human Interactive Communication. 2001. p. 442-449 981944.Research output: Chapter in Book/Report/Conference proceeding › Conference contribution
}
TY - GEN
T1 - Real-time vision system for autonomous mobile robot
AU - Doi, Masataka
AU - Nakakita, Manabu
AU - Aoki, Yoshimitsu
AU - Hashimoto, Shuji
PY - 2001
Y1 - 2001
N2 - In order to realize a vision system for an autonomous mobile robot used in a human living environment, it is necessary to observe human behaviors and to react to those actions. In this paper, we propose a real-time human tracking method based on vision system for an autonomous mobile robot. First, the system detects body parts as moving areas in the scene, and a face region or region specific to human is extracted in the detected area using color. Next, the facial gesture and head gesture are recognized. We implement the vision system on a mobile robot, and experimentally show the system can detect and track a human and his face in real-time.
AB - In order to realize a vision system for an autonomous mobile robot used in a human living environment, it is necessary to observe human behaviors and to react to those actions. In this paper, we propose a real-time human tracking method based on vision system for an autonomous mobile robot. First, the system detects body parts as moving areas in the scene, and a face region or region specific to human is extracted in the detected area using color. Next, the facial gesture and head gesture are recognized. We implement the vision system on a mobile robot, and experimentally show the system can detect and track a human and his face in real-time.
UR - http://www.scopus.com/inward/record.url?scp=84881090669&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84881090669&partnerID=8YFLogxK
U2 - 10.1109/ROMAN.2001.981944
DO - 10.1109/ROMAN.2001.981944
M3 - Conference contribution
AN - SCOPUS:84881090669
SN - 0780372220
SN - 9780780372221
SP - 442
EP - 449
BT - Proceedings - IEEE International Workshop on Robot and Human Interactive Communication
ER -