We present a method that can automatically annotate when and who is appearing in a video stream that is shot in an unstaged condition. Previous face recognition methods were not robust against different shooting conditions, such as those with variable lighting, face directions, and other factors, in a video stream and had difficulties identifying a person and the scenes the person appears in. To overcome such difficulties, our method groups consecutive video frames (scenes) into clusters that each have the same person's face, which we call a "facial-temporal continuum," and identifies a person by using many video frames in each cluster. In our experiments, accuracy with our method was approximately two or three times higher than a previous method that recognizes a face in each frame.
|Journal||Kyokai Joho Imeji Zasshi/Journal of the Institute of Image Information and Television Engineers|
|Publication status||Published - 2012 Jul 16|
ASJC Scopus subject areas
- Media Technology
- Computer Science Applications
- Electrical and Electronic Engineering