Identifying scenes with the same person in video content on the basis of scene continuity and face similarity measurement

Tatsunori Hirai, Tomoyasu Nakano, Masataka Goto, Shigeo Morishima

    Research output: Contribution to journalArticle

    2 Citations (Scopus)

    Abstract

    We present a method that can automatically annotate when and who is appearing in a video stream that is shot in an unstaged condition. Previous face recognition methods were not robust against different shooting conditions, such as those with variable lighting, face directions, and other factors, in a video stream and had difficulties identifying a person and the scenes the person appears in. To overcome such difficulties, our method groups consecutive video frames (scenes) into clusters that each have the same person's face, which we call a "facial-temporal continuum," and identifies a person by using many video frames in each cluster. In our experiments, accuracy with our method was approximately two or three times higher than a previous method that recognizes a face in each frame.

    Original languageEnglish
    JournalKyokai Joho Imeji Zasshi/Journal of the Institute of Image Information and Television Engineers
    Volume66
    Issue number7
    DOIs
    Publication statusPublished - 2012

    Fingerprint

    Face recognition
    Lighting
    Experiments

    ASJC Scopus subject areas

    • Electrical and Electronic Engineering
    • Media Technology
    • Computer Science Applications

    Cite this

    @article{e4ac491b45c4410ba0169c57c852caae,
    title = "Identifying scenes with the same person in video content on the basis of scene continuity and face similarity measurement",
    abstract = "We present a method that can automatically annotate when and who is appearing in a video stream that is shot in an unstaged condition. Previous face recognition methods were not robust against different shooting conditions, such as those with variable lighting, face directions, and other factors, in a video stream and had difficulties identifying a person and the scenes the person appears in. To overcome such difficulties, our method groups consecutive video frames (scenes) into clusters that each have the same person's face, which we call a {"}facial-temporal continuum,{"} and identifies a person by using many video frames in each cluster. In our experiments, accuracy with our method was approximately two or three times higher than a previous method that recognizes a face in each frame.",
    author = "Tatsunori Hirai and Tomoyasu Nakano and Masataka Goto and Shigeo Morishima",
    year = "2012",
    doi = "10.3169/itej.66.J251",
    language = "English",
    volume = "66",
    journal = "Kyokai Joho Imeji Zasshi/Journal of the Institute of Image Information and Television Engineers",
    issn = "1342-6907",
    publisher = "Institute of Image Information and Television Engineers",
    number = "7",

    }

    TY - JOUR

    T1 - Identifying scenes with the same person in video content on the basis of scene continuity and face similarity measurement

    AU - Hirai, Tatsunori

    AU - Nakano, Tomoyasu

    AU - Goto, Masataka

    AU - Morishima, Shigeo

    PY - 2012

    Y1 - 2012

    N2 - We present a method that can automatically annotate when and who is appearing in a video stream that is shot in an unstaged condition. Previous face recognition methods were not robust against different shooting conditions, such as those with variable lighting, face directions, and other factors, in a video stream and had difficulties identifying a person and the scenes the person appears in. To overcome such difficulties, our method groups consecutive video frames (scenes) into clusters that each have the same person's face, which we call a "facial-temporal continuum," and identifies a person by using many video frames in each cluster. In our experiments, accuracy with our method was approximately two or three times higher than a previous method that recognizes a face in each frame.

    AB - We present a method that can automatically annotate when and who is appearing in a video stream that is shot in an unstaged condition. Previous face recognition methods were not robust against different shooting conditions, such as those with variable lighting, face directions, and other factors, in a video stream and had difficulties identifying a person and the scenes the person appears in. To overcome such difficulties, our method groups consecutive video frames (scenes) into clusters that each have the same person's face, which we call a "facial-temporal continuum," and identifies a person by using many video frames in each cluster. In our experiments, accuracy with our method was approximately two or three times higher than a previous method that recognizes a face in each frame.

    UR - http://www.scopus.com/inward/record.url?scp=84863662298&partnerID=8YFLogxK

    UR - http://www.scopus.com/inward/citedby.url?scp=84863662298&partnerID=8YFLogxK

    U2 - 10.3169/itej.66.J251

    DO - 10.3169/itej.66.J251

    M3 - Article

    VL - 66

    JO - Kyokai Joho Imeji Zasshi/Journal of the Institute of Image Information and Television Engineers

    JF - Kyokai Joho Imeji Zasshi/Journal of the Institute of Image Information and Television Engineers

    SN - 1342-6907

    IS - 7

    ER -