Lip-motion events analysis and lip segmentation using optical flow
2012 (English)Conference paper (Refereed)
We propose an algorithm for detecting the mouth events of opening and closing. Our method is translation and ro- tation invariant, works at very fast speeds, and does not re- quire segmented lips. The approach is based on a recently developed optical flow algorithm that handles the motion of linear structure in a stable and consistent way.Furthermore, we provide a semi-automatic tool for gen- erating groundtruth segmentation of video data, also based on the optical flow algorithm used for tracking keypoints at faster than 200 frames/second. We provide groundtruth for 50 sessions of speech of the XM2VTS database  avail- able for download, and the means to segment further ses- sions at a relatively small amount of user interaction.We use the generated groundtruth to test the proposed al- gorithm for detecting events, and show it to yield promising result. The semi-automatic tool will be a useful resource for researchers in need of groundtruth segmentation from video for the XM2VTS database and others.
Place, publisher, year, edition, pages
Piscataway, N.J.: IEEE Press, 2012. 138-145 p.
, IEEE Computer Society Conference on Computer Vision and Pattern Recognition workshops. Proceedings, ISSN 2160-7508
Keypoints, Linear structures, Lip segmentation, Optical flow algorithm, Rotation invariant, Semi-automatic tools, User interaction, Video data
IdentifiersURN: urn:nbn:se:hh:diva-19645DOI: 10.1109/CVPRW.2012.6239228ScopusID: 2-s2.0-84864974582ISBN: 978-1-4673-1612-5ISBN: 978-146731611-8OAI: oai:DiVA.org:hh-19645DiVA: diva2:552889
2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2012, 16-21 June, 2012, Rhode Island, USA
FunderSwedish Research Council
©2012 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.2012-10-042012-09-172014-11-18Bibliographically approved