I. Amara, E. Granger and A. Hadid, "Contextual Weighting of Patches for Local Matching in Still-to-Video Face Recognition," 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi’an, 2018, pp. 756-763. doi: 10.1109/FG.2018.00119
Contextual weighting of patches for local matching in still-to-video face recognition
|Author:||Amara, Ibtihel1; Granger, Eric1; Hadid, Abdenour2|
1Laboratory for Imagery, Vision and Artificial Intelligence Ecole de technologie supérieure, Université de Québec, Canada
2Center for Machine Vision and Signal Analysis (CMVS), University of Oulu, Finland
|Online Access:||PDF Full Text (PDF, 0.5 MB)|
|Persistent link:|| http://urn.fi/urn:nbn:fi-fe202003248952
Institute of Electrical and Electronics Engineers,
|Publish Date:|| 2020-03-24
Still-to-video face recognition (FR) systems for watchlist screening seek to recognize individuals of interest given faces captured over a network of video surveillance cameras. Screening faces against a watchlist is a challenging application because only a limited number of reference stills is available per individual during enrollment, and the appearance of face captures in videos changes from camera to camera, due to variations in illumination, pose, blur, scale, expression and occlusion. In order to improve the robustness of FR systems, several local matching techniques have been proposed that rely on static or dynamic weighting of patches. However, these approaches are not suitable for watchlist screening applications where the capturing conditions vary significantly over different camera fields of view (FoV). In this paper, a new dynamic weighting technique is proposed for weighting facial patches based on video data collected a priori from the specific operational domain (camera FoV) and on image quality assessment. Results obtained on videos from the Chokepoint dataset indicate that the proposed approach can significantly outperform the reference local matching methods because patch weights tend to grow for discriminant facial regions.
|Pages:||756 - 763|
13th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2018
IEEE International Conference on Automatic Face and Gesture Recognition
|Type of Publication:||
A4 Article in conference proceedings
|Field of Science:||
213 Electronic, automation and communications engineering, electronics
The financial support from the Academy of Finland is fully acknowledged.
© 2018 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.