Abstract
We describe QCompere consortium submissions to the REPERE 2013 evaluation campaign. The REPERE challenge aims at gathering four communities (face recognition, speaker identification, optical character recognition and named entity detection) towards the same goal: multimodal person recognition in TV broadcast. First, four mono-modal components are introduced (one for each foregoing community) constituting the elementary building blocks of our various submissions. Then, depending on the target modality (speaker or face recognition) and on the task (supervised or unsupervised recognition), four different fusion techniques are introduced: they can be summarized as propagation-, classifier-, rule- or graph-based approaches. Finally, their performance is evaluated on REPERE 2013 test set and their advantages and limitations are discussed.
Original language | English |
---|---|
Pages (from-to) | 49-54 |
Number of pages | 6 |
Journal | CEUR Workshop Proceedings |
Volume | 1012 |
Publication status | Published - 2013 |
Externally published | Yes |
Event | 1st Workshop on Speech, Language and Audio in Multimedia, SLAM 2013 - Marseille, France Duration: 22 Aug 2013 → 23 Aug 2013 |
Bibliographical note
Publisher Copyright:Copyright © 2013 for the individual papers by the papers' authors.
Keywords
- Face recognition
- Multimodal fusion
- Named entity detection
- Speaker identification
- Video optical character recognition