Sciweavers

CORR
2010
Springer

Learning Multi-modal Similarity

13 years 9 months ago
Learning Multi-modal Similarity
In many applications involving multi-media data, the definition of similarity between items is integral to several key tasks, including nearest-neighbor retrieval, classification, and recommendation. Data in such regimes typically exhibits multiple modalities, such as acoustic and visual content of video. Integrating such heterogeneous data to form a holistic similarity space is therefore a key challenge to be overcome in many real-world applications. We present a novel multiple kernel learning technique for integrating heterogeneous data into a single, unified similarity space. Our algorithm learns an optimal ensemble of kernel transformations which conform to measurements of human perceptual similarity, as expressed by relative comparisons. To cope with the ubiquitous problems of subjectivity and inconsistency in multimedia similarity, we develop graph-based techniques to filter similarity measurements, resulting in a simplified and robust training procedure.
Brian McFee, Gert R. G. Lanckriet
Added 01 Mar 2011
Updated 01 Mar 2011
Type Journal
Year 2010
Where CORR
Authors Brian McFee, Gert R. G. Lanckriet
Comments (0)