Semantic Multi-modality Fusion for Video Search
Keywords:
multimodality, relevance score, concept detectionAbstract
We collect information from different sources in different forms. Multimodality fusion can be solution for various information retrieval problems. In this paper, we propose multimodality fusion approach for video search, where search modalities are derived from a set of knowledge sources, such as text, images and videos. We break down the query modality relationship into two components that are much easier to calculate: the relationship between the query and the concept and the relevance of the concept of modality. The first can be estimated effectively online using visual mapping techniques, while the seconds can be calculated off-line based on the concept-detection precision of each modality.
References
- Wei, X. Y., Jiang, Y. G., & Ngo, C. W. (2011). Concept-driven multi-modality fusion for video search. Circuits and Systems for Video Technology, IEEE Transactions on, 21(1), 62-73.
- Hu, W., Xie, N., Li, L., Zeng, X. Maybank, (2011). A survey on visual content-based video indexing and retrieval. Systems, Man, and Cybernetics, Part C: Applications and Reviews,IEEE Transactions on, 41(6), 797-819.
- M. Naphade, J. R. Smith, J. Tesi c, S.-F.Chang, W. Hsu, L. Kennedy, A. Hauptmann, and J. Curtis, “Large-scale concept ontology for mul- timedia,” IEEE Trans. Multimedia, vol. 13, no. 3, pp. 86–91, Jul.–Sep. 2006.
- Kennedy, L.,Chang, S. F., &Natsev, A. (2008). Query-adaptive fusion for multimodal search. Proceedings of the IEEE, 96(4), 567-588.
Downloads
Published
Issue
Section
License
Copyright (c) IJSRCSEIT

This work is licensed under a Creative Commons Attribution 4.0 International License.