Abstract

Video event carries very rich and complex semantics, and video event analysis is intrinsically multi-view learning problem. In this paper, by revealing the relationship between canonical correlation analysis (CCA) and linear discriminant analysis (LDA), we propose a new convenient multi-view learning architecture, i.e., multi-view discriminative fusion on CCA (MvDF-CCA). Instead of pair-wise relation consideration among different views, MvDF-CCA separately aligns each view space to the target one comprised by labeled indicators, which is easy expanding to additional view data. MvDF-CCA leverages the discriminative ability in LDA and the strengths in CCA, and maps all view spaces to the same target labeled space, which can be concurrent execution on large scale video categorization. Competitive results are reported on the well-known UCF101 and Columbia Consumer Video (CCV) benchmarks.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call