Abstract

Canonical Correlation Analysis (CCA) is a linear representation learning method that seeks maximally correlated variables in multi-view data. Nonlinear CCA extends this notion to a broader family of transformations, which are more powerful in many real-world applications. Given the joint probability, the Alternating Conditional Expectation (ACE) algorithm provides an optimal solution to the nonlinear CCA problem. However, it suffers from limited performance and an increasing computational burden when only a finite number of samples is available. In this work, we introduce an information-theoretic compressed representation framework for the nonlinear CCA problem (CRCCA), which extends the classical ACE approach. Our suggested framework seeks compact representations of the data that allow a maximal level of correlation. This way, we control the trade-off between the flexibility and the complexity of the model. CRCCA provides theoretical bounds and optimality conditions, as we establish fundamental connections to rate-distortion theory, the information bottleneck and remote source coding. In addition, it allows a soft dimensionality reduction, as the compression level is determined by the mutual information between the original noisy data and the extracted signals. Finally, we introduce a simple implementation of the CRCCA framework, based on lattice quantization.

Highlights

  • IntroductionCanonical correlation analysis (CCA) seeks linear projections of two given random vectors so that the extracted (possibly lower dimensional) variables are maximally correlated [1]

  • Canonical correlation analysis (CCA) seeks linear projections of two given random vectors so that the extracted variables are maximally correlated [1]

  • Inspired by Breiman and Friedman [8], we suggest an iterative approach for the CRCCA problem

Read more

Summary

Introduction

Canonical correlation analysis (CCA) seeks linear projections of two given random vectors so that the extracted (possibly lower dimensional) variables are maximally correlated [1]. CCA is a powerful tool in the analysis of paired data ( X, Y ), where X and Y are two different representations of the same set of objects. It is commonly used in a variety of applications, such as speech recognition [2], natural language processing [3], cross-modal retrieval [4], multimodal signal processing [5], computer vision [6], and many others. One of its major drawbacks is its restriction to linear projections, whereas many real-world setups exhibit highly nonlinear relationships To overcome this limitation, several nonlinear CCA extensions have been proposed. Van Der Burg and De Leeuw [7]

Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.