Abstract
We investigate dimensionality reduction methods from the perspective of their ability to produce a low-rank customer-product matrix representation. We analyze the results of using collaborative filtering based on SVD, RI, Reflective Random Indexing (RRI) and Randomized Singular Value Decomposition (RSVD) from the perspective of selected algebraic (i.e. application-independent) properties. We show that the Frobenius-norm optimality of SVD does not correspond to the optimal recommendation accuracy, when measured in terms of F1. On the other hand, a high collaborative filtering quality is achievable when a matrix decomposition - based on a combination of RRI and SVD referred to as RSVD-RRI - leads to increased diversity of low-dimensional eigenvectors. The diversity is observable from the perspective of cosine similarities analyzed in comparison to the analogical case of SVD. Such a feature is more desirable than the fidelity of the input matrix spectrum representation, despite the MSE-optimality of SVD.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: International Journal of Machine Learning and Computing
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.