Abstract

The POWER (PrObabilistic Weighted Ensemble of Roles) model[1] is a Bayesian mixture model where a single data point is generated by multiple components with varying degrees of influence. This is unlike standard mixture models where each data point is generated by a single mixture component. The POWER model allows for capturing various hidden and complexly overlapping components and roles of entities that contribute to generating the data points. However, the POWER model suffers from a very slow learning time. The highest complexity time of the parameters’ learning steps is \(O\left(n \cdot k \cdot d^{2}\right)\), with n being the number of data points, k the number of components of the model and d the number of data attributes.In this paper, we propose an approximation to the POWER model weight parameter learning algorithm that reduces the computational time significantly. The overall complexity of the new algorithm is \(O\left(n \cdot k \cdot d \cdot p\right)\) in practice (where p is an approximation parameter much smaller than d). This allows the model learning time to be linear in the number of attributes of the data and provides a significant speedup over the original algorithm.We demonstrate the accuracy of our approximation technique using synthetic and real datasets. We also provide experimental results of the approximate POWER model on a dataset of NIPS papers and a dataset of wireless web access patterns and show that the model learnt are similar. An implementation of the approximate POWER model on the dataset of NIPS papers is about 27 times faster than the original version.KeywordsMixture ModelData MiningApproximation

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.