Abstract
In this study, we focus on non-parametric probabilistic modeling for general regression analysis with large amounts of data and present an algorithm called the robust sparse Bayesian broad learning system. Robust sparse Bayesian learning is employed to infer the posterior distribution of the sparse connecting weight parameters in broad learning system. Regardless of the number of candidate features, our algorithm can always produce a compact subset of hidden-layer neurons of almost the same size learned from the data, which allows the algorithm to automatically adjust the model complexity of the network. This algorithm not only solves the regression problem of large amounts of data robustly but also possesses high computational efficiency and low requirements for computing hardware. Moreover, as a Bayesian probabilistic algorithm, it can provide the posterior uncertainty quantification of the predicted output, giving a measure of prediction confidence. The proposed algorithm is verified using simulated data generated by a benchmark function and also applied in non-parametric probabilistic modeling using high-speed railway track monitoring data. The results show that compared with several existing neural network algorithms, our proposed algorithm has strong model robustness, excellent prediction accuracy, and computational efficiency for regression analysis with large amounts of data, and has the potential to be widely used in general regression problems in science and engineering.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.