Abstract
Multiple-instance learning (MIL) can solve supervised learning tasks, where only a bag of multiple instances is labeled, instead of a single instance. It is considerably important to develop effective and efficient MIL algorithms, because real-world datasets usually contain large instances. Known for its good generalization performance, MIL based on extreme learning machines (ELM-MIL) has proven to be more efficient than several typical MIL classification methods. ELM-MIL selects the most qualified instances from each bag through a single hidden layer feedforward network (SLFN) and trains modified ELM models to update the output weights. This learning approach often performs susceptible to the number of hidden nodes and can easily suffer from over-fitting problem. Using Bayesian inferences, this study introduces a Bayesian ELM (BELM)-based MIL algorithm (BELM-MIL) to address MIL classification problems. First, weight self-learning method based on a Bayesian network is applied to determine the weights of instance features. The most qualified instances are then selected from each bag to represent the bag. Second, BELM can improve the classification model via regularization of automatic estimations to reduce possible over-fitting during the calibration process. Experiments and comparisons are conducted with several competing algorithms on Musk datasets, images datasets, and inductive logic programming datasets. Superior classification accuracy and performance are demonstrated by BELM-MIL.
Highlights
Multiple-instance learning (MIL) was firstly proposed by Dietterich et al [1] as an approach to predict drug activities
Bayesian extreme learning machine (ELM) (BELM)–MIL is a better algorithm overall and needs a little more running time, which is attributed to the combination of weight update method and Bayesian classifier training used in BELM–MIL
Full Bayesian method of ELM has more advantages than BELM, for example, the variational approximation inference is employed in the Bayesian model to compute the posterior distribution and the independent variational hyperparameters approximately, which can be used to select the hidden nodes automatically, and it can achieve more stable performance with more compact architectures [57]
Summary
Multiple-instance learning (MIL) was firstly proposed by Dietterich et al [1] as an approach to predict drug activities. Afterwards, MIL enjoyed many successful applications, such as improved drug activity prediction [2], text categorization [3], image classification [4], object detection [5], and stock prediction [6]. Different from other machine-learning frameworks, MIL is novel, because it contains bags with labels, instead of labeled instances. If there is at least one positive instance in the bag, the label is positive.
Published Version (Free)
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have