Abstract

Generally, fuzzy models, especially rule-based models, are designed in a monolithic manner, meaning that all data are used <italic xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">en bloc</i> to design the model. At the same time, there is a visible need to cope with the ever-increasing volumes of data (both in terms of the number of data and their dimensionality) as well as being faced with distributed data located at various locations. The objective of this study is to develop a concept and provide a design framework as well as assess its performance for constructing a collection of rule-based models on a basis of a randomly sampled repository of data and then realize their aggregation. More specifically, for the sampled data, the design of each model is carried out in a standard way as commonly encountered in the case of Takagi-Sugeno (TS) rule-based models and next augmented by gradient boosting. The aggregation is realized by optimizing a weighting scheme applied to the results of the individual models. Our intent is also to carefully demonstrate the performance offered by the mechanisms of machine learning applied in the setting of rule-based models, which is an original task completed before. A number of high-dimensional data are used in the experimental studies to complete a thorough assessment. A comparative performance analysis is reported with respect to the monolithically developed TS models.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call