Transfer learning is aimed at supporting the design of machine learning models in the target domain Dt, given that the knowledge (model) has already been constructed in the source domain Ds. The domains Dt and Ds (as well as the corresponding tasks Ts and Tt) are similar, yet not identical. As a result, the model transferred from Ds to Dt in this new environment exhibits its relevance (credibility) only to some limited extent. In this study, we develop an original approach, where we advocate that the knowledge transfer (model transfer) gives rise to a granular model where the level of information granularity associated with the produced results quantifies the relevance (quality or credibility) of the transferred model. In other words, we stress that the quality of knowledge transferred to Dt becomes captured through a granular generalization of the original numeric model. The overall systematic design process is elaborated on by focusing on the development process of granular neural networks carried out on a basis of the numeric neural networks coming from Ds. The key aspect of the design is to elevate the existing numeric neural network to its granular counterpart by admitting that the connections of the developed model come in the form of information granules, in particular intervals and fuzzy sets. The optimization process is guided by adjusting (optimizing) the level of information granularity being regarded as an essential design asset. The optimized performance index builds upon the descriptors of information granules commonly encountered in Granular Computing. In particular, coverage and specificity measures are treated as sound performance indicators of the quality of knowledge transfer (viz. the performance of the granular neural network expressed in the target domain). Several illustrative examples are provided to visualize the performance of the established design environment.
Read full abstract