Abstract

Artificial Neural Networks (ANNs) are weighted directed graphs of interconnected neurons widely employed to model complex problems. However, the selection of the optimal ANN architecture and its training parameters is not enough to obtain reliable models. The data preprocessing stage is fundamental to improve the model’s performance. Specifically, Feature Normalisation (FN) is commonly utilised to remove the features’ magnitude aiming at equalising the features’ contribution to the model training. Nevertheless, this work demonstrates that the FN method selection affects the model performance. Also, it is well-known that ANNs are commonly considered a “black box” due to their lack of interpretability. In this sense, several works aim to analyse the features’ contribution to the network for estimating the output. However, these methods, specifically those based on network’s weights, like Garson’s or Yoon’s methods, do not consider preprocessing factors, such as dispersion factors , previously employed to transform the input data. This work proposes a new features’ relevance analysis method that includes the dispersion factors into the weight matrix analysis methods to infer each feature’s actual contribution to the network output more precisely. Besides, in this work, the Proportional Dispersion Weights (PWD) are proposed as explanatory factors of similarity between models’ performance results. The conclusions from this work improve the understanding of the features’ contribution to the model that enhances the feature selection strategy, which is fundamental for reliably modelling a given problem.

Highlights

  • Artificial Neural Networks (ANNs) are algorithms that simulate the human brain learning behaviour, modelled by a weighted directed graph of interconnected nodes or neurons

  • As stated and demonstrated in this work, the Feature Normalisation (FN) method selection significantly affects the ANN-based model’s performance and the inclusion of dispersion factors when estimating the features’ contribution improves the understanding of the features’ influence on the model. The former point emphasises the influence of the FN method selection; it remains open the question about which FN to employ to transform a given dataset in order to reach the best model’s performance; or even if it is recommendable the application of FN or discard the magnitude of the features by removing the 10nj factors from (4)

  • A weight that does not correspond to the real relative importance of a given feature can result in a performance loss

Read more

Summary

Introduction

Artificial Neural Networks (ANNs) are algorithms that simulate the human brain learning behaviour, modelled by a weighted directed graph of interconnected nodes or neurons. These neurons are simple functions whose arguments are the weighted summation of the inputs to the node [1]. Due to their ability to solve challenging computational problems [2], [3], ANNs are widely applied in different fields, like industry among others [4]–[8].

Objectives
Methods
Findings
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call