Abstract

AbstractThere is a huge discrepancy in how researchers implement, evaluate, and report the performance of a machine learning method for classification or segmentation of biomedical data. Poor reporting and inadequate inferences are, however, not unusual to see in current literature. More specifically, vague aims and scope, missing details about the data, ambiguous preprocessing procedures, lack of clarity regarding the method's implementation, poor validation and testing, invalid comparisons between methods, and the absence of a clear rationale for performance metric choices are making it difficult to draw the right conclusions from many studies in the field. This report suggests 10 guidelines and principles that should be followed when reporting the implementation of a method and the evaluation of its performance in order to make the study transparent, interpretable, replicable, and useful. All stages of data processing and method's performance evaluation should be clearly described, and parameters and metric choices must be justified in order to aid readers in appreciating the performance of the method or in comparing it with other relevant methods. We feel that these guidelines are important for clear scientific communication in the field of biomedical data processing.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call