Abstract

In this digital era, machine learning is a technology that is in demand by organizations and individuals. In the age of data and digital information, the ability to process data efficiently is needed. As the amount of data grows, there are various problems in machine learning. One of them is that with the increasing amount of data, class imbalance is also often found. Class imbalance is a condition where a class dominates another class, in one example case is when the positive value class has less number than the negative class. The class that is less in number is categorized as the minority class, while the class that dominates the dataset is called the majority class. Class imbalance can affect classification performance in a bad way, so handling imbalanced classes is needed to improve classification results. Classification of imbalanced data using Random Forest has satisfactory results, as well as by implementing SMOTE and ADASYN as sampling methods because they are highly popular and easy to implement. The best model produced in this study is the model that applies SMOTE oversampling on a dataset with 10% IR with a balanced accuracy of 98.75%, and the best result when applying ADASYN oversampling is on a dataset with 13% IR and a balanced accuracy of 99.03%.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call