Abstract
The aim of classification in machine learning is to utilize knowledge gained from applying learning algorithms on a given data so as determine what class an unlabelled data having same pattern belongs to. However, algorithms do not learn properly when a massive difference in size between data classes exist. This classification problem exists in many real world application domains and has been a popular area of focus by machine learning and data mining researchers. The class imbalance problem is further made complex with the presence of associative data difficult factors. The duo have proven to greatly deteriorate classification performance. This paper introduces a two-phased data level approach for binary classes which entails the temporary re-labelling of classes. The proposed approach takes advantage of the local neighbourhood of the minority instances to identify and treat difficult examples belonging to both classes. Its outcome was satisfactory when compared against various data-level methods using datasets extracted from KEEL and UCI datasets repository.
Original language | English |
---|---|
Title of host publication | Proceedings of the 2016 International Joint Conference on Nerual Networks (IJCNN) |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
ISBN (Electronic) | 978-1509006205 |
ISBN (Print) | 978-1509006212 |
DOIs | |
Publication status | Published - 3 Nov 2016 |
Event | 2016 IEEE World Congress on Computational Intelligence - Vancouver, Canada Duration: 25 Jul 2016 → 29 Jul 2016 |
Conference
Conference | 2016 IEEE World Congress on Computational Intelligence |
---|---|
Abbreviated title | IEEE WCCI |
Country/Territory | Canada |
City | Vancouver |
Period | 25/07/16 → 29/07/16 |