Background: Telemarketing is an effective marketing strategy lately, because it allows long-distance interaction making it easier for marketing promotion management to market their products. But sometimes with incessant phone calls to clients that are less potential to cause inconvenience, so we need predictions that produce good probabilities so that it can be the basis for making decisions about how many potential clients can be contacted which results in time and costs can be minimized, telephone calls can be more effective, client stress and intrusion will be reduced. strong. Method: This study will compare the classification performance of Bank Marketing datasets from the UCI Machine Learning Repository using data mining with the Adaboost and Bagging ensemble approach, base algorithm using J48 Weka, and Wrapper subset evaluation feature selection techniques and previously data balancing was performed on the dataset, where the expected results can be known the best ensemble method that produces the best performance of both. Results: In the Bagging experiment, the best performance of Adaboost and J48 with an accuracy rate of 86.6%, Adaboost 83.5% and J48 of 85.9%Conclusion: The conclusion obtained from this study that the use of data balancing and feature selection techniques can help improve classification performance, Bagging is the best ensemble algorithm from this study, while for Adaboost is not productive for this study because the basic algorithm used is a strong learner where Adaboost has Weaknesses to improve strong basic algorithm.
[1]
Ron Kohavi,et al.
Supervised and Unsupervised Discretization of Continuous Features
,
1995,
ICML.
[2]
Li Zhu,et al.
Data Mining on Imbalanced Data Sets
,
2008,
2008 International Conference on Advanced Computer Theory and Engineering.
[3]
Sotiris B. Kotsiantis,et al.
Bagging and boosting variants for handling classifications problems: a survey
,
2013,
The Knowledge Engineering Review.
[4]
Isabelle Guyon,et al.
An Introduction to Variable and Feature Selection
,
2003,
J. Mach. Learn. Res..
[5]
Yoav Freund,et al.
A Short Introduction to Boosting
,
1999
.
[6]
Huan Liu,et al.
Feature Selection for Classification
,
1997,
Intell. Data Anal..
[7]
Michael Stonebraker,et al.
The Morgan Kaufmann Series in Data Management Systems
,
1999
.
[8]
Robert C. Holte,et al.
Severe Class Imbalance: Why Better Algorithms Aren't the Answer
,
2005,
ECML.
[9]
Bernard F. Buxton,et al.
Performance Degradation in Boosting
,
2001,
Multiple Classifier Systems.