SUMMARYThe rapid development of information techniques has lead to more and more high-dimensional datasets, making classification more difficult. However, not all of the features are useful for classification, and some of these features may even cause low classification accuracy. Feature selection is a useful technique, which aims to reduce the dimensionality of datasets, for solving classification problems. In this paper, we propose a modified bat algorithm (BA) for feature selection, called MBAFS, using a SVM. Some mechanisms are designed for avoiding the premature convergence. On the one hand, in order to maintain the diversity of bats, they are guided by the combination of a random bat and the global best bat. On the other hand, to enhance the ability of escaping from local optimization, MBAFS employs one mutation mechanism while the algorithm trapped into local optima. Furthermore, the performance of MBAFS was tested on twelve benchmark datasets, and was compared with other BA based algorithms and some well-known BPSO based algorithms. Experimental results indicated that the proposed algorithm outperforms than other methods. Also, the comparison details showed that MBAFS is competitive in terms of computational time. key words: feature selection, wrapper model, bat algorithm, premature convergence, SVM
Transfer boosting, a branch of instance-based transfer learning, is a commonly adopted transfer learning method. However, currently popular transfer boosting methods focus on binary classification problems even though there are many multi-classification tasks in practice. In this paper, we developed a new algorithm called MultiTransferBoost on the basis of TransferBoost for multi-classification. MultiTransferBoost firstly separated the multi-classification problem into several orthogonal binary classification problems. During each iteration, MultiTransferBoost boosted weighted instances from different source domains while each instance's weight was assigned and updated by evaluating the difficulty of the instance being correctly classified and the "transferability" of the instance's corresponding source domain to the target. The updating process repeated until it reached the predefined training error or iteration number. The weight update factors, which were analyzed and adjusted to minimize the Hamming loss of the output coding, strengthened the connections among the sub binary problems during each iteration. Experimental results demonstrated that MultiTransferBoost had better classification performance and less computational burden than existing instance-based algorithms using the One-Against-One (OAO) strategy.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.