logo
banner

Journals & Publications

Journals Publications Papers

Papers

Iteratively Divide-and-Conquer Learning for Nonlinear Classification and Ranking
Mar 19, 2018Author:
PrintText Size A A

Title: Iteratively Divide-and-Conquer Learning for Nonlinear Classification and Ranking

 Authors: Wu, O; Mao, X; Hu, WM

 Author Full Names: Wu, Ou; Mao, Xue; Hu, Weiming

 Source: ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 9 (2):10.1145/3122802 JAN 2018

 Language: English

 Abstract: Nonlinear classifiers (i.e., kernel support vector machines (SVMs)) are effective for nonlinear data classification. However, nonlinear classifiers are usually prohibitively expensive when dealing with large nonlinear data. Ensembles of linear classifiers have been proposed to address this inefficiency, which is called the ensemble linear classifiers for nonlinear data problem. In this article, a new iterative learning approach is introduced that involves two steps at each iteration: partitioning the data into clusters according to Gaussian mixture models with local consistency and then training basic classifiers (i.e., linear SVMs) for each cluster. The two divide-and-conquer steps are combined into a graphical model. Meanwhile, with training, each classifier is regarded as a task; clustered multitask learning is employed to capture the relatedness among different tasks and avoid overfitting in each task. In addition, two novel extensions are introduced based on the proposed approach. First, the approach is extended for quality-aware web data classification. In this problem, the types of web data vary in terms of information quality. The ignorance of the variations of information quality of web data leads to poor classification models. The proposed approach can effectively integrate quality-aware factors into web data classification. Second, the approach is extended for listwise learning to rank to construct an ensemble of linear ranking models, whereas most existing listwise ranking methods construct a solely linear ranking model. Experimental results on benchmark datasets show that our approach outperforms state-of-the-art algorithms. During prediction for nonlinear classification, it also obtains comparable classification performance to kernel SVMs, with much higher efficiency.

 ISSN: 2157-6904

 eISSN: 2157-6912

 Article Number: 18

 IDS Number: FX0FU

 Unique ID: WOS:000425717900005

*Click Here to View Full Record