Pareto-Path Multitask Multiple Kernel Learning

Keywords

Machine learning; optimization methods; pattern recognition; supervised learning; support vector machines (SVM).

Abstract

A traditional and intuitively appealing Multitask Multiple Kernel Learning (MT-MKL) method is to optimize the sum (thus, the average) of objective functions with (partially) shared kernel function, which allows information sharing among the tasks. We point out that the obtained solution corresponds to a single point on the Pareto Front (PF) of a multiobjective optimization problem, which considers the concurrent optimization of all task objectives involved in the Multitask Learning (MTL) problem. Motivated by this last observation and arguing that the former approach is heuristic, we propose a novel support vector machine MT-MKL framework that considers an implicitly defined set of conic combinations of task objectives. We show that solving our framework produces solutions along a path on the aforementioned PF and that it subsumes the optimization of the average of objective functions as a special case. Using the algorithms we derived, we demonstrate through a series of experimental results that the framework is capable of achieving a better classification performance, when compared with other similar MTL approaches.

Publication Date

1-1-2015

Publication Title

IEEE Transactions on Neural Networks and Learning Systems

Volume

26

Issue

1

Number of Pages

51-61

Document Type

Article

Personal Identifier

scopus

DOI Link

https://doi.org/10.1109/TNNLS.2014.2309939

Socpus ID

84919617362 (Scopus)

Source API URL

https://api.elsevier.com/content/abstract/scopus_id/84919617362

This document is currently not available here.

Share

COinS