An Extended Level Method for Multiple Kernel Learning Zenglin Xu

advertisement
An Extended Level Method for Multiple Kernel Learning
Zenglin
The Chinese University of Hong Kong
1
Xu ,
Rong
2
Jin ,
Irwin
1
King ,
1 {zlxu, king, lyu}@cse.cuhk.edu.hk
Department of Computer Science and Engineering
The Chinese University of Hong Kong
Shatin, N.T., Hong Kong
1. Multiple Kernel Learning (MKL)
Given a list of base kernel functions/matrices Ki , i = 1, . . . ,m, MKL
searches for a linear combination of the base kernel functions that
maximizes a generalized performance measure.
3. Level Method for MKL
3.1 Motivation
Michael R.
1
Lyu
2 rongjin@cse.msu.edu
Department of Computer Science and Engineering
Michigan State University
East Lansing, MI, 48824
Michigan State University
4. Experiments
4.1 Settings
Kernel combination parameters
4.2 Results
Kernel weights evolution
SVM parameters
Properties
SILP / breast
3.2 Algorithms
Question?
How to efficiently solve the optimization problem?
2. Optimization Method
Small or medium scale:
Time-saving-ratio
Properties:
Evolution of objective values
 Semi-definite Programming (SDP) [Lanckriet et al., 2004]
 Second Order Cone Programming (SOCP) [Bach et al., 2004]
SD / breast
Breast data set
Large scale:
 Semi-Infinite Linear Programming (SILP) [Sonnenburg et al., 2006]
 Subgradient Descent (SD) [Rakotomamonjy et al., 2008]
Properties of Bounds
Properties of Gap
Steps of SILP and SD
4.3 Discussion
SILP
pro
con
SD
pro
3.3 Level method illustration
SILP:
 oscillation of solutions
SD:
 a large number of calls to SVM required to compute
the optimal step size via a line search
 e.g., for “iono”, 1231 times for SD, while 47 for level
method
Level method:
 the cutting plane model
 the projection to level sets ensures the stability of solutions
5 Conclusion
An efficient method for large scale multiple kernel learning
con
 utilize the gradients of all the past solutions
 introduce a projection step for regularization
Level / breast
Download