华南理工大学学报(自然科学版) ›› 2015, Vol. 43 ›› Issue (5): 78-85.doi: 10.3969/j.issn.1000-565X.2015.05.013

• 计算机科学与技术 • 上一篇    下一篇

基于原问题求解的非稀疏多核学习方法

胡庆辉1,2 丁立新1† 刘晓刚2 李照奎1   

  1. 1. 武汉大学 计算机学院∥软件工程国家重点实验室,湖北 武汉 430072;2. 桂林航天工业学院 广西高校机器人与焊接技术重点实验室培育基地,广西 桂林 541004
  • 收稿日期:2014-09-19 修回日期:2015-01-25 出版日期:2015-05-25 发布日期:2015-05-07
  • 通信作者: 丁立新(1967-),男,教授,博士生导师,主要从事机器学习、智能计算及其理论研究. E-mail:lxding@whu.edu.cn
  • 作者简介:胡庆辉(1976-),男,在职博士生,桂林航天工业学院副教授,主要从事多核学习、监督学习、半监督学习及数据挖 掘研究. E-mail: huqinghui2004@126. com
  • 基金资助:

    国家自然科学基金资助项目(11301106);广西自然科学基金资助项目(2014GXNSFAA1183105);广西高校科研重点资助项目(ZD2014147);广西高校科研项目(YB2014431);桂林航天工业学院科研基金资助项目(Y12Z028)

A Non-Sparse Multi-Kernel Learning Method Based on Primal Problem

Hu Qing-hui1,2 Ding Li-xin1 Liu Xiao-gang2 Li Zhao-kui1   

  1. 1. School of Computer//State Key Laboratory of Software Engineering,Wuhan University,Wuhan 430072,Hubei,China;
    2. Guangxi Colleges and Universities Key Laboratory Breeding Base of Robot and Welding Technology,Guilin University of Aerospace Technology,Guilin 541004,Guangxi,China
  • Received:2014-09-19 Revised:2015-01-25 Online:2015-05-25 Published:2015-05-07
  • Contact: 丁立新(1967-),男,教授,博士生导师,主要从事机器学习、智能计算及其理论研究. E-mail:lxding@whu.edu.cn
  • About author:胡庆辉(1976-),男,在职博士生,桂林航天工业学院副教授,主要从事多核学习、监督学习、半监督学习及数据挖 掘研究. E-mail: huqinghui2004@126. com
  • Supported by:
    Supported by the National Natural Science Foundation of China(11301106) and the Natural Science Foundation of Guangxi Province(2014GXNSFAA1183105)

摘要: 传统的多核学习方法通常将原问题转换为其对偶问题再进行求解,但直接求解原问题比求解对偶问题有更好的收敛属性. 为此,文中提出了一种在原问题上求解、L P 范数约束的非稀疏多核学习算法,首先采用次梯度和改进的拟牛顿法求解支持向量机(SVM),然后通过简单计算求解基本核的权系数. 由于拟牛顿法具有二次收敛性,并且不需要计算二阶导数来得到 Hessian 矩阵的逆,因此文中算法具有更快的收敛速度. 仿真结果表明,文中算法不仅具有较好的分类精度和泛化性能,还具有较快的收敛速度及很好的可扩展性.

关键词: 多核学习, 拟牛顿法, 两步优化, 支持向量机

Abstract: Traditional multi-kernel learning (MKL) methods mainly solve primal problems in the dual. However,the solving in the primal may result in better convergence property. In this paper,a novel L p -norm-constraint non-sparse MKL method,which optimizes the modal in the primal,is proposed. In this method,firstly,support vector machine (SVM) is solved by means of subgradient and improved quasi-Newton method. Then,basic kernel weights are obtained via simple calculations. As quasi-Newton method is of second-order convergence property and acquires inverse Hessian matrix without computing the second-order derivative,the proposed method is of higher convergence speed than that of conventional ones. Simulated results show that the proposed method is of comparable classifica-tion accuracy,strong generalization capability,high convergence speed and good scalability.

Key words: multi-kernel learning, quasi-Newton method, alternating optimization, support vector machines

中图分类号: