Parsimonious Least Squares Support Vector Regression Using Orthogonal Forward Selection With The Generalised Kernel Model

INTERNATIONAL JOURNAL OF MODELLING IDENTIFICATION AND CONTROL(2006)

引用 14|浏览5
暂无评分
摘要
A sparse regression modelling technique is developed using a generalised kernel model in which each kernel regressor has its individually tuned position (centre) vector and diagonal covariance matrix. An orthogonal least squares forward selection procedure is employed to append the regressors one by one. After the determination of the model structure, namely the selection of an appropriate number of regressors, the model weight parameters are calculated from the Lagrange dual problem of the original least squares problem. Different from the least squares support vector regression, this regression modelling procedure involves neither reproducing kernel Hilbert space nor Mercer decomposition concepts. As the regressors used are not restricted to be positioned at training input points and each regressor has its own diagonal covariance matrix, a very sparse representation can be obtained with excellent generalisation capability. Experimental results involving two real data sets demonstrate the effectiveness of the proposed regression modelling approach.
更多
查看译文
关键词
generalised kernel model, least squares support vector machine, orthogonal least squares forward selection, regression, sparse modelling
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要