An Efficient V-Minimum Absolute Deviation Distribution Regression Machine

IEEE ACCESS(2020)

引用 3|浏览56
暂无评分
摘要
Support Vector Regression (SVR) and its variants are widely used regression algorithms, and they have demonstrated high generalization ability. This research proposes a new SVR-based regressor : v-minimum absolute deviation distribution regression (v-MADR) machine. Instead of merely minimizing structural risk, as with v-SVR, v-MADR aims to achieve better generalization performance by minimizing both the absolute regression deviation mean and the absolute regression deviation variance, which takes into account the positive and negative values of the regression deviation of sample points. For optimization, we propose a dual coordinate descent (DCD) algorithm for small sample problems, and we also propose an averaged stochastic gradient descent (ASGD) algorithm for large-scale problems. Furthermore, we study the statistical property of v-MADR that leads to a bound on the expectation of error. The experimental results on both artificial and real datasets indicate that our nu-MADR has significant improvement in generalization performance with less training time compared to the widely used v-SVR, LS-SVR, "-TSVR, and linear "-SVR. Finally, we open source the code of nu-MADR at https://github.com/AsunaYY/v-MADR for wider dissemination.
更多
查看译文
关键词
nu-support vector regression, absolute regression deviation mean, absolute regression deviation variance, dual coordinate descent algorithm
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要