Abstract：Best subset selection aims to find a small subset of predictors that lead to the most desirable and pre-defined prediction accuracy in a linear regression model. It is not only the most fundamental problem in regression analysis, but also has far reaching applications in every facet of research including computer science and medicine. We introduce a polynomial algorithm which under mild conditions, solves the problem. This algorithm exploits the idea of sequencing and splicing to reach the stable solution in finite steps when the sparsity level of the model is fixed but unknown. We define a novel information criterion that the algorithm uses to select the true sparsity level with a high probability. We show when the algorithm produces a stable optimal solution that is the oracle estimator of the true parameters with probability one. We also demonstrate the power of the algorithm in several numerical studies.
报告人简介：王学钦，中国科学技术大学管理学院教授。2003年毕业于纽约州立大学宾厄姆顿分校。他现担任教育部高等学校统计学类专业教学指导委员会委员、统计学国际期刊《JASA》等的Associate Editor、高等教育出版社《Lecture Notes: Data Science, Statistics and Probability》系列丛书的副主编。