A minimax optimal approach to high-dimensional double sparse linear regression
Title: | A minimax optimal approach to high-dimensional double sparse linear regression |
---|---|
Authors: | Zhang, Yanhang, Li, Zhifan, Liu, Shixiang, Yin, Jianxin |
Source: | Journal of machine learning research, 2024 |
Publication Year: | 2023 |
Collection: | Mathematics Statistics |
Subject Terms: | Mathematics - Statistics Theory |
More Details: | In this paper, we focus our attention on the high-dimensional double sparse linear regression, that is, a combination of element-wise and group-wise sparsity. To address this problem, we propose an IHT-style (iterative hard thresholding) procedure that dynamically updates the threshold at each step. We establish the matching upper and lower bounds for parameter estimation, showing the optimality of our proposal in the minimax sense. More importantly, we introduce a fully adaptive optimal procedure designed to address unknown sparsity and noise levels. Our adaptive procedure demonstrates optimal statistical accuracy with fast convergence. Additionally, we elucidate the significance of the element-wise sparsity level $s_0$ as the trade-off between IHT and group IHT, underscoring the superior performance of our method over both. Leveraging the beta-min condition, we establish that our IHT-style procedure can attain the oracle estimation rate and achieve almost full recovery of the true support set at both the element level and group level. Finally, we demonstrate the superiority of our method by comparing it with several state-of-the-art algorithms on both synthetic and real-world datasets. Comment: 74 pages, 8 figures, 4 tables |
Document Type: | Working Paper |
Access URL: | http://arxiv.org/abs/2305.04182 |
Accession Number: | edsarx.2305.04182 |
Database: | arXiv |
Description not available. |