Variable selection in convex quantile regression: L1-norm or L0-norm regularization?

Sheng Dai

Research output: Contribution to journalArticleScientificpeer-review

21 Citations (Scopus)
97 Downloads (Pure)

Abstract

The curse of dimensionality is a recognized challenge in nonparametric estimation. This paper develops a new L0-norm regularization approach to the convex quantile and expectile regressions for subset selection. We show how to use mixed-integer programming to solve the proposed L0-norm regularization approach in practice and build a link to the commonly used L1-norm regularization approach. A Monte Carlo study is performed to compare the finite sample performances of the proposed L0-penalized convex quantile and expectile regression approaches with the L1-norm regularization approaches. The proposed approach is further applied to benchmark the sustainable development performance of the OECD countries and empirically analyze the accuracy in the dimensionality reduction of variables. The results from the simulation and application illustrate that the proposed L0-norm regularization approach can more effectively address the curse of dimensionality than the L0-norm regularization approach in multidimensional spaces.
Original languageEnglish
Pages (from-to)338-355
Number of pages18
JournalEuropean Journal of Operational Research
Volume305
Issue number1
Early online date30 May 2022
DOIs
Publication statusPublished - 16 Feb 2023
MoE publication typeA1 Journal article-refereed

Keywords

  • Variable selection
  • Convex quantile regression
  • Regularization
  • SDG evaluation

Fingerprint

Dive into the research topics of 'Variable selection in convex quantile regression: L1-norm or L0-norm regularization?'. Together they form a unique fingerprint.
  • Science-IT

    Hakala, M. (Manager)

    School of Science

    Facility/equipment: Facility

Cite this