Auxiliary Learning and its Statistical Understanding

Type: Preprint

Publication Date: 2025-01-06

Citations: 0

DOI: https://doi.org/10.48550/arxiv.2501.03463

Abstract

Modern statistical analysis often encounters high-dimensional problems but with a limited sample size. It poses great challenges to traditional statistical estimation methods. In this work, we adopt auxiliary learning to solve the estimation problem in high-dimensional settings. We start with the linear regression setup. To improve the statistical efficiency of the parameter estimator for the primary task, we consider several auxiliary tasks, which share the same covariates with the primary task. Then a weighted estimator for the primary task is developed, which is a linear combination of the ordinary least squares estimators of both the primary task and auxiliary tasks. The optimal weight is analytically derived and the statistical properties of the corresponding weighted estimator are studied. We then extend the weighted estimator to generalized linear regression models. Extensive numerical experiments are conducted to verify our theoretical results. Last, a deep learning-related real-data example of smart vending machines is presented for illustration purposes.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ Learning a high-dimensional classification rule using auxiliary outcomes 2020 Muxuan Liang
Xiang Zhong
Jaeyoung Park
+ A Na¨Ive Bayes Regularized Logistic Regression Model for Classification 2024 Yi Tan
Ben Sherwood
Prakash P. Shenoy
+ SOFARI: High-Dimensional Manifold-Based Inference 2023 Zemin Zheng
Xin Zhou
Yingying Fan
Jinchi Lv
+ PDF Chat Are Latent Factor Regression and Sparse Regression Adequate? 2023 Jianqing Fan
Zhipeng Lou
Mengxin Yu
+ Are Latent Factor Regression and Sparse Regression Adequate? 2022 Jianqing Fan
Zhipeng Lou
Mengxin Yu
+ PDF Chat Statistical Learning from a Regression Perspective 2020 Richard A. Berk
+ Robust and flexible learning of a high-dimensional classification rule using auxiliary outcomes 2020 Muxuan Liang
Jaeyoung Park
Xiang Zhong
+ PDF Chat Task-Agnostic Machine Learning-Assisted Inference 2024 Jiacheng Miao
Qiongshi Lu
+ PDF Chat Robust and flexible learning of a high-dimensional classification rule using auxiliary outcomes 2024 Muxuan Liang
Jaeyoung Park
Qing Lu
Xiang Zhong
+ Nonsparse learning with latent variables 2017 Zemin Zheng
Jinchi Lv
Wei Lin
+ Robust and Sparse Multinomial Regression in High Dimensions 2022 Fatma Sevinç Kurnaz
Peter Filzmoser
+ Nonparametric learning in high dimensions 2010 John Lafferey
Larry Wasserman
Han Liu
+ PDF Chat Factor Augmented Sparse Throughput Deep ReLU Neural Networks for High Dimensional Regression 2023 Jianqing Fan
Yihong Gu
+ Statistical Learning as a Regression Problem 2020 Richard A. Berk
+ TSVR+: Twin support vector regression with privileged information 2023 Anuradha Kumari
M. Tanveer
+ Wasserstein Logistic Regression with Mixed Features 2022 Aras Selvi
Mohammad Reza Belbasi
Martin B. Haugh
Wolfram Wiesemann
+ Statistical Learning 2020 Jimmie Leppink
+ A proximal forward-backward splitting based algorithmic framework for Wasserstein logistic regression using heavy ball strategy 2023 Bo Zhou
Yuefei Yuan
Qiankun Song
+ Simultaneous Dimension Reduction and Variable Selection for Multinomial Logistic Regression 2023 Canhong Wen
Zhenduo Li
Ruipeng Dong
Yijin Ni
Wenliang Pan
+ An Extended Newton-type Algorithm for $\ell_2$-Regularized Sparse Logistic Regression and Its Efficiency for Classifying Large-scale Datasets. 2021 Rui Wang
Naihua Xiu
Shenglong Zhou

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (0)

Action Title Year Authors