Aggregating Regression Procedures for a Better Performance

Thumbnail Image
Date
1999-12-01
Authors
Yang, Yuhong
Major Professor
Advisor
Committee Member
Journal Title
Journal ISSN
Volume Title
Publisher
Authors
Research Projects
Organizational Units
Organizational Unit
Statistics
As leaders in statistical research, collaboration, and education, the Department of Statistics at Iowa State University offers students an education like no other. We are committed to our mission of developing and applying statistical methods, and proud of our award-winning students and faculty.
Journal Issue
Is Version Of
Versions
Series
Department
Statistics
Abstract

A fundamental question regarding combining procedures concerns the potential gain and how much one needs to pay for it in terms of statistical risk. Juditsky and Nemirovski considered the case where a large number of procedures are to be combined. We give upper and lower bounds for complementary cases. Under an l1 constraint on the linear coefficients, it is shown that for pursuing the best linear combination of nτ procedures, in terms of rate of convergence under the squared L2 loss, one can pay a price of order O(log n/n1-τ) when 0 < τ < 1/2 and a price of order O((log n/n)1/2) when 1/2 ≤ τ < ∞. These rates cannot be improved or essentially improved in a uniform sense. This result suggests that one should be cautious in pursuing the best linear combination, because one may end up paying a high price for nothing when linear combination in fact does not help. We show that with care in aggregation, the final procedure can automatically avoid paying the high price for such a case and then behaves as well as the best candidate procedure.

Comments

This preprint was published as Yuhong Yang, "Aggregating Regression Procedures to Improve Performance", Bernoulli (2004): 25-47, doi: 10.3150/bj/1077544602.

Description
Keywords
Citation
DOI
Source
Subject Categories
Copyright
Collections