Dissemin is shutting down on January 1st, 2025

Published in

Elsevier, Computational Statistics & Data Analysis, 3(50), p. 642-658

DOI: 10.1016/j.csda.2004.09.007

Links

Tools

Export citation

Search in Google Scholar

An even faster algorithm for ridge regression of reduced rank data

Journal article published in 2006 by Berwin A. Turlach ORCID
This paper is available in a repository.
This paper is available in a repository.

Full text: Download

Green circle
Preprint: archiving allowed
Red circle
Postprint: archiving forbidden
Red circle
Published version: archiving forbidden
Data provided by SHERPA/RoMEO

Abstract

Hawkins and Yin (Comput. Statist. Data Anal. 40 (2002) 253) describe an algorithm for ridge regression of reduced rank data, i.e.data where p, the number of variables, is larger than n, the number of observations. Whereas a direct implementation of ridge regression in this setting requires calculations of order O(np2+p3), their algorithm uses only calculations of order O(np2). In this paper, we describe an alternative algorithm based on a factorization of the (transposed) design matrix. This approach is numerically more stable, further reduces the amount of calculations and needs less memory. In particular, we show that the factorization can be calculated in O(n2p) operations. Once the factorization is obtained, for any value of the ridge parameter the ridge regression estimator can be calculated in O(np) operations and the generalized cross-validation score in O(n) operations.