Package Biglasso': R Topics Documented
Package Biglasso': R Topics Documented
Package Biglasso': R Topics Documented
URL https://github.com/YaohuiZeng/biglasso,
https://arxiv.org/abs/1701.05936
BugReports https://github.com/YaohuiZeng/biglasso/issues
Depends R (>= 3.2.0), bigmemory (>= 4.5.0), Matrix, ncvreg
Imports Rcpp (>= 0.12.1), methods
LinkingTo Rcpp, RcppArmadillo (>= 0.8.600), bigmemory, BH
VignetteBuilder knitr
Suggests parallel, testthat, R.rsp, glmnet, survival, knitr, rmarkdown
RoxygenNote 7.1.1
NeedsCompilation yes
Repository CRAN
Date/Publication 2021-01-31 15:30:06 UTC
R topics documented:
biglasso-package . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
biglasso . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
colon . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1
2 biglasso-package
cv.biglasso . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
plot.biglasso . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
plot.cv.biglasso . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
predict.biglasso . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
predict.cv.biglasso . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
setupX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
summary.cv.biglasso . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
Index 21
Description
Extend lasso and elastic-net linear, logistic and cox regression models for ultrahigh-dimensional,
multi-gigabyte data sets that cannot be loaded into available RAM. This package utilizes memory-
mapped files to store the massive data on the disk and only read those into memory whenever
necessary during model fitting. Moreover, some advanced feature screening rules are proposed and
implemented to accelerate the model fitting. As a result, this package is much more memory- and
computation-efficient and highly scalable as compared to existing lasso-fitting packages such as
glmnet and ncvreg, thus allowing for powerful big data analysis even with only an ordinary laptop.
Details
Package: biglasso
Type: Package
Version: 1.4-1
Date: 2021-01-29
License: GPL-3
Penalized regression models, in particular the lasso, have been extensively applied to analyzing
high-dimensional data sets. However, due to the memory limit, existing R packages are not ca-
pable of fitting lasso models for ultrahigh-dimensional, multi-gigabyte data sets which have been
increasingly seen in many areas such as genetics, biomedical imaging, genome sequencing and
high-frequency finance.
This package aims to fill the gap by extending lasso model fitting to Big Data in R. Version >=
1.2-3 represents a major redesign where the source code is converted into C++ (previously in C),
and new feature screening rules, as well as OpenMP parallel computing, are implemented. Some
key features of biglasso are summarized as below:
1. it utilizes memory-mapped files to store the massive data on the disk, only loading data into
memory when necessary during model fitting. Consequently, it’s able to seamlessly data-
larger-than-RAM cases.
2. it is built upon pathwise coordinate descent algorithm with warm start, active set cycling, and
feature screening strategies, which has been proven to be one of fastest lasso solvers.
biglasso-package 3
3. in incorporates our newly developed hybrid and adaptive screening that outperform state-of-
the-art screening rules such as the sequential strong rule (SSR) and the sequential EDPP rule
(SEDPP) with additional 1.5x to 4x speedup.
4. the implementation is designed to be as memory-efficient as possible by eliminating extra
copies of the data created by other R packages, making it at least 2x more memory-efficient
than glmnet.
5. the underlying computation is implemented in C++, and parallel computing with OpenMP is
also supported.
For more information:
Note
The input design matrix X must be a big.matrix object. This can be created by the function
as.big.matrix in the R package bigmemory. If the data (design matrix) is very large (e.g. 10 GB)
and stored in an external file, which is often the case for big data, X can be created by calling the
function setupX. In this case, there are several restrictions about the data file:
1. the data file must be a well-formated ASCII-file, with each row corresponding to an observa-
tion and each column a variable;
2. the data file must contain only one single type. Current version only supports double type;
3. the data file must contain only numeric variables. If there are categorical variables, the user
needs to create dummy variables for each categorical varable (by adding additional columns).
Future versions will try to address these restrictions.
Denote the number of observations and variables be, respectively, n and p. It’s worth noting that the
package is more suitable for wide data (ultrahigh-dimensional, p >> n) as compared to long data (n
>> p). This is because the model fitting algorithm takes advantage of sparsity assumption of high-
dimensional data. To just give the user some ideas, below are some benchmarking results of the
total computing time (in seconds) for solving lasso-penalized linear regression along a sequence of
100 values of the tuning parameter. In all cases, assume 20 non-zero coefficients equal +/- 2 in the
true model. (Based on Version 1.2-3, screening rule "SSR-BEDPP" is used)
• For wide data case (p > n), n = 1,000:
Author(s)
Yaohui Zeng, Chuyi Wang and Patrick Breheny
Maintainer: Yaohui Zeng <yaohui.zeng@gmail.com> and Chuyi Wang <wwaa0208@gmail.com>
4 biglasso-package
References
• Zeng, Y., and Breheny, P. (2017). The biglasso Package: A Memory- and Computation-
Efficient Solver for Lasso Model Fitting with Big Data in R. https://arxiv.org/abs/1701.
05936.
• Tibshirani, R., Bien, J., Friedman, J., Hastie, T., Simon, N., Taylor, J., and Tibshirani, R. J.
(2012). Strong rules for discarding predictors in lasso-type problems. Journal of the Royal
Statistical Society: Series B (Statistical Methodology), 74(2), 245-266.
• Wang, J., Zhou, J., Wonka, P., and Ye, J. (2013). Lasso screening rules via dual polytope
projection. In Advances in Neural Information Processing Systems, pp. 1070-1078.
• Xiang, Z. J., and Ramadge, P. J. (2012). Fast lasso screening tests based on correlations. In
Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE International Conference on
(pp. 2137-2140). IEEE.
• Wang, J., Zhou, J., Liu, J., Wonka, P., and Ye, J. (2014). A safe screening rule for sparse
logistic regression. In Advances in Neural Information Processing Systems, pp. 1053-1061.
Examples
## Not run:
## Example of reading data from external big data file, fit lasso model,
## and run cross validation in parallel
# the big.matrix then can be retrieved by its descriptor file (.desc) in any new R session.
rm(X)
xdesc <- 'x_e3_5e5.desc'
X <- attach.big.matrix(xdesc)
dim(X)
plot(cvfit)
summary(cvfit)
## End(Not run)
Description
Extend lasso model fitting to big data that cannot be loaded into memory. Fit solution paths for
linear or logistic regression models penalized by lasso, ridge, or elastic-net over a grid of values for
the regularization parameter lambda.
Usage
biglasso(
X,
y,
row.idx = 1:nrow(X),
penalty = c("lasso", "ridge", "enet"),
family = c("gaussian", "binomial", "cox"),
alg.logistic = c("Newton", "MM"),
screen = c("Adaptive", "SSR", "Hybrid", "None"),
safe.thresh = 0,
update.thresh = 1,
ncores = 1,
alpha = 1,
lambda.min = ifelse(nrow(X) > ncol(X), 0.001, 0.05),
nlambda = 100,
lambda.log.scale = TRUE,
lambda,
eps = 1e-07,
max.iter = 1000,
dfmax = ncol(X) + 1,
penalty.factor = rep(1, ncol(X)),
warn = TRUE,
output.time = FALSE,
return.time = TRUE,
verbose = FALSE
)
Arguments
X The design matrix, without an intercept. It must be a big.matrix object. The
function standardizes the data and includes an intercept internally by default
during the model fitting.
6 biglasso
α||β||1 + (1 − α)/2||β||22 .
alpha=1 is the lasso penalty, alpha=0 the ridge penalty, alpha in between 0 and
1 is the elastic-net ("enet") penalty.
lambda.min The smallest value for lambda, as a fraction of lambda.max. Default is .001
if the number of observations is larger than the number of covariates and .05
otherwise.
nlambda The number of lambda values. Default is 100.
lambda.log.scale
Whether compute the grid values of lambda on log scale (default) or linear scale.
biglasso 7
Details
The objective function for linear regression (family = "gaussian") is
1
RSS + λ ∗ penalty,
2n
for logistic regression (family = "binomial") it is
1
− loglike + λ ∗ penalty,
n
, for cox regression, breslow approximation for ties is applied.
Several advanced feature screening rules are implemented. For lasso-penalized linear regression, all
the options of screen are applicable. Our proposal adaptive rule - "Adaptive" - achieves highest
speedup so it’s the recommended one, especially for ultrahigh-dimensional large-scale data sets.
For cox regression and/or the elastic net penalty, only "SSR" is applicable for now. More efficient
rules are under development.
Value
An object with S3 class "biglasso" with following variables.
beta The fitted matrix of coefficients, store in sparse matrix representation. The
number of rows is equal to the number of coefficients, whereas the number of
columns is equal to nlambda.
8 biglasso
iter A vector of length nlambda containing the number of iterations until conver-
gence at each value of lambda.
lambda The sequence of regularization parameter values in the path.
penalty Same as above.
family Same as above.
alpha Same as above.
loss A vector containing either the residual sum of squares (for "gaussian") or
negative log-likelihood (for "binomial") of the fitted model at each value of
lambda.
penalty.factor Same as above.
n The number of observations used in the model fitting. It’s equal to length(row.idx).
center The sample mean vector of the variables, i.e., column mean of the sub-matrix of
X used for model fitting.
scale The sample standard deviation of the variables, i.e., column standard deviation
of the sub-matrix of X used for model fitting.
y The response vector used in the model fitting. Depending on row.idx, it could
be a subset of the raw input of the response vector y.
screen Same as above.
col.idx The indices of features that have ’scale’ value greater than 1e-6. Features with
’scale’ less than 1e-6 are removed from model fitting.
rejections The number of features rejected at each value of lambda.
safe_rejections
The number of features rejected by safe rules at each value of lambda.
Author(s)
Yaohui Zeng, Chuyi Wang and Patrick Breheny
Maintainer: Yaohui Zeng <yaohui.zeng@gmail.com> and Chuyi Wang <wwaa0208@gmail.com>
See Also
biglasso-package, setupX, cv.biglasso, plot.biglasso, ncvreg
Examples
## Linear regression
data(colon)
X <- colon$X
y <- colon$y
X.bm <- as.big.matrix(X)
# lasso, default
par(mfrow=c(1,2))
fit.lasso <- biglasso(X.bm, y, family = 'gaussian')
plot(fit.lasso, log.l = TRUE, main = 'lasso')
colon 9
# elastic net
fit.enet <- biglasso(X.bm, y, penalty = 'enet', alpha = 0.5, family = 'gaussian')
plot(fit.enet, log.l = TRUE, main = 'elastic net, alpha = 0.5')
## Logistic regression
data(colon)
X <- colon$X
y <- colon$y
X.bm <- as.big.matrix(X)
# lasso, default
par(mfrow = c(1, 2))
fit.bin.lasso <- biglasso(X.bm, y, penalty = 'lasso', family = "binomial")
plot(fit.bin.lasso, log.l = TRUE, main = 'lasso')
# elastic net
fit.bin.enet <- biglasso(X.bm, y, penalty = 'enet', alpha = 0.5, family = "binomial")
plot(fit.bin.enet, log.l = TRUE, main = 'elastic net, alpha = 0.5')
## Cox regression
set.seed(10101)
N <- 1000; p <- 30; nzc <- p/3
X <- matrix(rnorm(N * p), N, p)
beta <- rnorm(nzc)
fx <- X[, seq(nzc)] %*% beta/3
hx <- exp(fx)
ty <- rexp(N, hx)
tcens <- rbinom(n = N, prob = 0.3, size = 1) # censoring indicator
y <- cbind(time = ty, status = 1 - tcens) # y <- Surv(ty, 1 - tcens) with library(survival)
X.bm <- as.big.matrix(X)
fit <- biglasso(X.bm, y, family = "cox")
plot(fit)
Description
The data file contains gene expression data of 62 samples (40 tumor samples, 22 normal samples)
from colon-cancer patients analyzed with an Affymetrix oligonucleotide Hum6000 array.
Format
• X: a 62-by-2000 matrix that records the gene expression data. Used as design matrix.
• y: a binary vector of length 62 recording the sample status: 1 = tumor; 0 = normal. Used as
response vector.
10 cv.biglasso
Source
References
• U. Alon et al. (1999): Broad patterns of gene expression revealed by clustering analysis of
tumor and normal colon tissue probed by oligonucleotide arrays. Proc. Natl. Acad. Sci. USA
96, 6745-6750. https://www.pnas.org/content/96/12/6745.short.
Examples
data(colon)
X <- colon$X
y <- colon$y
str(X)
dim(X)
X.bm <- as.big.matrix(X, backingfile = "") # convert to big.matrix object
str(X.bm)
dim(X.bm)
Description
Perform k-fold cross validation for penalized regression models over a grid of values for the regu-
larization parameter lambda.
Usage
cv.biglasso(
X,
y,
row.idx = 1:nrow(X),
eval.metric = c("default", "MAPE"),
ncores = parallel::detectCores(),
...,
nfolds = 5,
seed,
cv.ind,
trace = FALSE
)
cv.biglasso 11
Arguments
X The design matrix, without an intercept, as in biglasso.
y The response vector, as in biglasso.
row.idx The integer vector of row indices of X that used for fitting the model. as in
biglasso.
eval.metric The evaluation metric for the cross-validated error and for choosing optimal
lambda. "default" for linear regression is MSE (mean squared error), for logistic
regression is misclassification error. "MAPE", for linear regression only, is the
Mean Absolute Percentage Error.
ncores The number of cores to use for parallel execution across a cluster created by the
parallel package. (This is different from ncores in biglasso, which is the
number of OpenMP threads.)
... Additional arguments to biglasso.
nfolds The number of cross-validation folds. Default is 5.
seed The seed of the random number generator in order to obtain reproducible results.
cv.ind Which fold each observation belongs to. By default the observations are ran-
domly assigned by cv.biglasso.
trace If set to TRUE, cv.biglasso will inform the user of its progress by announcing
the beginning of each CV fold. Default is FALSE.
Details
The function calls biglasso nfolds times, each time leaving out 1/nfolds of the data. The cross-
validation error is based on the residual sum of squares when family="gaussian" and the binomial
deviance when family="binomial".
The S3 class object cv.biglasso inherits class cv.ncvreg. So S3 functions such as "summary","plot"
can be directly applied to the cv.biglasso object.
Value
An object with S3 class "cv.biglasso" which inherits from class "cv.ncvreg". The following
variables are contained in the class (adopted from cv.ncvreg).
cve The error for each value of lambda, averaged across the cross-validation folds.
cvse The estimated standard error associated with each value of for cve.
lambda The sequence of regularization parameter values along which the cross-validation
error was calculated.
fit The fitted biglasso object for the whole data.
min The index of lambda corresponding to lambda.min.
lambda.min The value of lambda with the minimum cross-validation error.
null.dev The deviance for the intercept-only model.
pe If family="binomial", the cross-validation prediction error for each value of
lambda.
cv.ind Same as above.
12 plot.biglasso
Author(s)
Yaohui Zeng and Patrick Breheny
Maintainer: Yaohui Zeng <yaohui.zeng@gmail.com>
See Also
biglasso, plot.cv.biglasso, summary.cv.biglasso, setupX
Examples
## Not run:
## cv.biglasso
data(colon)
X <- colon$X
y <- colon$y
X.bm <- as.big.matrix(X)
## logistic regression
cvfit <- cv.biglasso(X.bm, y, family = 'binomial', seed = 1234, ncores = 2)
par(mfrow = c(2, 2))
plot(cvfit, type = 'all')
summary(cvfit)
## End(Not run)
Description
Produce a plot of the coefficient paths for a fitted biglasso object.
Usage
## S3 method for class 'biglasso'
plot(x, alpha = 1, log.l = TRUE, ...)
Arguments
x Fitted "biglasso" model.
alpha Controls alpha-blending, helpful when the number of covariates is large. Default
is alpha=1.
log.l Should horizontal axis be on the log scale? Default is TRUE.
... Other graphical parameters to plot
plot.cv.biglasso 13
Author(s)
Yaohui Zeng and Patrick Breheny
Maintainer: Yaohui Zeng <yaohui.zeng@gmail.com>
See Also
biglasso, cv.biglasso
Examples
Description
Plot the cross-validation curve from a cv.biglasso object, along with standard error bars.
Usage
## S3 method for class 'cv.biglasso'
plot(
x,
log.l = TRUE,
type = c("cve", "rsq", "scale", "snr", "pred", "all"),
selected = TRUE,
vertical.line = TRUE,
col = "red",
...
)
Arguments
x A "cv.biglasso" object.
log.l Should horizontal axis be on the log scale? Default is TRUE.
type What to plot on the vertical axis. cve plots the cross-validation error (deviance);
rsq plots an estimate of the fraction of the deviance explained by the model
(R-squared); snr plots an estimate of the signal-to-noise ratio; scale plots, for
family="gaussian", an estimate of the scale parameter (standard deviation);
pred plots, for family="binomial", the estimated prediction error; all pro-
duces all of the above.
selected If TRUE (the default), places an axis on top of the plot denoting the number of
variables in the model (i.e., that have a nonzero regression coefficient) at that
value of lambda.
14 predict.biglasso
vertical.line If TRUE (the default), draws a vertical line at the value where cross-validaton
error is minimized.
col Controls the color of the dots (CV estimates).
... Other graphical parameters to plot
Details
Error bars representing approximate 68% confidence intervals are plotted along with the estimates
at value of lambda. For rsq and snr, these confidence intervals are quite crude, especially near.
Author(s)
Yaohui Zeng and Patrick Breheny
Maintainer: Yaohui Zeng <yaohui.zeng@gmail.com>
See Also
biglasso, cv.biglasso
Examples
Description
Extract predictions (fitted reponse, coefficients, etc.) from a fitted biglasso object.
Usage
## S3 method for class 'biglasso'
predict(
object,
X,
row.idx = 1:nrow(X),
type = c("link", "response", "class", "coefficients", "vars", "nvars"),
lambda,
which = 1:length(object$lambda),
...
)
Arguments
object A fitted "biglasso" model object.
X Matrix of values at which predictions are to be made. It must be a big.matrix
object. Not used for type="coefficients".
row.idx Similar to that in biglasso, it’s a vector of the row indices of X that used for the
prediction. 1:nrow(X) by default.
type Type of prediction: "link" returns the linear predictors; "response" gives the
fitted values; "class" returns the binomial outcome with the highest probabil-
ity; "coefficients" returns the coefficients; "vars" returns a list containing
the indices and names of the nonzero variables at each value of lambda; "nvars"
returns the number of nonzero coefficients at each value of lambda.
lambda Values of the regularization parameter lambda at which predictions are requested.
Linear interpolation is used for values of lambda not in the sequence of lambda
values in the fitted models.
which Indices of the penalty parameter lambda at which predictions are required. By
default, all indices are returned. If lambda is specified, this will override which.
... Not used.
drop If coefficients for a single value of lambda are to be returned, reduce dimensions
to a vector? Setting drop=FALSE returns a 1-column matrix.
Value
The object returned depends on type.
Author(s)
Yaohui Zeng and Patrick Breheny
Maintainer: Yaohui Zeng <yaohui.zeng@gmail.com>
See Also
biglasso, cv.biglasso
Examples
## Logistic regression
data(colon)
X <- colon$X
y <- colon$y
X.bm <- as.big.matrix(X, backingfile = "")
fit <- biglasso(X.bm, y, penalty = 'lasso', family = "binomial")
coef <- coef(fit, lambda=0.05, drop = TRUE)
coef[which(coef != 0)]
predict(fit, X.bm, type="link", lambda=0.05)
predict(fit, X.bm, type="response", lambda=0.05)
predict(fit, X.bm, type="class", lambda=0.1)
predict(fit, type="vars", lambda=c(0.05, 0.1))
predict(fit, type="nvars", lambda=c(0.05, 0.1))
16 predict.cv.biglasso
Description
Extract predictions from a fitted cv.biglasso object.
Usage
## S3 method for class 'cv.biglasso'
predict(
object,
X,
row.idx = 1:nrow(X),
type = c("link", "response", "class", "coefficients", "vars", "nvars"),
lambda = object$lambda.min,
which = object$min,
...
)
Arguments
object A fitted "cv.biglasso" model object.
X Matrix of values at which predictions are to be made. It must be a big.matrix
object. Not used for type="coefficients".
row.idx Similar to that in biglasso, it’s a vector of the row indices of X that used for the
prediction. 1:nrow(X) by default.
type Type of prediction: "link" returns the linear predictors; "response" gives the
fitted values; "class" returns the binomial outcome with the highest probabil-
ity; "coefficients" returns the coefficients; "vars" returns a list containing
the indices and names of the nonzero variables at each value of lambda; "nvars"
returns the number of nonzero coefficients at each value of lambda.
lambda Values of the regularization parameter lambda at which predictions are requested.
The default value is the one corresponding to the minimum cross-validation er-
ror.
which Indices of the penalty parameter lambda at which predictions are requested. The
default value is the index of lambda corresponding to lambda.min. Note: this is
overridden if lambda is specified.
... Not used.
Value
The object returned depends on type.
setupX 17
Author(s)
Yaohui Zeng and Patrick Breheny
Maintainer: Yaohui Zeng <yaohui.zeng@gmail.com>
See Also
biglasso, cv.biglasso
Examples
## Not run:
## predict.cv.biglasso
data(colon)
X <- colon$X
y <- colon$y
X.bm <- as.big.matrix(X, backingfile = "")
fit <- biglasso(X.bm, y, penalty = 'lasso', family = "binomial")
cvfit <- cv.biglasso(X.bm, y, penalty = 'lasso', family = "binomial", seed = 1234, ncores = 2)
coef <- coef(cvfit)
coef[which(coef != 0)]
predict(cvfit, X.bm, type = "response")
predict(cvfit, X.bm, type = "link")
predict(cvfit, X.bm, type = "class")
## End(Not run)
setupX Set up design matrix X by reading data from big data file
Description
Set up the design matrix X as a big.matrix object based on external massive data file stored on
disk that cannot be fullly loaded into memory. The data file must be a well-formated ASCII-file, and
contains only one single type. Current version only supports double type. Other restrictions about
the data file are described in biglasso-package. This function reads the massive data, and creates
a big.matrix object. By default, the resulting big.matrix is file-backed, and can be shared across
processors or nodes of a cluster.
Usage
setupX(
filename,
dir = getwd(),
sep = ",",
backingfile = paste0(unlist(strsplit(filename, split = "\\."))[1], ".bin"),
descriptorfile = paste0(unlist(strsplit(filename, split = "\\."))[1], ".desc"),
type = "double",
...
)
18 setupX
Arguments
Details
For a data set, this function needs to be called only one time to set up the big.matrix object with
two backing files (.bin, .desc) created in current working directory. Once set up, the data can be
"loaded" into any (new) R session by calling attach.big.matrix(discriptorfile).
This function is a simple wrapper of read.big.matrix. See read.big.matrix and the package
bigmemory for more details.
Value
Author(s)
See Also
biglasso, cv.ncvreg
Examples
## see the example in "biglasso-package"
summary.cv.biglasso 19
Description
Summary method for cv.biglasso objects.
Usage
## S3 method for class 'cv.biglasso'
summary(object, ...)
Arguments
object A cv.biglasso object.
... Further arguments passed to or from other methods.
x A "summary.cv.biglasso" object.
digits Number of digits past the decimal point to print out. Can be a vector specifying
different display digits for each of the five non-integer printed values.
Value
summary.cv.biglasso produces an object with S3 class "summary.cv.biglasso" which inherits
class "summary.cv.ncvreg". The class has its own print method and contains the following list
elements:
penalty The penalty used by biglasso.
model Either "linear" or "logistic", depending on the family option in biglasso.
n Number of observations
p Number of regression coefficients (not including the intercept).
min The index of lambda with the smallest cross-validation error.
lambda The sequence of lambda values used by cv.biglasso.
cve Cross-validation error (deviance).
r.squared Proportion of variance explained by the model, as estimated by cross-validation.
snr Signal to noise ratio, as estimated by cross-validation.
sigma For linear regression models, the scale parameter estimate.
pe For logistic regression models, the prediction error (misclassification error).
Author(s)
Yaohui Zeng and Patrick Breheny
Maintainer: Yaohui Zeng <yaohui.zeng@gmail.com>
20 summary.cv.biglasso
See Also
biglasso, cv.biglasso, plot.cv.biglasso
Examples
## See examples in "cv.biglasso" and "biglasso-package"
Index
∗ datasets
colon, 9
∗ models
plot.biglasso, 12
plot.cv.biglasso, 13
predict.biglasso, 14
predict.cv.biglasso, 16
summary.cv.biglasso, 19
∗ package
biglasso-package, 2
∗ regression
plot.biglasso, 12
plot.cv.biglasso, 13
predict.biglasso, 14
predict.cv.biglasso, 16
summary.cv.biglasso, 19
big.matrix, 3, 5, 15, 16
biglasso, 5, 11–18, 20
biglasso-package, 2
coef.biglasso (predict.biglasso), 14
coef.cv.biglasso (predict.cv.biglasso),
16
colon, 9
cv.biglasso, 8, 10, 13–18, 20
cv.ncvreg, 11, 18
ncvreg, 8
plot.biglasso, 8, 12
plot.cv.biglasso, 12, 13, 20
predict.biglasso, 14
predict.cv.biglasso, 16
print.summary.cv.biglasso
(summary.cv.biglasso), 19
read.big.matrix, 18
setupX, 3, 8, 12, 17
summary.cv.biglasso, 12, 19
21