Rig Notes 17
Rig Notes 17
Rig Notes 17
Lecture Notes
Spring 2017
Preface
These lecture notes were written for the course 18.657, High Dimensional
Statistics at MIT. They build on a set of notes that was prepared at Prince-
ton University in 2013-14 that was modified (and hopefully improved) over the
years.
Over the past decade, statistics have undergone drastic changes with the
development of high-dimensional statistical inference. Indeed, on each indi-
vidual, more and more features are measured to a point that their number
usually far exceeds the number of observations. This is the case in biology and
specifically genetics where millions of (combinations of) genes are measured
for a single individual. High resolution imaging, finance, online advertising,
climate studies . . . the list of intensive data producing fields is too long to be
established exhaustively. Clearly not all measured features are relevant for a
given task and most of them are simply noise. But which ones? What can be
done with so little data and so much noise? Surprisingly, the situation is not
that bad and on some simple models we can assess to which extent meaningful
statistical methods can be applied. Regression is one such simple model.
Regression analysis can be traced back to 1632 when Galileo Galilei used
a procedure to infer a linear relationship from noisy data. It was not until
the early 19th century that Gauss and Legendre developed a systematic pro-
cedure: the least-squares method. Since then, regression has been studied
in so many forms that much insight has been gained and recent advances on
high-dimensional statistics would not have been possible without standing on
the shoulders of giants. In these notes, we will explore one, obviously subjec-
tive giant on whose shoulders high-dimensional statistics stand: nonparametric
statistics.
The works of Ibragimov and Has’minskii in the seventies followed by many
researchers from the Russian school have contributed to developing a large
toolkit to understand regression with an infinite number of parameters. Much
insight from this work can be gained to understand high-dimensional or sparse
regression and it comes as no surprise that Donoho and Johnstone have made
the first contributions on this topic in the early nineties.
Therefore, while not obviously connected to high dimensional statistics, we
will talk about nonparametric estimation. I borrowed this disclaimer (and the
template) from my colleague Ramon van Handel. It does apply here.
I have no illusions about the state of these notes—they were written
i
Preface ii
iii
Preface iv
Preface i
Notation iii
Contents v
Introduction 1
v
Contents vi
4.7 Lower bounds for estimating the `1 norm via moment matching 112
4.8 Problem Set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 120
Bibliography 157
Introduction
1
Introduction 2
variable Z small if IE[Z 2 ] = [IEZ]2 + var[Z] is small. Indeed in this case, the
expectation of Z is small and the fluctuations of Z around this value are also
small. The function R(g) = IE[Y − g(X)]2 is called the L2 risk of g defined for
IEY 2 < ∞.
For any measurable function g : X → IR, the L2 risk of g can be decom-
posed as
Note that khk22 is the Hilbert norm associated to the inner product
Z
hh, h0 i2 = hh0 dPX .
X
When the reference measure is clear from the context, we will simply write
khk2 = khkL2 (PX ) and hh, h0 i2 := hh, h0 iL2 (PX ) .
Introduction 3
It follows from the proof of the best prediction property above that
In particular, the prediction risk will always be at least equal to the positive
constant IE[Y − f (X)]2 . Since we tend to prefer a measure of accuracy to be
able to go to zero (as the sample size increases), it is equivalent to study the
estimation error kfˆn − f k22 . Note that if fˆn is random, then kfˆn − f k22 and
R(fˆn ) are random quantities and we need deterministic summaries to quantify
their size. It is customary to use one of the two following options. Let {φn }n
be a sequence of positive numbers that tends to zero as n goes to infinity.
1. Bounds in expectation. They are of the form:
IEkfˆn − f k22 ≤ φn ,
Often, bounds with high probability follow from a bound in expectation and a
concentration inequality that bounds the following probability
This equality allowed us to consider only the part kfˆn − f k22 as a measure of
error. While this decomposition may not hold for other risk measures, it may
be desirable to explore other distances (or pseudo-distances). This leads to two
distinct ways to measure error. Either by bounding a pseudo-distance d(fˆn , f )
(estimation error ) or by bounding the risk R(fˆn ) for choices other than the L2
risk. These two measures coincide up to the additive constant IE[Y − f (X)]2
in the case described above. However, we show below that these two quantities
may live independent lives. Bounding the estimation error is more customary
in statistics whereas risk bounds are preferred in learning theory.
Here is a list of choices for the pseudo-distance employed in the estimation
error.
• Pointwise error. Given a point x0 , the pointwise error measures only
the error at this point. It uses the pseudo-distance:
Note that these three examples can be split into two families: global (Sup-norm
and Lp ) and local (pointwise).
For specific problems, other considerations come into play. For example,
if Y ∈ {0, 1} is a label, one may be interested in the classification risk of a
classifier h : X → {0, 1}. It is defined by
It turns out that this principle can be extended even if an optimization follows
the substitution. Recall that the L2 risk is defined by R(g) = IE[Y −g(X)]2 . See
the expectation? Well, it can be replaced by an average to form the empirical
risk of g defined by
n
1X 2
Rn (g) = Yi − g(Xi ) .
n i=1
We can now proceed to minimizing this risk. However, we have to be careful.
Indeed, Rn (g) ≥ 0 for all g. Therefore any function g such that Yi = g(Xi ) for
all i = 1, . . . , n is a minimizer of the empirical risk. Yet, it may not be the best
choice (Cf. Figure 1). To overcome this limitation, we need to leverage some
prior knowledge on f : either it may belong to a certain class G of functions (e.g.,
linear functions) or it is smooth (e.g., the L2 -norm of its second derivative is
2 ERM may also mean Empirical Risk Minimizer
Introduction 6
1.5
1.5
1.5
1.0
1.0
1.0
0.5
0.5
0.5
0.0
0.0
0.0
y
y
−0.5
−0.5
−0.5
−1.0
−1.0
−1.0
−1.5
−1.5
−1.5
−2.0
−2.0
−2.0
0.0 0.2 0.4 0.6 0.8 0.0 0.2 0.4 0.6 0.8 0.0 0.2 0.4 0.6 0.8
x x x
Figure 1. It may not be the best choice idea to have fˆn (Xi ) = Yi for all i = 1, . . . , n.
small). In both cases, this extra knowledge can be incorporated to ERM using
either a constraint:
min Rn (g)
g∈G
or a penalty: n o
min Rn (g) + pen(g) ,
g
or both n o
min Rn (g) + pen(g) ,
g∈G
These schemes belong to the general idea of regularization. We will see many
variants of regularization throughout the course.
Unlike traditional (low dimensional) statistics, computation plays a key role
in high-dimensional statistics. Indeed, what is the point of describing an esti-
mator with good prediction properties if it takes years to compute it on large
datasets? As a result of this observation, much of the modern estimators, such
as the Lasso estimator for sparse linear regression can be computed efficiently
using simple tools from convex optimization. We will not describe such algo-
rithms for this problem but will comment on the computability of estimators
when relevant.
In particular computational considerations have driven the field of com-
pressed sensing that is closely connected to the problem of sparse linear regres-
sion studied in these notes. We will only briefly mention some of the results and
refer the interested reader to the book [FR13] for a comprehensive treatment.
Introduction 7
Linear models
When X = IRd , an all time favorite constraint G is the class of linear functions
that are of the form g(x) = x> θ, that is parametrized by θ ∈ IRd . Under
this constraint, the estimator obtained by ERM is usually called least squares
estimator and is defined by fˆn (x) = x> θ̂, where
n
1X
θ̂ ∈ argmin (Yi − Xi> θ)2 .
θ∈IRd n i=1
Note that θ̂ may not be unique. In the case of a linear model, where we assume
that the regression function is of the form f (x) = x> θ∗ for some unknown
θ∗ ∈ IRd , we will need assumptions to ensure identifiability if we want to prove
bounds on d(θ̂, θ∗ ) for some specific pseudo-distance d(· , ·). Nevertheless, in
other instances such as regression with fixed design, we can prove bounds on
the prediction error that are valid for any θ̂ in the argmin. In the latter case,
we will not even require that f satisfies the linear model but our bound will
be meaningful only if f can be well approximated by a linear function. In this
case, we talk about a misspecified model, i.e., we try to fit a linear model to
data that may not come from a linear model. Since linear models can have
good approximation properties especially when the dimension d is large, our
hope is that the linear model is never too far from the truth.
In the case of a misspecified model, there is no hope to drive the estimation
error d(fˆn , f ) down to zero even with a sample size that tends to infinity.
Rather, we will pay a systematic approximation error. When G is a linear
subspace as above, and the pseudo distance is given by the squared L2 norm
d(fˆn , f ) = kfˆn − f k22 , it follows from the Pythagorean theorem that
then it’s easy for the statistician to mimic it but it may be very far from the
true regression function; on the other hand, if the oracle is strong, then it is
harder to mimic but it is much closer to the truth.
Oracle inequalities were originally developed as analytic tools to prove adap-
tation of some nonparametric estimators. With the development of aggregation
[Nem00, Tsy03, Rig06] and high dimensional statistics [CT07, BRT09, RT11],
they have become important finite sample results that characterize the inter-
play between the important parameters of the problem.
In some favorable instances, that is when the Xi s enjoy specific properties,
it is even possible to estimate the vector θ accurately, as is done in parametric
statistics. The techniques employed for this goal will essentially be the same
as the ones employed to minimize the prediction risk. The extra assumptions
on the Xi s will then translate in interesting properties on θ̂ itself, including
uniqueness on top of the prediction properties of the function fˆn (x) = x> θ̂.
d
IEkfˆn − f k22 ≤ C , (1)
n
where C > 0 is a constant and in Chapter 4, we will show that this cannot
be improved apart perhaps for a smaller multiplicative constant. Clearly such
a bound is uninformative if d n and actually, in view of its optimality,
we can even conclude that the problem is too difficult statistically. However,
the situation is not hopeless if we assume that the problem has actually less
degrees of freedom than it seems. In particular, it is now standard to resort to
the sparsity assumption to overcome this limitation.
A vector θ ∈ IRd is said to be k-sparse for some k ∈ {0, . . . , d} if it has
at most k non-zero coordinates. We denote by |θ|0 the number of nonzero
coordinates of θ, which is also known as sparsity or “`0 -norm” though it is
clearly not a norm (see footnote 3). Formally, it is defined as
d
X
|θ|0 = 1I(θj 6= 0) .
j=1
Introduction 9
Sparsity is just one of many ways to limit the size of the set of potential
θ vectors to consider. One could consider vectors θ that have the following
structure for example (see Figure 2):
• Monotonic: θ1 ≥ θ2 ≥ · · · ≥ θd
• Smooth: |θi − θj | ≤ C|i − j|α for some α > 0
Pd−1
• Piecewise constant: j=1 1I(θj+1 6= θj ) ≤ k
• Structured in another basis: θ = Ψµ, for some orthogonal matrix and µ
is in one of the structured classes described above.
Sparsity plays a significant role in statistics because, often, structure trans-
lates into sparsity in a certain basis. For example, a smooth function is sparse
in the trigonometric basis and a piecewise constant function has sparse incre-
ments. Moreover, real images are approximately sparse in certain bases such as
wavelet or Fourier bases. This is precisely the feature exploited in compression
schemes such as JPEG or JPEG-2000: only a few coefficients in these images
are necessary to retain the main features of the image.
We say that θ is approximately sparse if |θ|0 may be as large as d but many
coefficients |θj | are small rather than exactly equal to zero. There are several
mathematical ways to capture this phenomena, including `q -“balls” for q ≤ 1.
For q > 0, the unit `q -ball of IRd is defined as
d
X
Bq (R) = θ ∈ IRd : |θ|qq = |θj |q ≤ 1
j=1
where |θ|q is often called `q -norm3 . As we will see, the smaller q is, the better
vectors in the unit `q ball can be approximated by sparse vectors.
Pk
Note that the set of k-sparse vectors of IRd is a union of j=0 dj linear
subspaces with dimension at most k and that are spanned by at most k vectors
in the canonical basis of IRd . If we knew that θ∗ belongs to one of these
subspaces, we could simply drop irrelevant coordinates and obtain an oracle
inequality such as (1), with d replaced by k. Since we do not know what
subspace θ∗ lives exactly, we will have to pay an extra term to find in which
subspace θ∗ lives. It turns out that this term is exactly of the order of
P
k d
log k log ed
j=0 j k
'C
n n
Therefore, the price to pay for not knowing which subspace to look at is only
a logarithmic factor.
3 Strictly speaking, |θ|q is a norm and the `q ball is a ball only for q ≥ 1.
Introduction 10
y1 Monotone Smooth
y2
x x
Constant Basis
y3
y4
x x
Nonparametric regression
Nonparametric does not mean that there is no parameter to estimate (the
regression function is a parameter) but rather that the parameter to estimate
is infinite dimensional (this is the case of a function). In some instances, this
parameter can be identified with an infinite sequence of real numbers, so that
we are still in the realm of countable infinity. Indeed, observe that since L2 (PX )
equipped with the inner product h· , ·i2 is a separable Hilbert space, it admits an
orthonormal basis {ϕk }k∈Z and any function f ∈ L2 (PX ) can be decomposed
as X
f= αk ϕk ,
k∈Z
where αk = hf, ϕk i2 .
Therefore estimating a regression function f amounts to estimating the
infinite sequence {αk }k∈Z ∈ `2 . You may argue (correctly) that the basis
{ϕk }k∈Z is also unknown as it depends on the unknown PX . This is absolutely
Introduction 11
where α̂k are some data-driven coefficients (obtained by least-squares for ex-
ample). Then by the Pythagorean theorem and Parseval’s identity, we have
We can even work further on this oracle inequality using the fact that |αk | ≤
C|k|−γ . Indeed, we have4
X X
αk2 ≤ C 2 k −2γ ≤ Ck01−2γ .
|k|>k0 |k|>k0
out these notes: a constant C may be different from line to line. This will not affect the
interpretation of our results since we are interested in the order of magnitude of the error
bounds. Nevertheless we will, as much as possible, try to make such constants explicit. As
an exercise, try to find an expression of the second C as a function of the first one and of γ.
Introduction 12
Matrix models
In the previous examples, the response variable is always assumed to be a scalar.
What if it is a higher dimensional signal? In Chapter 5, we consider various
problems of this form: matrix completion a.k.a. the Netflix problem, structured
graph estimation and covariance matrix estimation. All these problems can be
described as follows.
Let M, S and N be three matrices, respectively called observation, signal
and noise, and that satisfy
M =S+N.
Here N is a random matrix such that IE[N ] = 0, the all-zero matrix. The goal
is to estimate the signal matrix S from the observation of M .
The structure of S can also be chosen in various ways. We will consider the
case where S is sparse in the sense that it has many zero coefficients. In a way,
this assumption does not leverage much of the matrix structure and essentially
treats matrices as vectors arranged in the form of an array. This is not the case
of low rank structures where one assumes that the matrix S has either low rank
or can be well approximated by a low rank matrix. This assumption makes
sense in the case where S represents user preferences as in the Netflix example.
In this example, the (i, j)th coefficient Sij of S corresponds to the rating (on a
scale from 1 to 5) that user i gave to movie j. The low rank assumption simply
materializes the idea that there are a few canonical profiles of users and that
each user can be represented as a linear combination of these users.
At first glance, this problem seems much more difficult than sparse linear
regression. Indeed, one needs to learn not only the sparse coefficients in a given
Introduction 13
basis, but also the basis of eigenvectors. Fortunately, it turns out that the latter
task is much easier and is dominated by the former in terms of statistical price.
Another important example of matrix estimation is high-dimensional co-
variance estimation, where the goal is to estimate the covariance matrix of a
random vector X ∈ IRd , or its leading eigenvectors, based on n observations.
Such a problem has many applications including principal component analysis,
linear discriminant analysis and portfolio optimization. The main difficulty is
that n may be much smaller than the number of degrees of freedom in the
covariance matrix, which can be of order d2 . To overcome this limitation,
assumptions on the rank or the sparsity of the matrix can be leveraged.
d
IEkfˆn − f k22 > c
n
for some positive constant c. Here we used a different notation for the constant
to emphasize the fact that lower bounds guarantee optimality only up to a
constant factor. Such a lower bound on the risk is called minimax lower bound
for reasons that will become clearer in chapter 4.
How is this possible? How can we make a statement for all estimators?
We will see that these statements borrow from the theory of tests where we
know that it is impossible to drive both the type I and the type II error to
zero simultaneously (with a fixed sample size). Intuitively this phenomenon
is related to the following observation: Given n observations X1 , . . . , Xn , it is
hard to tell if they are distributed according to N (θ, 1) or to N (θ0 , 1) for a
Euclidean distance |θ − p θ0 |2 is small enough. We will see that it is the case for
0
example if |θ − θ |2 ≤ C d/n, which will yield our lower bound.
Chapter
1
Sub-Gaussian Random Variables
1 (x − µ)2
p(x) = √ exp − , x ∈ IR ,
2πσ 2 2σ 2
where µ = IE(X) ∈ IR and σ 2 = var(X) > 0 are the mean and variance of
X. We write X ∼ N (µ, σ 2 ). Note that X = σZ + µ for Z ∼ N (0, 1) (called
standard Gaussian) and where the equality holds in distribution. Clearly, this
distribution has unbounded support but it is well known that it has almost
bounded support in the following sense: IP(|X − µ| ≤ 3σ) ' 0.997. This is due
to the fast decay of the tails of p as |x| → ∞ (see Figure 1.1). This decay can
be quantified using the following proposition (Mill’s inequality).
14
1.1. Gaussian tails and MGF 15
68%
95%
99.7%
µ − 3σ µ − 2σ µ−σ µ µ+σ µ + 2σ µ + 3σ
Figure 1.1. Probabilities of falling within 1, 2, and 3 standard deviations close to the
mean in a Gaussian distribution. Source http://www.openintro.org/
and
t2
2 e− 2σ2
r
IP(|X − µ| > t) ≤ σ .
π t
Proof. Note that it is sufficient to prove the theorem for µ = 0 and σ 2 = 1 by
simple translation and rescaling. We get for Z ∼ N (0, 1),
Z ∞ x2
1
IP(Z > t) = √ exp − dx
2π t 2
Z ∞ x2
1 x
≤√ exp − dx
2π t t 2
Z ∞ x2
1 ∂
= √ − exp − dx
t 2π t ∂x 2
1
= √ exp(−t2 /2) .
t 2π
The second inequality follows from symmetry and the last one using the union
bound:
IP(|Z| > t) = IP({Z > t} ∪ {Z < −t}) ≤ IP(Z > t) + IP(Z < −t) = 2IP(Z > t) .
The fact that a Gaussian random variable Z has tails that decay to zero
exponentially fast can also be seen in the moment generating function (MGF)
M : s 7→ M (s) = IE[exp(sZ)] .
1.2. Sub-Gaussian random variables and Chernoff bounds 16
σ 2 s2
It follows that if X ∼ N (µ, σ 2 ), then IE[exp(sX)] = exp sµ + 2 .
7
6
5
y
4
3
0 5 10 15 20
x
Figure 1.2. Width of confidence intervals from exact computation in R (red dashed)
and from the approximation (1.1) (solid black). Here x = δ and y is the width of the
confidence intervals.
σ 2 if IE[X] = 0 and u> X is sub-Gaussian with variance proxy σ 2 for any vector
u ∈ S d−1 . In this case we write X ∼ subGd (σ 2 ). Note that if X ∼ subGd (σ 2 ),
then for any v such that |v|2 ≤ 1, we have v > X ∼ subG(σ 2 ). Indeed, denoting
u = v/|v|2 ∈ S d−1 , we have
> > σ 2 s2 |v|2
2 σ 2 s2
IE[esv X
] = IE[es|v|2 u X
]≤e 2 ≤e 2 .
A random matrix X ∈ IRd×T is said to be sub-Gaussian with variance proxy
σ 2 if IE[X] = 0 and u> Xv is sub-Gaussian with variance proxy σ 2 for any unit
vectors u ∈ S d−1 , v ∈ S T −1 . In this case we write X ∼ subGd×T (σ 2 ).
This property can equivalently be expressed in terms of bounds on the tail
of the random variable X.
Lemma 1.3. Let X ∼ subG(σ 2 ). Then for any t > 0, it holds
t2 t2
IP[X > t] ≤ exp − 2 , and IP[X < −t] ≤ exp − 2 . (1.3)
2σ 2σ
Proof. Assume first that X ∼ subG(σ 2 ). We will employ a very useful tech-
nique called Chernoff bound that allows to translate a bound on the moment
generating function into a tail bound. Using Markov’s inequality, we have for
any s > 0,
sX st
IE esX
IP(X > t) ≤ IP e > e ≤ .
est
1.2. Sub-Gaussian random variables and Chernoff bounds 18
inequality in this equation follows in the same manner (recall that (1.2) holds
for any s ∈ IR).
Moments
Recall that the absolute moments of Z ∼ N (0, σ 2 ) are given by
1 k + 1
IE[|Z|k ] = √ (2σ 2 )k/2 Γ
π 2
where Γ(·) denotes the Gamma function defined by
Z ∞
Γ(t) = xt−1 e−x dx , t > 0 .
0
The next lemma shows that the tail bounds of Lemma 1.3 are sufficient to
show that the absolute moments of X ∼ subG(σ 2 ) can be bounded by those of
Z ∼ N (0, σ 2 ) up to multiplicative constants.
Lemma 1.4. Let X be a random variable such that
t2
IP[|X| > t] ≤ 2 exp − 2 ,
2σ
then for any positive integer k ≥ 1,
IE[|X|k ] ≤ (2σ 2 )k/2 kΓ(k/2) .
In particular, √
IE[|X|k ])1/k ≤ σe1/e k , k ≥ 2.
√
and IE[|X|] ≤ σ 2π .
Proof.
Z ∞
IE[|X|k ] = IP(|X|k > t)dt
0
Z ∞
= IP(|X| > t1/k )dt
0
Z ∞
t2/k
≤2 e− 2σ2 dt
0
∞
t2/k
Z
= (2σ 2 )k/2 k e−u uk/2−1 du , u=
0 2σ 2
2 k/2
= (2σ ) kΓ(k/2)
1.2. Sub-Gaussian random variables and Chernoff bounds 19
The second statement follows from Γ(k/2) ≤ (k/2)k/2 and k 1/k ≤ e1/e for any
k ≥ 2. It yields
r
2 k/2
1/k 1/k 2σ 2 k √
(2σ ) kΓ(k/2) ≤k ≤ e1/e σ k .
2
√ √
Moreover, for k = 1, we have 2Γ(1/2) = 2π.
Lemma 1.5. If (1.3) holds and IE[X] = 0, then for any s > 0, it holds
2 2
IE[exp(sX)] ≤ e4σ s
.
From the above Lemma, we see that sub-Gaussian random variables can
be equivalently defined from their tail bounds and their moment generating
functions, up to constants.
1.2. Sub-Gaussian random variables and Chernoff bounds 20
If we only care about the tails, this property is preserved for sub-Gaussian
random variables.
and
n
hX i t2
IP ai Xi < −t ≤ exp −
i=1
2σ 2 |a|22
Of special interest
Pnis the case where ai = 1/n for all i. Then, we get that
the average X̄ = n1 i=1 Xi , satisfies
nt2 nt2
IP(X̄ > t) ≤ e− 2σ2 and IP(X̄ < −t) ≤ e− 2σ2
Hoeffding’s inequality
The class of sub-Gaussian random variables is actually quite large. Indeed,
Hoeffding’s lemma below implies that all random variables that are bounded
uniformly are actually sub-Gaussian with a variance proxy that depends on the
size of their support.
1.2. Sub-Gaussian random variables and Chernoff bounds 21
and 2n2 t2
IP(X̄ − IE(X̄) < −t) ≤ exp − Pn 2
.
i=1 (bi − ai )
Note that Hoeffding’s lemma holds for any bounded random variable. For
example, if one knows that X is a Rademacher random variable,
es + e−s s2
IE(esX ) = = cosh(s) ≤ e 2 .
2
Note that 2 is the best possible constant in the above approximation. For such
variables, a = −1, b = 1, and IE(X) = 0, so Hoeffding’s lemma yields
s2
IE(esX ) ≤ e 2 .
1.3. Sub-exponential random variables 22
Hoeffding’s inequality is very general but there is a price to pay for this gen-
erality. Indeed, if the random variables have small variance, we would like to
see it reflected in the exponential tail bound (as for the Gaussian case) but the
variance does not appear in Hoeffding’s inequality. We need a more refined
inequality.
In particular, the tails of this distribution do not decay as fast as the Gaussian
2
ones (that decays as e−t /2 ). Such tails are said to be heavier than Gaussian.
This tail behavior is also captured by the moment generating function of X.
Indeed, we have
1
IE esX =
if |s| < 1 ,
1 − s2
and it is not defined for s ≥ 1. It turns out that a rather weak condition on
the moment generating function is enough to partially reproduce some of the
bounds that we have proved for sub-Gaussian random variables. Observe that
for X ∼ Lap(1)
2
IE esX ≤ e2s if |s| < 1/2 ,
Lemma 1.10. Let X be a centered random variable such that IP(|X| > t) ≤
2e−2t/λ for some λ > 0. Then, for any positive integer k ≥ 1,
IE[|X|k ] ≤ λk k! .
Moreover,
IE[|X|k ])1/k ≤ 2λk ,
and the moment generating function of X satisfies
2 2 1
IE esX ≤ e2s λ ,
∀|s| ≤ .
2λ
1.3. Sub-exponential random variables 23
Proof.
Z ∞
k
IE[|X| ] = IP(|X|k > t)dt
0
Z ∞
= IP(|X| > t1/k )dt
Z0 ∞
2t1/k
≤ 2e− λ dt
0
∞
2t1/k
Z
= 2(λ/2)k k e−u uk−1 du , u=
0 λ
≤ λk kΓ(k) = λk k!
The second statement follows from Γ(k) ≤ k k and k 1/k ≤ e1/e ≤ 2 for any
k ≥ 1. It yields
1/k
λk kΓ(k) ≤ 2λk .
To control the MGF of X, we use the Taylor expansion of the exponential
function as follows. Observe that by the dominated convergence theorem, for
any s such that |s| ≤ 1/2λ
∞
X |s|k IE[|X|k ]
IE esX ≤ 1 +
k!
k=2
X∞
≤1+ (|s|λ)k
k=2
∞
X
= 1 + s2 λ2 (|s|λ)k
k=0
2 2 1
≤ 1 + 2s λ |s| ≤
2λ
2
λ2
≤ e2s
Bernstein’s inequality
Theorem 1.13 (Bernstein’s inequality). Let X1 , . . . , Xn be independent ran-
dom variables such that IE(Xi ) = 0 and Xi ∼ subE(λ). Define
n
1X
X̄ = Xi ,
n i=1
n t2
t
IP(X̄ > t) ∨ IP(X̄ < −t) ≤ exp − ( 2 ∧ ) .
2 λ λ
Proof. Without loss of generality, assume that λ = 1 (we can always replace
Xi by Xi /λ and t by t/λ). Next, using a Chernoff bound, we get for any s > 0
n
Y
IE esXi e−snt .
IP(X̄ > t) ≤
i=1
1.4. Maximal inequalities 25
2
Next, if |s| ≤ 1, then IE esXi ≤ es /2 by definition of sub-exponential distri-
butions. It yields
ns2
−snt
IP(X̄ > t) ≤ e 2
Choosing s = 1 ∧ t yields
n 2
IP(X̄ > t) ≤ e− 2 (t ∧t)
We obtain the same bound for IP(X̄ < −t) which concludes the proof.
The exponential inequalities of the previous section are valid for linear com-
binations of independent random variables, and, in particular, for the average
X̄. In many instances, we will be interested in controlling the maximum over
the parameters of such linear combinations (this is because of empirical risk
minimization). The purpose of this section is to present such results.
Note that the random variables in this theorem need not be independent.
1.4. Maximal inequalities 26
log N σ2 s
= + .
s 2
p
Taking s = 2(log N )/σ 2 yields the first inequality in expectation.
The first inequality in probability is obtained by a simple union bound:
[
IP max Xi > t = IP {Xi > t}
1≤i≤N
1≤i≤N
X
≤ IP(Xi > t)
1≤i≤N
t2
≤ N e− 2σ2 ,
On the opposite side of the picture, if all the Xi s are equal to the same random
variable X, we have for any t > 0,
In the Gaussian case, lower bounds are also available. They illustrate the effect
of the correlation between the Xi s.
1.4. Maximal inequalities 27
It yields
max c> x ≤ max c> x ≤ max c> x
x∈P x∈V(P) x∈P
n d
X o
B1 = x ∈ IRd : |xi | ≤ 1 .
i=1
n d
X o
B2 = x ∈ IRd : x2i ≤ 1 .
i=1
Clearly, this ball is not a polytope and yet, we can control the maximum of
random variables indexed by B2 . This is due to the fact that there exists a
finite subset of B2 such that the maximum over this finite set is of the same
order as the maximum over the entire ball.
Lemma 1.18. Fix ε ∈ (0, 1). Then the unit Euclidean ball B2 has an ε-net N
with respect to the Euclidean distance of cardinality |N | ≤ (3/ε)d
This is equivalent to
ε d
ε d
≥ |N |
1+ .
2 2
Therefore, we get the following bound
2 d 3 d
|N | ≤ 1 + ≤ .
ε ε
But
1
max x> X = max x> X
1
x∈ 2 B2 2 x∈B2
Therefore, using Theorem 1.14, we get
p p √
IE[max θ> X] ≤ 2IE[max z > X] ≤ 2σ 2 log(|N |) ≤ 2σ 2(log 6)d ≤ 4σ d .
θ∈B2 z∈N
Preliminaries
We denote the set of symmetric, positive semi-definite, and positive definite
matrices in IRd×d by Sd , Sd+ , and Sd++ , respectively, and will omit the subscript
when the dimensionality is clear. Here, positive semi-definite means that for
a matrix X ∈ S + , v > Xv ≥ 0 for all v ∈ IRd , |v|2 = 1, and positive definite
means that the equality holds strictly. This is equivalent to all eigenvalues of
X being larger or equal (strictly larger, respectively) than 0, λj (X) ≥ 0 for all
j = 1, . . . , d.
The cone of positive definite matrices induces an order on matrices by set-
ting A B if B − A ∈ S + .
Since we want to extend the notion of exponentiating random variables that
was essential to our derivation of the Chernoff bound to matrices, we will make
use of the matrix exponential and the matrix logarithm. For a symmetric matrix
X, we can define a functional calculus by applying a function f : IR → IR to
the diagonal elements of its spectral decomposition, i.e., if X = QΛQ> , then
where
[f (Λ)]i,j = f (Λi,j ), i, j ∈ [d], (1.5)
is only non-zero on the diagonal and f (X) is well-defined because the spectral
decomposition is unique up to the ordering of the eigenvalues and the basis of
the eigenspaces, with respect to both of which (1.4) is invariant, as well. From
the definition, it is clear that for the spectrum of the resulting matrix,
Similarly, we can define the matrix logarithm as the inverse function of exp on
S, log(eA ) = A, which defines it on S + .
Some nice properties of these functions on matrices are their monotonicity:
For A B,
Tr exp A ≤ Tr exp B, (1.8)
and for 0 ≺ A B,
log A log B. (1.9)
Note that the analogue of (1.9) is in general not true for the matrix exponential.
Proof. We multiply both sides of the inequality λmax (Y) ≥ t by θ, take ex-
ponentials, apply the spectral theorem (1.6) and then estimate the maximum
eigenvalue by the sum over all eigenvalues, the trace.
Recall that the crucial step in proving Bernstein’s and Hoeffding’s inequal-
ity was to exploit the independence of the summands by the fact that the
exponential function turns products into sums.
P Y Y
IE[e i θXi ] = IE[ eθXi ] = IE[eθXi ].
i i
1.5. Sums of independent random matrices 32
This property, eA+B = eA eB , no longer holds true for matrices, unless they
commute.
We could try to replace it with a similar property, the Golden-Thompson
inequality,
Tr[eθ(X1 +X2 ) ] ≤ Tr[eθX1 eθX2 ].
Unfortunately, this does not generalize to more than two matrices, and when
trying to peel off factors, we would have to pull a maximum eigenvalue out of
the trace,
Tr[eθX1 eθX2 ] ≤ λmax (eθX2 )Tr[eθX1 ].
This is the approach followed by Ahlswede-Winter [AW02], which leads to
worse constants for concentration than the ones we obtain below.
Instead, we are going to use the following deep theorem due to Lieb [Lie73].
A sketch of the proof can be found in the appendix of [Rus02].
is a concave function.
IE[TreH+X ] = IE[TreH+log Y ]
≤ TreH+log(IE[Y])
X
= TreH+log IE[e ]
With this, we can establish a better bound on the moment generating func-
tion of sums of independent matrices.
Proof. Without loss of generality, we can assume θ = 1. Write IEk for the
expectation conditioned on X1 , . . . , Xk , IEk [ . ] = IE[ . |X1 , . . . , Xk ]. By the
1.5. Sums of independent random matrices 33
tower property,
Xn Xn
IE[Tr exp( Xi )] = IE0 . . . IEn−1 Tr exp( Xi )
i=1 i=1
n−1
X
Xi + log IEn−1 eXn
≤ IE0 . . . IEn−2 Tr exp
| {z }
i=1
=IEeXn
..
.
Xn
≤ Tr exp( log IE[eXi ]),
i=1
Proof. Using the operator monoticity of log, (1.9), and the monotonicity of
Tr exp, (1.8), we can plug the estimates for the matrix mgfs into the master
inequality, Theorem 1.24, to obtain
X X
IP(λmax ( Xi ) ≥ t) ≤ e−θt Tr exp(g(θ) Ai )
i i
X
−θt
≤ de λmax (exp(g(θ) Ai ))
i
X
= de−θt exp(g(θ)λmax ( Ai )),
i
Pd
where we estimated Tr(X) = j=1 λj ≤ dλmax and used the spectral theorem.
1.5. Sums of independent random matrices 34
Proof. Define θx
e − θx − 1
, x 6= 0,
x2
f (x) =
θ2
, x = 0,
2
and verify that this is a smooth and increasing function on IR. Hence, f (x) ≤
f (1) for x ≤ 1. By the transfer rule, (1.7), f (X) f (I) = f (1)I. Therefore,
By Corollary 1.25,
X
IP(λmax ( Xi ) ≥ t) ≤ d exp(−θt + g(θ)σ 2 ).
i
1.5. Sums of independent random matrices 35
With similar techniques, one can also prove a version of Hoeffding’s inequal-
ity for matrices, see [Tro12, Theorem 1.3].
1.6. Problem set 36
t2
t
IP(|S(a)| > t) ≤ 2 exp −C ∧ .
λ2 |a|22 λ|a|∞
Problem 1.4. Let A = {Ai,j } 1≤i≤n be a random matrix such that its entries
1≤j≤m
are i.i.d. sub-Gaussian random variables with variance proxy σ 2 .
(a) Show that the matrix A is sub-Gaussian. What is its variance proxy?
(b) Let kAk denote the operator norm of A defined by
|Ax|2
maxm .
x∈IR |x|2
Show that there exits a constant C > 0 such that
√ √
IEkAk ≤ C( m + n) .
Problem 1.5. Recall that for any q ≥ 1, the `q norm of a vector x ∈ IRn is
defined by
Xn q1
|x|q = |xi |q .
i=1
Problem 1.6. Let K be a compact subset of the unit sphere of IRp that
admits an ε-net Nε with respect to the Euclidean distance of IRp that satisfies
|Nε | ≤ (C/ε)d for all ε ∈ (0, 1). Here C ≥ 1 and d ≤ p are positive constants.
Let X ∼ subGp (σ 2 ) be a centered random vector.
Show that there exists positive constants c1 and c2 to be made explicit such
that for any δ ∈ (0, 1), it holds
p p
max θ> X ≤ c1 σ d log(2p/d) + c2 σ log(1/δ)
θ∈K
with probability at least 1−δ. Comment on the result in light of Theorem 1.19 .
1.6. Problem set 38
Problem 1.7. For any K ⊂ IRd , distance d on IRd and ε > 0, the ε-covering
number C(ε) of K is the cardinality of the smallest ε-net of K. The ε-packing
number P (ε) of K is the cardinality of the largest set P ⊂ K such that
d(z, z 0 ) > ε for all z, z 0 ∈ P, z 6= z 0 . Show that
2
(Hint: Show that IE[eλX ] = √ 1 for λ < 1
if X ∼ N (0, 1).)
1−2λ 2
6. Show that for a random projection operator Q ∈ IRk×d and a fixed vector
x ∈ IRd ,
a) IE[kQxk2 ] = kd kxk2 .
b) For ε ∈ (0, 1), there is a constant c > 0 such that with probability
at least 1 − 2 exp(−ckε2 ),
k k
(1 − ε) kxk2 ≤ kQxk22 ≤ (1 + ε) kxk22 .
d d
(Hint: Think about how to apply the previous results in this case
and use the inequalities log(1 − ε) ≤ −ε − ε2 /2 and log(1 + ε) ≤
ε − ε2 /2 + ε3 /3.)
7. Prove Theorem 1.28.
Chapter
2
Linear Regression Model
Yi = f (Xi ) + εi , i = 1, . . . , n , (2.1)
where ε = (ε1 , . . . , εn )> is sub-Gaussian with variance proxy σ 2 and such that
IE[ε] = 0. Our goal is to estimate the function f under a linear assumption.
Namely, we assume that x ∈ IRd and f (x) = x> θ∗ for some unknown θ∗ ∈ IRd .
Random design
The case of random design corresponds to the statistical learning setup. Let
(X1 , Y1 ), . . . , (Xn+1 , Yn+1 ) be n + 1 i.i.d. random couples. Given the pairs
(X1 , Y1 ), . . . , (Xn , Yn ), the goal is construct a function fˆn such that fˆn (Xn+1 )
is a good predictor of Yn+1 . Note that when fˆn is constructed, Xn+1 is still
unknown and we have to account for what value it is likely to take.
Consider the following example from [HTF01, Section 3.2]. The response
variable Y is the log-volume of a cancerous tumor, and the goal is to predict
it based on X ∈ IR6 , a collection of variables that are easier to measure (age
of patient, log-weight of prostate, . . . ). Here the goal is clearly to construct f
for prediction purposes. Indeed, we want to find an automatic mechanism that
40
2.1. Fixed design linear regression 41
outputs a good prediction of the log-weight of the tumor given certain inputs
for a new (unseen) patient.
A natural measure of performance here is the L2 -risk employed in the in-
troduction:
R(fˆn ) = IE[Yn+1 − fˆn (Xn+1 )]2 = IE[Yn+1 − f (Xn+1 )]2 + kfˆn − f k2L2 (PX ) ,
Fixed design
In fixed design, the points (or vectors) X1 , . . . , Xn are deterministic. To em-
phasize this fact, we use lowercase letters x1 , . . . , xn to denote fixed design. Of
course, we can always think of them as realizations of a random variable but
the distinction between fixed and random design is deeper and significantly
affects our measure of performance. Indeed, recall that for random design, we
look at the performance in average over realizations of Xn+1 . Here, there is no
such thing as a marginal distribution of Xn+1 . Rather, since the design points
x1 , . . . , xn are considered deterministic, our goal is estimate f only at these
points. This problem is sometimes called denoising since our goal is to recover
f (x1 ), . . . , f (xn ) given noisy observations of these values.
In many instances, fixed designs exhibit particular structures. A typical
example is the regular design on [0, 1], given by xi = i/n, i = 1, . . . , n. Inter-
polation between these points is possible under smoothness assumptions.
>
Note that in fixed design, we observe µ∗ +ε, where µ∗ = f (x1 ), . . . , f (xn ) ∈
IRn and ε = (ε1 , . . . , εn )> ∈ IRn is sub-Gaussian with variance proxy σ 2 . In-
stead of a functional estimation problem, it is often simpler to view this problem
as a vector problem in IRn . This point of view will allow us to leverage the
Euclidean geometry of IRn .
In the case of fixed design, we will focus on the Mean Squared Error (MSE)
as a measure of performance. It is defined by
n
1X ˆ 2
MSE(fˆn ) = fn (xi ) − f (xi ) .
n i=1
1 X> X
MSE(Xθ̂) = |X(θ̂ − θ∗ )|22 = (θ̂ − θ∗ )> (θ̂ − θ∗ ) . (2.3)
n n
A natural example of fixed design regression is image denoising. Assume
that µ∗i , i ∈ 1, . . . , n is the grayscale value of pixel i of an image. We do not
get to observe the image µ∗ but rather a noisy version of it Y = µ∗ + ε. Given
a library of d images {x1 , . . . , xd }, xj ∈ IRn , our goal is to recover the original
image µ∗ using linear combinations of the images x1 , . . . , xd . This can be done
fairly accurately (see Figure 2.1).
Figure 2.1. Reconstruction of the digit “6”: Original (left), Noisy (middle) and Recon-
struction (right). Here n = 16 × 16 = 256 pixels. Source [RT11].
As we will see in Remark 2.3, properly choosing the design also ensures that
if MSE(fˆ) is small for some linear estimator fˆ(x) = x> θ̂, then |θ̂ − θ∗ |22 is also
small.
Throughout this section, we consider the regression model (2.2) with fixed
design.
Note that we are interested in estimating Xθ∗ and not θ∗ itself, so by exten-
sion, we also call µ̂ls = Xθ̂ls least squares estimator. Observe that µ̂ls is the
projection of Y onto the column span of X.
It is not hard to see that least squares estimators of θ∗ and µ∗ = Xθ∗ are
maximum likelihood estimators when ε ∼ N (0, σ 2 In ).
Proposition 2.1. The least squares estimator µ̂ls = Xθ̂ls ∈ IRn satisfies
∇θ |Y − Xθ|22 = 0,
∇θ |Y − Xθ|22 = ∇θ |Y |22 − 2Y > Xθ + θ> X> Xθ = −2(Y > X − θ> X> X)> .
X> Xθ = X> Y .
It concludes the proof of the first statement. The second statement follows
from the definition of the Moore-Penrose pseudoinverse.
We are now going to prove our first result on the finite sample performance
of the least squares estimator for fixed design.
Theorem 2.2. Assume that the linear model (2.2) holds where ε ∼ subGn (σ 2 ).
Then the least squares estimator θ̂ls satisfies
1 r
IE MSE(Xθ̂ls ) = IE|Xθ̂ls − Xθ∗ |22 . σ 2 ,
n n
where r = rank(X> X). Moreover, for any δ > 0, with probability at least 1 − δ,
it holds
r + log(1/δ)
MSE(Xθ̂ls ) . σ 2 .
n
Proof. Note that by definition
Moreover,
|Y − Xθ̂ls |22 = |Xθ∗ + ε − Xθ̂ls |22 = |Xθ̂ls − Xθ∗ |22 − 2ε> X(θ̂ls − θ∗ ) + |ε|22 .
2.2. Least squares estimators 44
Therefore, we get
ε> X(θ̂ls − θ∗ )
|Xθ̂ls − Xθ∗ |22 ≤ 2ε> X(θ̂ls − θ∗ ) = 2|Xθ̂ls − Xθ∗ |2 (2.5)
|X(θ̂ls − θ∗ )|2
Note that it is difficult to control
ε> X(θ̂ls − θ∗ )
|X(θ̂ls − θ∗ )|2
as θ̂ls depends on ε and the dependence structure of this term may be compli-
cated. To remove this dependency, a traditional technique is to “sup-out” θ̂ls .
This is typically where maximal inequalities are needed. Here we have to be a
bit careful.
Let Φ = [φ1 , . . . , φr ] ∈ IRn×r be an orthonormal basis of the column span
of X. In particular, there exists ν ∈ IRr such that X(θ̂ls − θ∗ ) = Φν. It yields
ε> X(θ̂ls − θ∗ ) ε> Φν ε> Φν ν
= = = ε̃> ≤ sup ε̃> u ,
|X(θ̂ls − θ∗ )|2 |Φν|2 |ν|2 |ν|2 u∈B2
Moreover, with probability 1 − δ, it follows from the last step in the proof1 of
Theorem 1.19 that
sup (ε̃> u)2 ≤ 8 log(6)σ 2 r + 8σ 2 log(1/δ) .
u∈B2
X> X
Remark 2.3. If d ≤ n and B := n has rank d, then we have
MSE(Xθ̂ls )
|θ̂ls − θ∗ |22 ≤ ,
λmin (B)
and we can use Theorem 2.2 to bound |θ̂ls − θ∗ |22 directly. By contrast, in the
high-dimensional case, we will need more structural assumptions to come to a
similar conclusion.
1 we could use Theorem 1.19 directly here but at the cost of a factor 2 in the constant.
2.2. Least squares estimators 45
The fundamental inequality (2.4) would still hold and the bounds on the MSE
may be smaller. Indeed, (2.5) can be replaced by
ls
|Xθ̂K − Xθ∗ |22 ≤ 2ε> X(θ̂K
ls
− θ∗ ) ≤ 2 sup (ε> Xθ) ,
θ∈K−K
and it has exactly 2d vertices V = {e1 , −e1 , . . . , ed , −ed }, where ej is the j-th
vector of the canonical basis of IRd and is defined by
ej = (0, . . . , 0, 1
|{z} , 0, . . . , 0)> .
jth position
Observe now that since ε ∼ subGn (σ 2 ), for any column Xj such that |Xj |2 ≤
√
n, the random variable ε> Xj ∼ subG(nσ 2 ). Therefore, applying Theo-
rem 1.16, we get the bound on IE MSE(Xθ̂K ls
) and for any t ≥ 0,
nt2
) > t ≤ IP sup (ε> v) > nt/4 ≤ 2de− 32σ2
ls
IP MSE(Xθ̂K
v∈XK
Note that the proof of Theorem 2.2 also applies to θ̂Bls1 (exercise!) so that
Xθ̂Bls1
benefits from the best of both rates,
r r
log d
IE MSE(Xθ̂Bls1 ) . min σ 2 , σ
.
n n
√
This is called an elbow effect. The elbow takes place around r ' n (up to
logarithmic terms).
Remark 2.5. The `0 terminology and notation comes from the fact that
d
X
lim |θj |q = |θ|0
q→0+
j=1
Therefore it is really limq→0+ |θ|qq but the notation |θ|00 suggests too much that
it is always equal to 1.
By extension, denote by B0 (k) the `0 ball of IRd , i.e., the set of k-sparse
vectors, defined by
B0 (k) = {θ ∈ IRd : |θ|0 ≤ k} .
ls
In this section, our goal is to control the MSE of θ̂K when K = B0 (k). Note that
d
ls
computing θ̂B0 (k) essentially requires computing k least squares estimators,
which is an exponential number in k. In practice this will be hard (or even
impossible) but it is interesting to understand the statistical properties of this
estimator and to use them as a benchmark.
Theorem 2.6. Fix a positive integer k ≤ d/2. Let K = B0 (k) be set of
k-sparse vectors of IRd and assume that θ∗ ∈ B0 (k). Moreover, assume the
conditions of Theorem 2.2. Then, for any δ > 0, with probability 1 − δ, it holds
σ2 σ2 k σ2
ls d
MSE(Xθ̂B0 (k) ) . log + + log(1/δ) .
n 2k n n
ε> X(θ̂K
ls
− θ∗ )
ls
|Xθ̂K − Xθ∗ |22 ≤ 2ε> X(θ̂K
ls
− θ∗ ) = 2|Xθ̂K
ls
− Xθ∗ |2 .
ls − θ ∗ )|
|X(θ̂K 2
Therefore,
ε> X(θ̂K
ls
− θ∗ ) ε> ΦŜ ν
= ≤ max sup [ε> ΦS ]u
ls − θ ∗ )|
|X(θ̂K 2 |ν|2 |S|=2k u∈BrS
2
where B2rS rS
is the unit ball of IR . It yields
ls
|Xθ̂K − Xθ∗ |22 ≤ 4 max sup (ε̃> 2
S u) ,
|S|=2k u∈BrS
2
2.2. Least squares estimators 48
It follows from the proof of Theorem 1.19 that for any |S| ≤ 2k,
t t
sup (ε̃> u)2 > t ≤ 6|S| e− 8σ2 ≤ 62k e− 8σ2 .
IP
r
u∈B2S
d
How large is log 2k ? It turns out that it is not much larger than k.
Lemma 2.7. For any integers 1 ≤ k ≤ n, it holds
n en k
≤ .
k k
Corollary 2.8. Under the assumptions of Theorem 2.6, for any δ > 0, with
probability at least 1 − δ, it holds
σ2 k ed σ 2 k σ2
MSE(Xθ̂Bls0 (k) ) . log + log(6) + log(1/δ) .
n 2k n n
Note that for any fixed δ, there exits a constant Cδ > 0 such that for any
n ≥ 2k, with high probability,
σ2 k ed
MSE(Xθ̂Bls0 (k) ) ≤ Cδ log .
n 2k
Comparing this result with Theorem 2.2 with r = k, we see that the price to
pay for not knowing the support of θ∗ but only its size, is a logarithmic factor
in the dimension d.
This result immediately leads the following bound in expectation.
Corollary 2.9. Under the assumptions of Theorem 2.6,
σ2 k ed
IE MSE(Xθ̂Bls0 (k) ) . log .
n k
Proof. It follows from (2.6) that for any H ≥ 0,
Z ∞
− Xθ∗ |22 > nu)du
IE MSE(Xθ̂Bls0 (k) ) = ls
IP(|Xθ̂K
0
Z ∞
≤H+ ls
IP(|Xθ̂K − Xθ∗ |22 > n(u + H))du
0
2k Z ∞
X d n(u+H)
≤H+ 62k
e− 32σ 2 du
j=1
j 0
2k
X d nH 32σ 2
=H+ 62k e− 32σ2 .
j=1
j n
This yields
σ2 k ed
H. log ,
n k
which completes the proof.
The Gaussian Sequence Model is a toy model that has received a lot of
attention, mostly in the eighties. The main reason for its popularity is that
2.3. The Gaussian Sequence Model 50
y = θ∗ + ξ ∈ IRd , (2.9)
2
where ξ ∼ subGd (σ /n).
In this section, we can actually completely “forget” about our original
model (2.2). In particular we can define this model independently of Assump-
tion ORT and thus for any values of n and d.
The sub-Gaussian sequence model and the Gaussian sequence model are
called direct (observation) problems as opposed to inverse problems where the
goal is to estimate the parameter θ∗ only from noisy observations of its image
through an operator. The linear regression model is one such inverse problem
where the matrix X plays the role of a linear operator. However, in these notes,
we never try to invert the operator. See [Cav11] for an interesting survey on
the statistical theory of inverse problems.
σ2 k ed
MSE(Xθ̂Bls0 (k) ) ≤ Cδ log .
n 2k
As we will see, the assumption ORT gives us the luxury to not know k and yet
adapt to its value. Adaptation means that we can construct an estimator that
does not require the knowledge of k (the smallest such that |θ∗ |0 ≤ k) and yet
perform as well as θ̂Bls0 (k) , up to a multiplicative constant.
Let us begin with some heuristic considerations to gain some intuition.
Assume the sub-Gaussian sequence model (2.9). If nothing is known about θ∗
2.3. The Gaussian Sequence Model 52
supp(θ̂hrd ) = supp(θ∗ ) .
and recall that Theorem 1.14 yields IP(A) ≥ 1 − δ. On the event A, the
following holds for any j = 1, . . . , d.
First, observe that
|yj | > 2τ ⇒ |θj∗ | ≥ |yj | − |ξj | > τ (2.12)
and
|yj | ≤ 2τ ⇒ |θj∗ | ≤ |yj | + |ξj | ≤ 3τ. (2.13)
It yields
|θ̂jhrd − θj∗ | = |yj − θj∗ |1I(|yj | > 2τ ) + |θj∗ |1I(|yj | ≤ 2τ )
≤ τ 1I(|yj | > 2τ ) + |θj∗ |1I(|yj | ≤ 2τ )
≤ τ 1I(|θj∗ | > τ ) + |θj∗ |1I(|θj∗ | ≤ 3τ ) by (2.12) and (2.13)
≤ 4 min(|θj∗ |, τ )
It yields
d
X d
X
|θ̂hrd − θ∗ |22 = |θ̂jhrd − θj∗ |2 ≤ 16 min(|θj∗ |2 , τ 2 ) ≤ 16|θ∗ |0 τ 2 .
j=1 j=1
Therefore, |θj∗ | =
6 0 and supp(θ̂hrd ) ⊂ supp(θ∗ ).
Similar results can be obtained for the soft thresholding estimator θ̂sft
defined by
yj − 2τ if yj > 2τ ,
θ̂jsft = yj + 2τ if yj < −2τ ,
0 if |yj | ≤ 2τ ,
Hard Soft
2
2
1
1
0
0
y
y
−2 −1
−2 −1
−2 −1 0 1 2 −2 −1 0 1 2
x x
Figure 2.2. Transformation applied to yj with 2τ = 1 to obtain the hard (left) and soft
(right) thresholding estimators
In view of (2.8), under the assumption ORT, the above variational definitions
can be written as
n1 o
θ̂hrd = argmin |Y − Xθ|22 + 4τ 2 |θ|0
θ∈IRd n
n1 o
θ̂sft = argmin |Y − Xθ|22 + 4τ |θ|1
θ∈IRd n
When the assumption ORT is not satisfied, they no longer correspond to thresh-
olding estimators but can still be defined as above. We change the constant in
the threshold parameters for future convenience.
Definition 2.12. Fix τ > 0 and assume the linear regression model (2.2). The
2.4. High-dimensional linear regression 55
4. Recently there has been a lot of interest around this objective for very
large d and very large n. In this case, even computing |Y − Xθ|22 may
be computationally expensive and solutions based on stochastic gradient
descent are flourishing.
Theorem 2.14. Assume that the linear model (2.2) holds, where ε ∼ subGn (σ 2 )
and that |θ∗ |0 ≥ 1. Then, the BIC estimator θ̂bic with regularization parameter
σ2 σ 2 log(ed)
τ 2 = 16 log(6) + 32 . (2.14)
n n
satisfies
1 log(ed/δ)
MSE(Xθ̂bic ) = |Xθ̂bic − Xθ∗ |22 . |θ∗ |0 σ 2 (2.15)
n n
with probability at least 1 − δ.
where we use the inequality 2ab ≤ 2a2 + 12 b2 . Together with the previous
display, it yields
2
|Xθ̂bic − Xθ∗ |22 ≤ 2nτ 2 |θ∗ |0 + 4 ε> U(θ̂bic − θ∗ ) − 2nτ 2 |θ̂bic |0
(2.16)
2.4. High-dimensional linear regression 57
where
z
U(z) =
|z|2
Next, we need to “sup out” θ̂bic . To that end, we decompose the sup into a
max over cardinalities as follows:
Moreover, using the ε-net argument from Theorem 1.19, we get for |S| = k,
t
2 t 1 + 1 nτ 2 k
sup ε> ΦS,∗ u ≥ + nτ 2 k ≤ 2 · 6rS,∗ exp − 4 2 2
IP
r
u∈B2S,∗
4 2 8σ
t nτ 2 k
≤ 2 exp − 2
− + (k + |θ∗ |0 ) log(6)
32σ 16σ 2
t ∗
≤ exp − − 2k log(ed) + |θ |0 log(12)
32σ 2
where, in the last inequality, we used the definition (2.14) of τ .
2.4. High-dimensional linear regression 58
It follows from Theorem 2.14 that θ̂bic adapts to the unknown sparsity of
∗
θ , just like θ̂hrd . Moreover, this holds under no assumption on the design
matrix X.
Theorem 2.15. Assume that the linear model (2.2) holds where ε ∼ subGn (σ 2 ).
Moreover, assume
√ that the columns of X are normalized in such a way that
maxj |Xj |2 ≤ n. Then, the Lasso estimator θ̂L with regularization parameter
r r
2 log(2d) 2 log(1/δ)
2τ = 2σ + 2σ (2.17)
n n
satisfies
r r
L 1 L ∗ 2 ∗ 2 log(2d) ∗ 2 log(1/δ)
MSE(Xθ̂ ) = |Xθ̂ − Xθ |2 ≤ 4|θ |1 σ + 4|θ |1 σ
n n n
with probability at least 1 − δ.
Proof. From the definition of θ̂L , it holds
1 1
|Y − Xθ̂L |22 + 2τ |θ̂L |1 ≤ |Y − Xθ∗ |22 + 2τ |θ∗ |1 .
n n
Using Hölder’s inequality, it implies
|Xθ̂L − Xθ∗ |22 ≤ 2ε> X(θ̂L − θ∗ ) + 2nτ |θ∗ |1 − |θ̂L |1
≤ 2|X> ε|∞ |θ̂L |1 − 2nτ |θ̂L |1 + 2|X> ε|∞ |θ∗ |1 + 2nτ |θ∗ |1
= 2(|X> ε|∞ − nτ )|θ̂L |1 + 2(|X> ε|∞ + nτ )|θ∗ |1
Observe now that for any t > 0,
t2
− 2nσ2
IP(|X> ε|∞ ≥ t) = IP( max |X>
j ε| > t) ≤ 2de
1≤j≤d
p p
Therefore, taking t = σ 2n log(2d) + σ 2n log(1/δ) = nτ , we get that with
probability at least 1 − δ,
|Xθ̂L − Xθ∗ |22 ≤ 4nτ |θ∗ |1 .
Notice that the regularization parameter (2.17) depends on the confidence
level δ. This not the case for the BIC estimator (see (2.14)).
p
The rate in Theorem 2.15 is of order (log d)/n (slow rate), which is much
slower than the rate of order (log d)/n (fast rate) for the BIC estimator. Here-
after, we show that fast rates can also be achieved by the computationally
efficient Lasso estimator, but at the cost of a much stronger condition on the
design matrix X.
Incoherence
Assumption INC(k) We say that the design matrix X has incoherence k for
some integer k > 0 if
X> X 1
− Id ∞ ≤
n 32k
where the |A|∞ denotes the largest element of A in absolute value. Equivalently,
2.4. High-dimensional linear regression 60
1. For all j = 1, . . . , d,
|Xj |22 1
−1 ≤ .
n 32k
2. For all 1 ≤ i, j ≤ d, i 6= j, we have
|X>
i Xj | 1
≤ .
n 32k
It implies that there exists matrices that satisfy Assumption INC(k) for
n ≥ Ck 2 log(d) ,
Proof. Let εij ∈ {−1, 1} denote the Rademacher random variable that is on
the ith row and jth column of X.
Note first that the jth diagonal entries of X> X/n are given by
n
1X 2
ε = 1.
n i=1 i,j
Moreover, for j 6= k, the (j, k)th entry of the d × d matrix X> X/n is given by
n n
1X 1 X (j,k)
εi,j εi,k = ξ ,
n i=1 n i=1 i
X> X 1 n
IP − Id ∞
> ≤ 2d2 e− 211 k2 ≤ δ
n 32k
for
n ≥ 211 k 2 log(1/δ) + 213 k 2 log(d) .
For any θ ∈ IRd , S ⊂ {1, . . . , d}, define θS to be the vector with coordinates
θj if j ∈ S ,
θS,j =
0 otherwise .
Lemma 2.17. Fix a positive integer k ≤ d and assume that X satisfies as-
sumption INC(k). Then, for any S ∈ {1, . . . , d} such that |S| ≤ k and any
θ ∈ IRd that satisfies the cone condition
it holds
|Xθ|22
|θ|22 ≤ 2 .
n
Proof. We have
X> X
>
|XθS |22 |θS |21
X X
= θS> θS = |θS |22 + θS> − Id θS ≥ |θS |22 − ,
n n n 32k
2.4. High-dimensional linear regression 62
(ii) Similarly,
|XθS c |22 |θS c |21 9|θS |21
≥ |θS c |22 − ≥ |θS c |22 − ,
n 32k 32k
where, in the last inequality, we used the cone condition (2.18).
(iii) Finally,
X> X 2 6
2 θS> θS c ≤ |θS |1 |θS c |1 ≤ |θS |21 .
n 32k 32k
where, in the last inequality, we used the cone condition (2.18).
Observe now that it follows from the Cauchy-Schwarz inequality that
|θS |21 ≤ |S||θS |22 .
Thus for |S| ≤ k,
|Xθ|22 16|S| 1
≥ |θS |22 + |θS c |22 − |θS |22 ≥ |θ|22 .
n 32k 2
|Xθ̂L −Xθ∗ |22 +nτ |θ̂L −θ∗ |1 ≤ 2ε> X(θ̂L −θ∗ )+nτ |θ̂L −θ∗ |1 +2nτ |θ∗ |1 −2nτ |θ̂L |1 .
Applying Hölder’s inequality and using the same steps as in the proof of The-
orem 2.15, we get that with probability 1 − δ, we get
ε> X(θ̂L − θ∗ ) ≤ |ε> X|∞ |θ̂L − θ∗ |1
nτ L
≤ |θ̂ − θ∗ |1 ,
2
2.4. High-dimensional linear regression 63
where we used the fact that |Xj |22 ≤ n + 1/(32k) ≤ 2n. Therefore, taking
S = supp(θ∗ ) to be the support of θ∗ , we get
|Xθ̂L − Xθ∗ |22 + nτ |θ̂L − θ∗ |1 ≤ 2nτ |θ̂L − θ∗ |1 + 2nτ |θ∗ |1 − 2nτ |θ̂L |1
= 2nτ |θ̂SL − θ∗ |1 + 2nτ |θ∗ |1 − 2nτ |θ̂SL |1
≤ 4nτ |θ̂SL − θ∗ |1 . (2.21)
so that θ = θ̂L − θ∗ satisfies the cone condition (2.18). Using now the Cauchy-
Schwarz inequality and Lemma 2.17 respectively, we get, since |S| ≤ k,
r
L ∗
p L ∗
p L ∗ 2k
|θ̂S − θ |1 ≤ |S||θ̂S − θ |2 ≤ |S||θ̂ − θ |2 ≤ |Xθ̂L − Xθ∗ |2 .
n
Combining this result with (2.21), we find
This concludes the proof of the bound on the MSE. To prove (2.20), we use
Lemma 2.17 once again to get
Note that all we required for the proof was not really incoherence but the
conclusion of Lemma 2.17:
|Xθ|22
inf inf ≥ κ, (2.23)
|S|≤k θ∈CS n|θ|22
or equivalently as
d
X
|θ|∗ = max λj |θφ(j) |. (2.25)
φ∈Sd
j=1
With this choice, we will exhibit a scaling in τ that leads to the desired high
probability bounds, following the proofs in [BLT16].
We begin with refined bounds on the suprema of Gaussians.
Lemma 2.21. Define [d] := {1, . . . , d}. Under the same assumptions as in
Lemma 2.20,
(g ∗ ) p
sup k ≤ 4 log(1/δ), (2.29)
k∈[d] σλk
For t > 8,
! d 1−3t/8
(gk∗ )2 X 2d
IP sup 2 2 > t ≤ (2.32)
k∈[d] σ λk j=1
j
d
X 1
≤ (2d)1−3t/8 2
(2.33)
j=1
j
≤ 4 · 2−3t/8 . (2.34)
for δ ≤ 1/2.
2.4. High-dimensional linear regression 66
Theorem 2.22. Fix n ≥ 2. Assume that the linear model (2.2) holds where
ε ∼ Nn (0, σ 2 In ). Moreover, assume that |θ∗ |0 ≤ k and that X satisfies as-
sumption INC(k 0 ) with k 0 ≥ 4k log(2de/k). Then the Slope estimator θ̂S with
regularization parameter defined by
r
√ log(1/δ)
τ = 8 2σ (2.36)
n
satisfies
1 k log(2d/k) log(1/δ)
MSE(Xθ̂S ) = |Xθ̂S − Xθ∗ |22 . σ 2 (2.37)
n n
and
k log(2d/k) log(1/δ)
|θ̂S − θ∗ |22 . σ 2 . (2.38)
n
with probability at least 1 − δ.
|Xθ̂S −Xθ∗ |22 +nτ |θ̂S −θ∗ |∗ ≤ 2ε> X(θ̂S −θ∗ )+nτ |θ̂S −θ∗ |∗ +2τ n|θ∗ |∗ −2τ n|θ̂S |∗ .
(2.39)
Set
u := θ̂S − θ∗ , gj = (X> ε)j , (2.40)
By Lemma 2.21, we can estimate
d
X d
X
ε> Xu = (X> ε)j uj ≤ gj∗ u∗j (2.41)
j=1 j=1
d
X gj∗
= (λj u∗j ) (2.42)
j=1
λ j
∗
gj
≤ sup |u|∗ (2.43)
j λ j
√ p nτ
≤ 4 2σ n log(1/δ)|u|∗ = |u|∗ , (2.44)
2
where we used that |Xj |22 ≤ 2n.
2.4. High-dimensional linear regression 67
Pk
Pick a permutation φ such that |θ∗ |∗ = j=1 λj |θφ(j) | and |uφ(k+1) | ≥ · · · ≥
|uφ(d) |. Then, noting that λj is monotonically decreasing,
k
X d
X
|θ∗ |∗ − |θ̂S |∗ = ∗
λj |θφ(j) |− λj (θ̂S )∗j (2.45)
j=1 j=1
k
X d
X
∗ S S
≤ λj (|θφ(j) | − |θ̂φ(j) |) − λj |θ̂φ(j) | (2.46)
j=1 j=k+1
k
X d
X
≤ λj |uφ(j) | − λj u∗j (2.47)
j=1 j=k+1
k
X d
X
≤ λj u∗j − λj u∗j . (2.48)
j=1 j=k+1
√ p
Combined with τ = 8 2σ log(1/δ)/n and the basic inequality (2.39), we
have that
whence
d
X k
X
λj u∗j ≤ 3 λj u∗j . (2.51)
j=k+1 j=1
We can now repeat the incoherence arguments from Lemma 2.17, with S
being the k largest entries of |u|, to get the same conclusion under the restriction
INC(k 0 ). First, by exactly the same argument as in Lemma 2.17, we have
|XuS c |22 9k
≥ |uS c |22 − log(2de/k)|uS |22 . (2.61)
n 32k 0 λ2d
|Xu|22 36 + 12 + 1 1
≥ |uS |22 + |uS c |22 − |uS |22 ≥ |u|22 . (2.62)
n 128 2
Hence, from (2.50),
1/2 1/2
k
X Xk
|Xu|22 + nτ |u|∗ ≤ 4nτ λ2j (u∗j )2 (2.63)
j=1 j=1
√ p
≤ 4 2τ nk log(2de/k)|Xu|2 (2.64)
p p
= 26 σ log(1/δ) k log(2de/k)|Xu|2 , (2.65)
2.4. High-dimensional linear regression 69
Problem 2.1. Consider the linear regression model with fixed design with
d ≤ n. The ridge regression estimator is employed when rank(X> X) < d but
we are interested in estimating θ∗ . It is defined for a given parameter τ > 0 by
n1 o
θ̂τridge = argmin |Y − Xθ|22 + τ |θ|22 .
θ∈IRd n
(a) Show that for any τ , θ̂τridge is uniquely defined and give its closed form
expression.
(b) Compute the bias of θ̂τridge and show that it is bounded in absolute value
by |θ∗ |2 .
Problem 2.2. Let X = (1, Z, . . . , Z d−1 )> ∈ IRd be a random vector where Z
is a random variable. Show that the matrix IE(XX > ) is positive definite if Z
admits a probability density with respect to the Lebesgue measure on IR.
Problem 2.3. Let θ̂hrd be the hard thresholding estimator defined in Defini-
tion 2.10.
1. Show that
|θ̂hrd − θ∗ |22
|θ̂hrd |0 ≤ |θ∗ |0 +
4τ 2
2. Conclude that if τ is chosen as in Theorem 2.11, then
|θ̂hrd |0 ≤ C|θ∗ |0
Problem 2.5. For any q > 0, a vector θ ∈ IRd is said to be in a weak `q ball
of radius R if the decreasing rearrangement |θ[1] | ≥ |θ[2] | ≥ . . . satisfies
|θ[j] | ≤ Rj −1/q .
Problem 2.7. Assume the linear model (2.2) with ε ∼ subGn (σ 2 ) and θ∗ 6= 0.
Show that the modified BIC estimator θ̂ defined by
n1 ed o
θ̂ ∈ argmin |Y − Xθ|22 + λ|θ|0 log
θ∈IRd n |θ|0
satisfies
ed
∗ 2
log |θ ∗ |0
MSE(Xθ̂) . |θ |0 σ .
n
with probability .99, for appropriately chosen λ. What do you conclude?
Problem 2.8. Assume that the linear model (2.2) holds where ε ∼ subGn (σ 2 ).
Moreover, assume the conditions of Theorem 2.2 √ and that the columns of X
are normalized in such a way that maxj |Xj |2 ≤ n. Then the Lasso estimator
θ̂L with regularization parameter
r
2 log(2d)
2τ = 8σ ,
n
satisfies
|θ̂L |1 ≤ C|θ∗ |1
with probability 1 − (2d)−1 for some constant C to be specified.
Chapter
3
Misspecified Linear Models
Yi = f (Xi ) + εi , i = 1, . . . , n , (3.1)
72
3.1. Oracle inequalities 73
Oracle inequalities
As mentioned in the introduction, an oracle is a quantity that cannot be con-
structed without the knowledge of the quantity of interest, here: the regression
function. Unlike the regression function itself, an oracle is constrained to take
a specific form. For all matter of purposes, an oracle can be viewed as an
estimator (in a given family) that can be constructed with an infinite amount
of data. This is exactly what we should aim for in misspecified models.
When employing the least squares estimator θ̂ls , we constrain ourselves to
estimating functions that are of the form x 7→ x> θ, even though f itself may
not be of this form. Therefore, the oracle fˆ is the linear function that is the
closest to f .
Rather than trying to approximate f by a linear function f (x) ≈ θ> x, we
make the model a bit more general and consider a dictionary H = {ϕ1 , . . . , ϕM }
of functions where ϕj : IRd → IR. In this case, we can actually remove the
assumption that X ∈ IRd . Indeed, the goal is now to estimate f using a linear
combination of the functions in the dictionary:
M
X
f ≈ ϕθ := θj ϕj .
j=1
R(ϕθ̄ ) ≤ R(ϕθ ) , ∀θ ∈ K .
or
IP R(fˆ) ≤ C inf R(ϕθ ) + φn,M,δ (K) ≥ 1 − δ ,
∀δ>0
θ∈K
where ϕθ̄ denotes the orthogonal projection of f onto the linear span of ϕ1 , . . . , ϕM .
Since Y = f + ε, we get
It yields
|ϕθ̂ls − ϕθ̄ |22 ≤ 2ε> (ϕθ̂ls − ϕθ̄ ) .
Using the same steps as the ones following equation (2.5) for the well specified
case, we get
σ2 M
|ϕθ̂ls − ϕθ̄ |22 . log(1/δ)
n
with probability 1 − δ. The result of the lemma follows.
Proof. Recall that the proof of Theorem 2.14 for the BIC estimator begins as
follows:
1 1
|Y − ϕθ̂bic |22 + τ 2 |θ̂bic |0 ≤ |Y − ϕθ |22 + τ 2 |θ|0 .
n n
This is true for any θ ∈ IRM . It implies
Cσ 2 h io
MSE(ϕθ̂bic ) ≤ 3MSE(ϕθ̄ ) + |θ̄|0 log(eM ) + log(1/δ)
n
If the linear model happens to be correct, then we simply have MSE(ϕθ̄ ) = 0.
Theorem 3.5. Assume the general regression model (3.1) with ε ∼ subGn (σ 2 ).
Moreover, assume that there exists an integer k such that the matrix Φ satisfies
assumption INC(k). Then, the Lasso estimator θ̂L with regularization param-
eter given by r r
2 log(2M ) 2 log(1/δ)
2τ = 8σ + 8σ (3.6)
n n
satisfies for some numerical constant C,
n Cσ 2 o
MSE(ϕθ̂L ) ≤ inf MSE(ϕθ ) + |θ|0 log(eM/δ)
θ∈IRM n
|θ|0 ≤k
Expanding the squares, adding τ |θ̂L − θ|1 on each side and multiplying by n,
we get
|ϕθ̂L − f |22 − |ϕθ − f |22 + nτ |θ̂L − θ|1 ≤ 2nτ |θ̂L − θ|1 + 2nτ |θ|1 − 2nτ |θ̂L |1
= 2nτ |θ̂SL − θ|1 + 2nτ |θ|1 − 2nτ |θ̂SL |1
≤ 4nτ |θ̂SL − θ|1 (3.8)
with probability 1 − δ.
It implies that either MSE(ϕθ̂L ) ≤ MSE(ϕθ ) or that
|θ̂SLc − θS c |1 ≤ 3|θ̂SL − θS |1 .
so that θ = θ̂L − θ satisfies the cone condition (2.18). Using now the Cauchy-
Schwarz inequality and Lemma 2.17, respectively, assuming that |θ|0 ≤ k, we
get p p
4nτ |θ̂SL − θ|1 ≤ 4nτ |S||θ̂SL − θ|2 ≤ 4τ 2n|θ|0 |ϕθ̂L − ϕθ |2 .
2 2
Using now the inequality 2ab ≤ αa + α2 b2 , we get
16τ 2 n|θ|0 α
4nτ |θ̂SL − θ|1 ≤ + |ϕθ̂L − ϕθ |22
α 2
2
16τ n|θ|0
≤ + α|ϕθ̂L − f |22 + α|ϕθ − f |22
α
Combining this result with (3.7) and (3.8), we find
16τ 2 |θ|0
(1 − α)MSE(ϕθ̂L ) ≤ (1 + α)MSE(ϕθ ) + .
α
To conclude the proof, it only remains to divide by 1 − α on both sides of the
above inequality and take α = 1/2.
Maurey’s argument
In there is no sparse θ such that MSE(ϕθ ) is small, Theorem 3.4 is useless
whereas the Lasso may still enjoy slow rates. In reality, no one really be-
lieves in the existence of sparse vectors but rather of approximately sparse
3.1. Oracle inequalities 78
vectors. Zipf’s law would instead favor the existence of vectors θ with abso-
lute coefficients that decay polynomially when ordered from largest to smallest
in absolute value. This is the case for example if θ has a small `1 norm but
is not sparse. For such θ, the Lasso estimator still enjoys slow rates as in
Theorem 2.15, which can be easily extended to the misspecified case (see Prob-
lem 3.2). As a result, it seems that the Lasso estimator is strictly better than
the BIC estimator as long as incoherence holds since it enjoys both fast and
slow rates, whereas the BIC estimator seems to be tailored to the fast rate.
Fortunately, such vectors can be well approximated by sparse vectors in the
following sense: for any vector θ ∈ IRM such that |θ|1 ≤ 1, there exists a vector
θ0 that is sparse and for which MSE(ϕθ0 ) is not much larger than MSE(ϕθ ). The
following theorem quantifies exactly the tradeoff between sparsity and MSE. It
is often attributed to B. Maurey and was published by Pisier [Pis81]. This is
why it is referred to as Maurey’s argument.
Then for any integer k such that 1 ≤ k ≤ M and any positive R, we have
D 2 R2
min MSE(ϕθ ) ≤ min MSE(ϕθ ) + .
θ∈IRM θ∈IRM k
|θ|0 ≤k |θ|1 ≤R
Proof. Define
θ̄ ∈ argmin |ϕθ − f |22
θ∈IRM
|θ|1 ≤R
|θ̄j |
IP(U = R sign(θ̄j )ϕj ) = , j = 1, . . . , M,
R
|θ̄|1
IP(U = 0) = 1 − .
R
√
Note that IE[U ] = ϕθ̄ and |U |2 ≤ RD n. Let now U1 , . . . , Uk be k independent
copies of U and define their average
k
1X
Ū = Ui .
k i=1
3.1. Oracle inequalities 79
Note that Ū = ϕθ̃ for some θ̃ ∈ IRM such that |θ̃|0 ≤ k. Moreover, using the
Pythagorean Theorem,
and divide by n.
Corollary 3.7. Assume that the assumptions of Theorem 3.4 hold and that
the dictionary {ϕ1 , . . . , ϕM } is normalized in such a way that
√
max |ϕj |2 ≤ n .
1≤j≤M
Then there exists a constant C > 0 such that the BIC estimator satisfies
r
n h σ 2 |θ| log(eM ) log(eM ) io
0
MSE(ϕθ̂bic ) ≤ inf 2MSE(ϕθ ) + C ∧ σ|θ|1
θ∈IRM n n
2
σ log(1/δ)
+C
n
with probability at least 1 − δ.
Let θ0 ∈ IRM . It follows from Maurey’s argument that for any k ∈ [M ], there
exists θ = θ(θ0 , k) ∈ IRM such that |θ|0 = k and
2|θ0 |21
MSE(ϕθ ) ≤ MSE(ϕθ0 ) +
k
It implies that
Since the above bound holds for any θ0 ∈ RM and k ∈ [M ], we can take an
infimum with respect to both θ0 and k on the right-hand side to get
n σ 2 |θ|0 log(eM ) o
inf MSE(ϕθ ) + C
θ∈IRM n
n |θ0 |2 σ 2 k log(eM ) o
1
≤ inf MSE(ϕθ0 ) + C min +C .
θ 0 ∈IRM k k n
To control the minimum over k, we need to consider three cases for the quantity
|θ0 |1
r
n
k̄ = .
σ log(eM )
1. If 1 ≤ k̄ ≤ M , then we get
r
|θ0 |2 σ 2 k log(eM ) log(eM )
min 1
+C ≤ Cσ|θ0 |1
k k n n
2. If k̄ ≤ 1, then
σ 2 log(eM )
|θ0 |21 ≤ C ,
n
which yields
|θ0 |2 σ 2 k log(eM ) σ 2 log(eM )
1
min +C ≤C
k k n n
3. If k̄ ≥ M , then
σ 2 M log(eM ) |θ0 |21
≤C .
n M
|θ|1
Therefore, on the one hand, if M ≥ √ , we get
σ log(eM )/n
r
|θ0 |2 σ 2 k log(eM ) |θ0 |21 log(eM )
min 1
+C ≤C ≤ Cσ|θ0 |1 .
k k n M n
|θ|1
On the other hand, if M ≤ √ , then for any θ ∈ IRM , we have
σ log(eM )/n
r
σ 2 |θ|0 log(eM ) σ 2 M log(eM ) log(eM )
≤ ≤ Cσ|θ0 |1 .
n n n
3.2. Nonparametric regression 81
Combined,
n |θ0 |2 σ 2 k log(eM ) o
1
inf MSE(ϕθ0 ) + C min +C
θ 0 ∈IRM k k n
r
n log(eM σ 2 log(eM ) o
≤ inf MSE(ϕθ0 ) + Cσ|θ0 |1 +C ,
θ 0 ∈IRM n n
which together with Theorem 3.4 yields the claim.
Note that this last result holds for any estimator that satisfies an oracle
inequality with respect to the `0 norm as in Theorem 3.4. In particular, this
estimator need not be the BIC estimator. An example is the Exponential
Screening estimator of [RT11].
Maurey’s argument allows us to enjoy the best of both the `0 and the
`1 world. The rate adapts to the sparsity of the problem and can be even
generalized to `q -sparsity (see Problem 3.3). However, it is clear from the proof
that this argument is limited to squared `2 norms such as the one appearing
in MSE and extension to other risk measures is non trivial. Some work has
been done for non-Hilbert spaces [Pis81, DDGS97] using more sophisticated
arguments.
So far, the oracle inequalities that we have derived do not deal with the
approximation error MSE(ϕθ ). We kept it arbitrary and simply hoped that
it was small. Note also that in the case of linear models, we simply assumed
that the approximation error was zero. As we will see in this section, this
error can be quantified under natural smoothness conditions if the dictionary
of functions H = {ϕ1 , . . . , ϕM } is chosen appropriately. In what follows, we
assume for simplicity that d = 1 so that f : IR → IR and ϕj : IR → IR.
Fourier decomposition
Historically, nonparametric estimation was developed before high-dimensional
statistics and most results hold for the case where the dictionary H = {ϕ1 , . . . , ϕM }
forms an orthonormal system of L2 ([0, 1]):
Z 1 Z 1
ϕ2j (x)dx = 1 , ϕj (x)ϕk (x)dx = 0, ∀ j 6= k .
0 0
Assume now that the regression function f admits the following decompo-
sition
∞
X
f= θj∗ ϕj .
j=1
There exists many choices for the orthonormal system and we give only two
as examples.
Example 3.8. Trigonometric basis. This is an orthonormal basis of L2 ([0, 1]).
It is defined by
ϕ1 ≡ 1
√
ϕ2k (x)= 2 cos(2πkx) ,
√
ϕ2k+1 (x) = 2 sin(2πkx) ,
for k = 1, 2, . . . and x ∈ [0, 1]. The fact that it is indeed an orthonormal system
can be easily check using trigonometric identities.
The next example has received a lot of attention in the signal (sound, image,
. . . ) processing community.
Example 3.9. Wavelets. Let ψ : IR → IR be a sufficiently smooth and
compactly supported function, called “mother wavelet”. Define the system of
functions
ψjk (x) = 2j/2 ψ(2j x − k) , j, k ∈ Z .
It can be shown that for a suitable ψ, the dictionary {ψj,k , j, k ∈ Z} forms an
orthonormal system of L2 ([0, 1]) and sometimes a basis. In the latter case, for
any function g ∈ L2 ([0, 1]), it holds
∞
X ∞
X Z 1
g= θjk ψjk , θjk = g(x)ψjk (x)dx .
j=−∞ k=−∞ 0
1
0
y
−1
Definition 3.10. Fix parameters β ∈ {1, 2, . . . } and L > 0. The Sobolev class
of functions W (β, L) is defined by
n
W (β, L) = f : [0, 1] → IR : f ∈ L2 ([0, 1]) , f (β−1) is absolutely continuous and
Z 1 o
[f (β) ]2 ≤ L2 , f (j) (0) = f (j) (1), j = 0, . . . , β − 1
0
jβ
for j even
aj = (3.9)
(j − 1)β for j odd
With these coefficients, we can define the Sobolev class of functions in terms
of Fourier coefficients.
1 In the sense that
Z 1 k
X
lim |f (t) − θj ϕj (t)|2 dt = 0
k→∞ 0 j=1
3.2. Nonparametric regression 84
Theorem 3.11. Fix β ≥ 1 and L > 0 and let {ϕj }j≥1 denote the trigonometric
basis of L2 ([0, 1]). Moreover, let {aj }j≥1 be defined as in (3.9). A function
f ∈ W (β, L) can be represented as
∞
X
f= θj∗ ϕj ,
j=1
where the sequence {θj∗ }j≥1 belongs to Sobolev ellipsoid of `2 (IN) defined by
n ∞
X o
Θ(β, Q) = θ ∈ `2 (IN) : a2j θj2 ≤ Q
j=1
for Q = L2 /π 2β .
Proof. Let us first define the Fourier coefficients {sk (j)}k≥1 of the jth deriva-
tive f (j) of f for j = 1, . . . , β:
Z 1
s1 (j) = f (j) (t)dt = f (j−1) (1) − f (j−1) (0) = 0 ,
0
√ Z 1
s2k (j) = 2 f (j) (t) cos(2πkt)dt ,
0
√ Z 1
s2k+1 (j) = 2 f (j) (t) sin(2πkt)dt ,
0
Moreover,
√ 1 √ Z 1
s2k+1 (β) = 2f (β−1) (t) sin(2πkt) − (2πk) 2 f (β−1) (t) cos(2πkt)dt
0 0
= −(2πk)s2k (β − 1) .
In particular, it yields
so that θ ∈ Θ(β, L2 /π 2β ) .
It can actually be shown that the reciprocal is true, that is, any function
with Fourier coefficients in Θ(β, Q) belongs to if W (β, L), but we will not be
needing this.
In what follows, we will define smooth functions as functions with Fourier
coefficients (with respect to the trigonometric basis) in a Sobolev ellipsoid. By
extension, we write f ∈ Θ(β, Q) in this case and consider any real value for β.
Proof. Fix j, j 0 ∈ {1, . . . , n − 1}, j 6= j 0 , and consider the inner product ϕ>
j ϕj 0 .
Write kj = bj/2c for the integer part of j/2 and define the vectors a, b, a0 , b0 ∈
i2πkj s i2πk 0 s
j
IRn with coordinates such that e n = as+1 +ibs+1 and e n = a0s+1 +ib0s+1
for s ∈ {0, . . . , n − 1}. It holds that
1 >
ϕ ϕj 0 ∈ {a> a0 , b> b0 , b> a0 , a> b0 } ,
2 j
depending on the parity of j and j 0 .
On the one hand, observe that if kj 6= kj 0 , we have for any σ ∈ {−1, +1},
n−1 i2πk 0 s
n−1 i2π(kj +σk 0 )s
X i2πkj s j X j
e n eσ n = e n = 0.
s=0 s=0
s=0
On the one hand if j 6= j 0 , it can only be the case that ϕ> > 0 > 0
j ϕj 0 ∈ {b a , a b }
> 0 > 0
but the same argument as above yields b a = ±a b = 0 since the imaginary
part of the inner product is still 0. Hence, in that case, ϕ> j ϕj 0 = 0. On the
other hand, if j = j 0 , then a = a0 and b = b0 so that it yields a> a0 = |a|22 = n
and b> b0 = |b|22 = n which is equivalent to ϕ> 2
j ϕj = |ϕj |2 = n. Therefore, the
design matrix Φ is such that
Φ> Φ = nIM .
for any fixed M . This truncation leads to a systematic error that vanishes as
M → ∞. We are interested in understanding the rate at which this happens.
The Sobolev assumption allows us to control precisely this error as a func-
tion of the tunable parameter M and the smoothness β.
Lemma 3.14. For any integer M ≥ 1, and f ∈ Θ(β, Q), β > 1/2, it holds
X
kϕM 2
θ ∗ − f kL2 = |θj∗ |2 ≤ QM −2β . (3.10)
j>M
Proof. Note that for any θ ∈ Θ(β, Q), if β > 1/2, then
∞ ∞
X X 1
|θj | = aj |θj |
j=2 j=2
aj
v
u ∞ 2 2X ∞
uX
1
≤t aj θj 2 by Cauchy-Schwarz
j=2
a
j=2 j
v
u ∞ 1
u X
≤ tQ <∞
j=1
j 2β
where in√the last inequality, we used the fact that for the trigonometric basis
|ϕj |2 ≤ 2n, j ≥ 1 regardless of the choice of the design X1 , . . . , Xn . When
θ∗ ∈ Θ(β, Q), we have
sX s
∗ 1
X X X 1 1
∗
p
|θj | = aj |θj | ≤ 2 ∗
aj |θj |2 . Qn 2 −β .
aj a2j
j≥n j≥n j≥n j≥n
3.2. Nonparametric regression 88
n
ls
X 2
θ̂ ∈ argmin Yi − ϕθ (Xi ) ,
θ∈IRM i=1
which should be such that ϕθ̂ls is close to ϕθ∗ . For this estimator, we have
proved (Theorem 3.3) an oracle inequality for the MSE that is of the form
|ϕM
θ̂ ls
− f |22 ≤ inf |ϕM 2 2
θ − f |2 + Cσ M log(1/δ) , C > 0.
θ∈IRM
It yields
M >
|ϕM
θ̂ ls
− ϕM 2 M M 2
θ ∗ |2 ≤ 2(ϕθ̂ ls − ϕθ ∗ ) (f − ϕθ ∗ ) + Cσ M log(1/δ)
X
>
= 2(ϕM
θ̂ ls
− ϕM
θ∗ ) ( θj∗ ϕj ) + Cσ 2 M log(1/δ)
j>M
X
>
= 2(ϕM
θ̂ ls
− ϕM
θ∗ ) ( θj∗ ϕj ) + Cσ 2 M log(1/δ) ,
j≥n
where we used Lemma 3.13 in the last equality. Together with (3.11) and
Young’s inequality 2ab ≤ αa2 + b2 /α, a, b ≥ 0 for any α > 0, we get
>
X C
2(ϕM
θ̂ ls
− ϕM
θ∗ ) ( θj∗ ϕj ) ≤ α|ϕM
θ̂ ls
− ϕM 2
θ ∗ |2 + Qn2−2β ,
α
j≥n
1 σ2 M
|ϕM
θ̂ ls
− ϕM 2
θ ∗ |2 . Qn2−2β + log(1/δ) (3.12)
α(1 − α) 1−α
M log(1/δ)
kϕM
θ̂ ls
− ϕM 2
θ ∗ kL2 ([0,1]) . n
1−2β
+ σ2 .
n
Using now Lemma 3.14 and σ 2 ≤ 1, we get
M log(1/δ)
kϕM
θ̂ ls
− f k2L2 ([0,1]) . M −2β + n1−2β + σ 2 .
n
1
Taking M = dn 2β+1 e ≤ n − 1 for n large enough yields
2β
kϕM
θ̂ ls
− f k2L2 ([0,1]) . n− 2β+1 + n1−2β σ 2 log(1/δ) .
To conclude the proof, simply note that for the prescribed β, we have n1−2β ≤
2β
n− 2β+1 .
Adaptive estimation
1
The rate attained by the projection estimator ϕθ̂ls with M = dn 2β+1 e is actually
optimal so, in this sense, it is a good estimator. Unfortunately, its implementa-
tion requires the knowledge of the smoothness parameter β which is typically
unknown, to determine the level M of truncation. The purpose of adaptive es-
timation is precisely to adapt to the unknown β, that is to build an estimator
2β
that does not depend on β and yet, attains a rate of the order of Cn− 2β+1 (up
to a logarithmic lowdown). To that end, we will use the oracle inequalities for
the BIC and Lasso estimator defined in (3.3) and (3.4) respectively. In view of
Lemma 3.13, the design matrix Φ actually satisfies the assumption ORT when
we work with the trigonometric basis. This has two useful implications:
1. Both estimators are actually thresholding estimators and can therefore
be implemented efficiently
2. The condition INC(k) is automatically satisfied for any k ≥ 1.
These observations lead to the following corollary.
√
Corollary 3.16. Fix β > (1 + 5)/4 ' 0.81, Q > 0, δ > 0 and n large enough
1
to ensure n − 1 ≥ dn 2β+1 e assume the general regression model (3.1) with
n−1
f ∈ Θ(β, Q) and ε ∼ subGn (σ 2 ), σ 2 ≤ 1. Let {ϕj }j=1 be the trigonometric
n−1 n−1
basis. Denote by ϕθ̂bic (resp. ϕθ̂L ) the BIC (resp. Lasso) estimator defined
in (3.3) (resp. (3.4)) over IRn−1 with regularization parameter given by (3.5)
(resp. (3.6)). Then ϕn−1
θ̂
, where θ̂ ∈ {θ̂bic , θ̂L } satisfies with probability 1 − δ,
2β
kϕn−1
θ̂
− f k2L2 ([0,1]) . (n/ log n)− 2β+1 (1 + σ 2 log(1/δ)) ,
Proof. For θ̂ ∈ {θ̂bic , θ̂L }, adapting the proofs of Theorem 3.4 for the BIC
estimator and Theorem 3.5 for the Lasso estimator, for any θ ∈ IRn−1 , with
probability 1 − δ
1 + α n−1
|ϕθ̂n−1 − f |22 ≤ |ϕ − f |22 + R(|θ|0 ) .
1−α θ
where
Cσ 2
R(|θ|0 ) := |θ0 | log(en/δ)
α(1 − α)
It yields
2α
|ϕn−1 − ϕn−1
θ |22 ≤ |ϕn−1 − f |22 + 2(ϕn−1 − ϕn−1 )> (f − ϕn−1 ) + R(|θ|0 )
θ̂ 1−α θ θ̂ θ θ
2α 1 n−1
≤ + |ϕθ − f |22 + α|ϕn−1 − ϕn−1
θ |22 + R(|θ|0 ) ,
1−α α θ̂
|ϕn−1
θ̂
− ϕn−1 2 n−1 2 n−1 n−1 2 n−1 2
θ ∗ |2 . |ϕθ ∗ − f |2 + R(M ) . |ϕθ ∗ − ϕθ ∗ |2 + |ϕθ ∗ − f |2 + R(M )
M M M
R(M )
kϕn−1 − ϕn−1 2 n−1 n−1 2
θ ∗ kL2 ([0,1]) . kϕθ ∗ − ϕθ ∗ kL2 ([0,1]) + Qn
1−2β
+ .
θ̂ M M n
Moreover, using (3.10), we find that
M 2
kϕn−1 − f k2L2 ([0,1]) . M −2β + Qn1−2β + σ log(en/δ) .
θ̂ n
1
To conclude the proof, choose M = d(n/ log n) 2β+1 e and observe that the choice
of β ensures that n1−2β . M −2β .
Problem 3.1. Show that the least-squares estimator θ̂ls defined in (3.2) sat-
isfies the following exact oracle inequality:
M
IEMSE(ϕθ̂ls ) ≤ inf MSE(ϕθ ) + Cσ 2
θ∈IRM n
for some constant C to be specified.
2
Problem 3.2. Assume that ε ∼ subG √ n (σ ) and the vectors ϕj are normalized
in such a way that maxj |ϕj |2 ≤ n. Show that there exists a choice of τ
such that the Lasso estimator θ̂L with regularization parameter 2τ satisfies the
following exact oracle inequality:
r
n log M o
MSE(ϕθ̂L ) ≤ inf MSE(ϕθ ) + Cσ|θ|1
θ∈IRM n
with probability at least 1 − M −c for some positive constants C, c.
Problem 3.3. Let √ {ϕ1 , . . . , ϕM } be a dictionary normalized in such a way
that maxj |ϕj |2 ≤ D n. Show that for any integer k such that 1 ≤ k ≤ M , we
have 1 1 2
M q̄ − k q̄
min MSE(ϕθ ) ≤ min MSE(ϕθ ) + Cq D2 ,
θ∈IRM θ∈IRM k
|θ|0 ≤2k |θ|w`q ≤1
1
where |θ|w`q denotes the weak `q norm and q̄ is such that q + 1q̄ = 1, for q > 1.
Problem 3.4. Show that the trigonometric basis and the Haar system indeed
form an orthonormal system of L2 ([0, 1]).
Problem 3.5. Consider the n × d random matrix Φ = {ϕj (Xi )}1≤i≤n where
1≤j≤d
X1 , . . . , Xn are i.i.d uniform random variables on the interval [0, 1] and φj is
the trigonometric basis as defined in Example 3.8. Show that Φ satisfies INC(k)
with probability at least .9 as long as n ≥ Ck 2 log(d) for some large enough
constant C > 0.
Problem 3.6. If f ∈ Θ(β, Q) for β > 1/2 and Q > 0, then f is continuous.
Chapter
4
Minimax Lower Bounds
In the previous chapters, we have proved several upper bounds and the goal of
this chapter is to assess their optimality. Specifically, our goal is to answer the
following questions:
1. Can our analysis be improved? In other words: do the estimators that
we have studied actually satisfy better bounds?
2. Can any estimator improve upon these bounds?
Both questions ask about some form of optimality. The first one is about
optimality of an estimator, whereas the second one is about optimality of a
bound.
The difficulty of these questions varies depending on whether we are looking
for a positive or a negative answer. Indeed, a positive answer to these questions
simply consists in finding a better proof for the estimator we have studied
(question 1.) or simply finding a better estimator, together with a proof that
it performs better (question 2.). A negative answer is much more arduous.
For example, in question 2., it is a statement about all estimators. How can
this be done? The answer lies in information theory (see [CT06] for a nice
introduction).
In this chapter, we will see how to give a negative answer to question 2. It
will imply a negative answer to question 1.
92
4.1. Optimality in a minimax sense 93
σ2 d
IRd θ̂ls Theorem 2.2
n
r
log d
B1 σ θ̂Bls1 Theorem 2.4
n
σ2 k
B0 (k) log(ed/k) θ̂Bls0 (k) Corollaries 2.8-2.9
n
2
where ε = (ε1 , . . . , εd )> ∼ Nd (0, σn Id ), θ∗ = (θ1∗ , . . . , θd∗ )> ∈ Θ is the parameter
of interest and Θ ⊂ IRd is a given set of parameters. We will need a more precise
notation for probabilities and expectations throughout this chapter. Denote by
IPθ∗ and IEθ∗ the probability measure and corresponding expectation that are
associated to the distribution of Y from the GSM (4.1).
Recall that GSM is a special case of the linear regression model when the
design matrix satisfies the ORT condition. In this case, we have proved several
performance guarantees (upper bounds) for various choices of Θ that can be
expressed either in the form
IE |θ̂n − θ∗ |22 ≤ Cφ(Θ)
(4.2)
or the form
|θ̂n − θ∗ |22 ≤ Cφ(Θ) , with prob. 1 − d−2 (4.3)
For some constant C. The rates φ(Θ) for different choices of Θ that we have
obtained are gathered in Table 41 together with the estimator (and the corre-
sponding result from Chapter 2) that was employed to obtain this rate. Can
any of these results be improved? In other words, does there exists another
estimator θ̃ such that supθ∗ ∈Θ IE|θ̃ − θ∗ |22 φ(Θ)?
A first step in this direction is the Cramér-Rao lower bound [Sha03] that
allows us to prove lower bounds in terms of the Fisher information. Neverthe-
less, this notion of optimality is too stringent and often leads to nonexistence
of optimal estimators. Rather, we prefer here the notion of minimax optimality
that characterizes how fast θ∗ can be estimated uniformly over Θ.
Definition 4.1. We say that an estimator θ̂n is minimax optimal over Θ if it
satisfies (4.2) and there exists C 0 > 0 such that
where the infimum is taker over all estimators (i.e., measurable functions of
Y). Moreover, φ(Θ) is called minimax rate of estimation over Θ.
4.2. Reduction to finite hypothesis testing 94
for some positive constants A and C”. The above inequality also implies a lower
bound with high probability. We can therefore employ the following alternate
definition for minimax optimality.
Definition 4.2. We say that an estimator θ̂ is minimax optimal over Θ if it
satisfies either (4.2) or (4.3) and there exists C 0 > 0 such that
where the infimum is taker over all estimators (i.e., measurable functions of
Y). Moreover, φ(Θ) is called minimax rate of estimation over Θ.
Minimax lower bounds rely on information theory and follow from a simple
principle: if the number of observations is too small, it may be hard to distin-
guish between two probability distributions that are close to each other. For
example, given n i.i.d. observations, it is impossible to reliably decide whether
they are drawn from N (0, 1) or N ( n1 , 1). This simple argument can be made
precise using the formalism of statistical hypothesis testing. To do so, we reduce
our estimation problem to a testing problem. The reduction consists of two
steps.
1. Reduction to a finite number of parameters. In this step the goal
is to find the largest possible number of parameters θ1 , . . . , θM ∈ Θ under
the constraint that
|θj − θk |22 ≥ 4φ(Θ) . (4.7)
This problem boils down to a packing of the set Θ.
Then we can use the following trivial observations:
inf sup IPθ |θ̂ − θ|22 > φ(Θ) ≥ inf max IPθj |θ̂ − θj |22 > φ(Θ) .
θ̂ θ∈Θ θ̂ 1≤j≤M
4.3. Lower bounds based on two hypotheses 95
where the infimum is taken over all tests ψ based on Y and that take
values in {1, . . . , M }.
Conclusion: it is sufficient for proving lower bounds to find θ1 , . . . , θM ∈ Θ
such that |θj − θk |22 ≥ 4φ(Θ) and
inf max IPθj ψ 6= j ≥ C 0 .
ψ 1≤j≤M
The above quantity is called minimax probability of error. In the next sections,
we show how it can be bounded from below using arguments from information
theory. For the purpose of illustration, we begin with the simple case where
M = 2 in the next section.
Lemma 4.3 (Neyman-Pearson). Let IP0 and IP1 be two probability measures.
Then for any test ψ, it holds
Z
IP0 (ψ = 1) + IP1 (ψ = 0) ≥ min(p0 , p1 )
Next for any test ψ, define its rejection region R = {ψ = 1}. Let R? = {p1 ≥
p0 } denote the rejection region of the likelihood ratio test ψ ? . It holds
IP0 (ψ = 1) + IP1 (ψ = 0) = 1 + IP0 (R) − IP1 (R)
Z
=1+ p0 − p1
ZR Z
=1+ p0 − p1 + p0 − p1
R∩R? R∩(R? )c
Z Z
=1− |p0 − p1 | + |p0 − p1 |
R∩R? R∩(R? )c
Z
= 1 + |p0 − p1 | 1I(R ∩ (R? )c ) − 1I(R ∩ R? )
Proof. Clearly (i) = (ii) and the Neyman-Pearson Lemma gives (iv) = (v).
Moreover, by identifying a test ψ to its rejection region, it is not hard to see
that (i) = (v). Therefore it remains only to show that (iii) is equal to any
of the other expressions. Hereafter, we show that (iii) = (iv). To that end,
observe that
Z Z Z
|p0 − p1 | = p1 − p0 + p0 − p1
p ≥p p1 <p0
Z 1 0 Z Z
= p1 + p0 − min(p0 , p1 )
p1 ≥p0 p1 <p0
Z Z Z
=1− p1 + 1 − p0 − min(p0 , p1 )
p1 <p0 p1 ≥p0
Z
= 2 − 2 min(p0 , p1 )
It can be shown [Tsy09] that the integral is always well defined when IP1
IP0 (though it can be equal to ∞ even in this case). Unlike the total variation
distance, the Kullback-Leibler divergence is not a distance. Actually, it is not
even symmetric. Nevertheless, it enjoys properties that are very useful for our
purposes.
Proposition 4.6. Let IP and Q be two probability measures. Then
1. KL(IP, Q) ≥ 0.
2. The function (IP, Q) 7→ KL(IP, Q) is convex.
3. If IP and Q are product measures, i.e.,
n
O n
O
IP = IPi and Q = Qi
i=1 i=1
then
n
X
KL(IP, Q) = KL(IPi , Qi ) .
i=1
Proof. If IP is not absolutely continuous then the result is trivial. Next, assume
that IP Q and let X ∼ IP.
1. Observe that by Jensen’s inequality,
dQ dQ
KL(IP, Q) = −IE log (X) ≥ − log IE (X) = − log(1) = 0 .
dIP dIP
2. Consider the function f : (x, y) 7→ x log(x/y) and compute its Hessian:
Theorem 4.9. Assume that Θ contains two hypotheses θ0 and θ1 such that
|θ0 − θ1 |22 = 8α2 σ 2 /n for some α ∈ (0, 1/2). Then
2ασ 2 1
inf sup IPθ (|θ̂ − θ|22 ≥ ) ≥ − α.
θ̂ θ∈Θ n 2
Proof. Write for simplicity IPj = IPθj , j = 0, 1. Recall that it follows from the
4.4. Lower bounds based on many hypotheses 101
2ασ 2
inf sup IPθ (|θ̂ − θ|22 ≥ ) ≥ inf max IPj (ψ 6= j)
θ̂ θ∈Θ n ψ j=0,1
1
≥ inf IP0 (ψ = 1) + IP1 (ψ = 0)
2 ψ
1h i
= 1 − TV(IP0 , IP1 ) (Prop.-def. 4.4)
2
1h p i
≥ 1 − KL(IP1 , IP0 ) (Lemma 4.8)
2 r
1h n|θ1 − θ0 |22 i
= 1− (Example 4.7)
2 2σ 2
1 h i
= 1 − 2α
2
Clearly the result of Theorem 4.9 matches the upper bound for Θ = IRd
only for d = 1. How about larger d? A quick inspection of our proof shows
that our technique, in its present state, cannot yield better results. Indeed,
there are only two known candidates for the choice of θ∗ . With this knowledge,
one can obtain upper bounds that do not depend on d by simply projecting
Y onto the linear span of θ0 , θ1 and then solving the GSM in two dimensions.
To obtain larger lower bounds, we need to use more than two hypotheses. In
particular, in view of the above discussion, we need a set of hypotheses that
spans a linear space of dimension proportional to d. In principle, we should
need at least order d hypotheses but we will actually need much more.
The reduction to hypothesis testing from Section 4.2 allows us to use more
than two hypotheses. Specifically, we should find θ1 , . . . , θM such that
where the infimum is taken over all tests with values in {1, . . . , M }.
4.4. Lower bounds based on many hypotheses 102
Proof. Define
M
1 X
pj = Pj (ψ = j) and qj = Pk (ψ = j)
M
k=1
so that
M M
1 X 1 X
p̄ = pj ∈ (0, 1) , q̄ = qj .
M j=1 M j=1
This result can be seen to follow directly from a well known inequality often
referred to as data processing inequality but we are going to prove it directly.
Denote by dPk=j (resp. dPk6=j ) the conditional density of Pk given ψ(X) = j
(resp. ψ(X) 6= j) and recall that
Z dP
j
KL(Pj , Pk ) = log dPj
dPk
Z dP Z dP
j j
= log dPj + log dPj
ψ=j dP k ψ6=j dP k
Z dP =j P (ψ = j)
j j
= Pj (ψ = j) log dPj=j
dPk=j Pk (ψ = j)
Z dP 6=j P (ψ 6= j)
j
+ Pj (ψ 6= j) log dPj6=j
dPk6=j Pk (ψ 6= j)
P (ψ = j)
j
= Pj (ψ = j) log + KL(Pj=j , Pj=j )
Pk (ψ = j)
P (ψ 6= j)
j
+ Pj (ψ 6= j) log + KL(Pj6=j , Pj6=j )
Pk (ψ 6= j)
≥ kl(Pj (ψ = j), Pk (ψ = j)) .
4.4. Lower bounds based on many hypotheses 103
n|θj − θk |22
KL(IPj , IPk ) = ≤ α log(M ) .
2σ 2
Moreover, since M ≥ 5,
1
PM
M2 j,k=1 KL(IPj , IPk ) + log 2 α log(M ) + log 2 1
≤ ≤ 2α + .
log(M − 1) log(M − 1) 2
of a packing of the discrete hypercube {0, 1}d with respect to the Hamming
distance defined by
d
X
ρ(ω, ω 0 ) = 1I(ωi 6= ωj0 ) , ∀ ω, ω 0 ∈ {0, 1}d .
i=1
Lemma 4.12 (Varshamov-Gilbert). For any γ ∈ (0, 1/2), there exist binary
vectors ω1 , . . . ωM ∈ {0, 1}d such that
1
(i) ρ(ωj , ωk ) ≥ − γ d for all j 6= k ,
2
2 γ2 d
(ii) M = beγ d c ≥ e 2 .
M (M − 1) d M (M − 1)
IP X − > γd ≤ exp − 2γ 2 d + log
<1
2 2 2
as soon as
M (M − 1) < 2 exp 2γ 2 d .
2
A sufficient condition for the above inequality to hold is to take M = beγ d c ≥
γ2 d
e 2 . For this value of M , we have
1
IP ∀j 6= k , ρ(ωj , ωk ) ≥ −γ d >0
2
and by virtue of the probabilistic method, there exist ω1 , . . . ωM ∈ {0, 1}d that
satisfy (i) and (ii)
α σ2 d 1
inf sup IPθ |θ̂ − θ|22 ≥ ≥ − 2α .
θ̂ θ∈IRd 256 n 2
h k k−1
X i h X i
IE exp s Zi = IE exp s Zi IE exp sZk Z1 , . . . , Zk=1
i=1 i=1
h k−1
X i
= IE exp s Zi (Qk (es − 1) + 1)
i=1
k−1
h X i 2k s
≤ IE exp s Zi (e − 1) + 1
i=1
d
..
.
2k s k
≤ (e − 1) + 1
d
= 2k
d
For s = log(1 + 2k ). Putting everything together, we get
sk
IP ∃ ωj 6= ωk : ρ(ωj , ωk ) < k ≤ exp log M + k log 2 −
2
k d
= exp log M + k log 2 − log(1 + )
2 2k
k d
≤ exp log M + k log 2 − log(1 + )
2 2k
k d
≤ exp log M − log(1 + ) (for d ≥ 8k)
4 2k
< 1,
for some β > 0 to be chosen later. We can check the conditions of Theorem 4.11:
β 2 σ2 d β 2 σ2 d
(i) |θj − θk |22 = ρ(ωj , ωk ) log(1 + )≥4 k log(1 + );
n 2k 8n 2k
β 2 σ2 d 2kβ 2 σ 2 d 2ασ 2
(ii) |θj −θk |22 = ρ(ωj , ωk ) log(1+ ) ≤ log(1+ ) ≤ log(M ) ,
n 2k n 2k n
4.5. Application to the Gaussian sequence model 108
pα
for β = 8. Applying now Theorem 4.11 yields
α2 σ 2 d 1
inf sup IPθ |θ̂ − θ|22 ≥ k log(1 + ) ≥ − 2α .
θ̂ θ∈IRd 64n 2k 2
|θ|0 ≤k
Note that the modified BIC estimator of Problem 2.7 and the Slope (2.26)
are also minimax optimal over B0 (k). However, unlike θ̂Bls0 (k) and the modified
BIC, the Slope is also adaptive to k. For any ε > 0, the Lasso estimator and the
BIC estimator are minimax optimal for sets of parameters such that k ≤ d1−ε .
Corollary 4.16. Recall that B1 (R) ⊂ IRd denotes the set vectors θ ∈ IRd such
that |θ|1 ≤ R. Then there exist a constant C > 0 such that if d ≥ n1/2+ε ,
ε > 0, the minimax rate of estimation over B1 (R) in the Gaussian sequence
model is
log d
φ(B0 (k)) = min(R2 , Rσ ).
n
Moreover, it is attained by the constrained least squares estimator θ̂Bls1 (R) if
R ≥ σ logn d and by the trivial estimator θ̂ = 0 otherwise.
Proof. To complete the proof of the statement, we need to study risk of the
trivial estimator equal to zero for small R. Note that if |θ∗ |1 ≤ R, we have
|0 − θ∗ |22 = |θ∗ |22 ≤ |θ∗ |21 = R2 .
Remark 4.17. Note that the inequality |θ∗ |22 ≤ |θ∗ |21 appears to be quite loose.
Nevertheless, it is tight up to a multiplicative constant for the vectors of the
log d
form θj = ωj Rk that are employed in the lower bound. Indeed, if R ≤ σ n ,
we have k ≤ 2/β
R2 β
|θj |22 = ≥ |θj |21 .
k 2
4.6 LOWER BOUNDS FOR SPARSE ESTIMATION VIA χ2 DIVERGENCE
In this section, we will show how to derive lower bounds by directly con-
trolling the distance between a simple and a composite hypothesis, which is
useful when investigating lower rates for decision problems instead of estima-
tion problems.
Define the χ2 divergence between two probability distributions IP, Q as
Z 2
dIP
− 1 dQ if IP Q,
χ2 (IP, Q) = dQ
∞ otherwise.
When we compare the expression in the case where IP Q, then wesee that
both the KL divergence and the χ2 divergence can be written as f dIP
R
dQ dQ
with f (x) = x log x and f (x) = (x − 1)2 , respectively. Also note that both of
these functions are convex in x and fulfill f (1) = 0, which by Jensen’s inequality
shows us that they are non-negative and equal to 0 if and only if IP = Q.
Firstly, let us derive another useful expression for the χ2 divergence. By
expanding the square, we have
2
(dIP)2
Z Z Z Z
2 dIP
χ (IP, Q) = − 2 dIP + dQ = dQ − 1
dQ dQ
Secondly, we note that we can bound the KL divergence from above by the
χ2 divergence, via Jensen’s inequality, as
(dIP)2
Z Z
dIP
KL(IP, Q) = log dIP ≤ log = log(1 + χ2 (IP, Q)) ≤ χ2 (IP, Q),
dQ dQ
(4.8)
4.6. Lower bounds for sparse estimation via χ2 divergence 110
The first step is to compute dIPS /(dIP0 ). Writing out the corresponding Gaus-
sian densities yields
1 n
√
√ d exp(− 2 kX − µ1IS / kk2 2)
dIPS 2σ
(X) = 2π 1 n
dIP0 √ d exp(− 2 kX − 0k2 2)
2π 2σ
n 2µ 2
= exp √ hX, 1IS i − µ
2σ 2 k
For convenience, we introduce the notation X = √σn Z for a standard normal
√ √
Z ∼ N (0, Id ), ν = µ n/(σ k), and write ZS for the restriction to the coor-
dinates in the set S. Multiplying two of these densities and integrating with
respect to IP0 in turn then reduces to computing the mgf of a Gaussian and
gives
n σ 2
IEX∼IP0 exp 2µ √ (ZS + ZT ) − 2µ
2σ 2 kn
= IEZ∼N (0,Id ) exp ν(ZS + ZT ) − ν 2 k .
P P
Decomposing ZS + ZT = 2 i∈S∩T Zi + i∈S∆T Zi and noting that |S∆T | ≤
2k, we see that
ν2
dIPS dIPT 4 2
IEIP0 = exp ν |S ∩ T + |S∆T | − ν 2 k
dIP0 dIP0 2 2
2
≤ exp 2ν |S ∩ T | .
Now, we need to take the expectation over two uniform draws of support
sets S and T , which reduces via conditioning and exploiting the independence
of the two draws to
¯ IP0 ) ≤ IES,T [exp(2ν 2 |S ∩ T |] − 1
χ2 (IP,
= IET IES [[exp(2ν 2 |S ∩ T | | T ]] − 1
= IES [exp(2ν 2 |S ∩ [k]|)] − 1.
Similar to the proof of the sparse Varshamov-Gilbert bound, Lemma 4.14, the
distribution of |S ∩ [k]| is stochastically dominated by a binomial distribution
Bin(k, k/d), so that
So far, we have seen many examples of rates tha scale with n−1 for the
squared error. In this section, we are going to see an example for estimating
4.7. Lower bounds for estimating the `1 norm via moment
matching 113
Y ∼ N (θ, In ). (4.10)
Moreover, write Fi for the marginal distributions over the priors and denote
their density with respect to the average of the two probability distributions
by fi . With this, we write the χ2 distance between f0 and f1 as
2 2
f1 (X) f1 (X)
I 2 = χ2 (IPf0 , IPf1 ) = IEf0 −1 = IEf0 − 1.
f0 (X) f0 (X)
Z Z
B(θ)dµ1 (θ) − B(θ)dµ0 (θ) ≥ |m1 − m0 | − (ε + v0 )I. (4.11)
in particular
(|m1 − m0 | − v0 I)2
Z
max IEθ (Tb(X) − T (θ))2 dµi (θ) ≥ , (4.13)
i∈{0,1} (I + 2)2
and
(|m1 − m0 | − v0 I)2
sup IEθ (Tb(X) − T (θ))2 ≥ . (4.14)
θ∈Θ (I + 2)2
Proof. Without loss of generality, assume m1 ≥ m0 . We start by considering
the term
f1 (X) − f0 (X)
IEf0 (T (X) − m0 )
b
f0 (X)
f1 (X) − f0 (X)
= IE T (X)
b
f0 (X)
h i h i
= IEf1 m1 + T (X) − m1 − IEf0 m0 + Tb(X) − m0
b
Z Z
= m1 + B(θ)dµ1 (θ) − m0 + B(θ)dµ0 (θ) .
Moreover,
Z
IEf0 (Tb(X) − m0 )2 = IEθ (Tb(X) − m0 )2 dµ0 (θ)
Z
= IEθ (Tb(X) − T (θ) + T (θ) − m0 )2 dµ0 (θ)
Z
= IEθ (Tb(X) − T (θ))2 dµ0 (θ)
Z
+ 2 B(θ)(T (θ) − m0 )dµ0 (θ)
Z
+ (T (θ) − m0 )2 dµ0 (θ)
≤ ε2 + 2εv0 + v02 = (ε + v0 )2 .
Therefore, by Cauchy-Schwarz,
f1 (X) − f0 (X)
IEf0 (Tb(X) − m0 ) ≤ (ε + v0 )I,
f0 (X)
and hence
Z Z
m1 + B(θ)dµ1 (θ) − m0 + B(θ)dµ0 (θ) ≤ (ε + v0 )I,
whence
Z Z
B(θ)dµ0 (θ) − B(θ)dµ1 (θ) ≥ m1 − m0 − (ε + v0 )I,
4.7. Lower bounds for estimating the `1 norm via moment
matching 115
in the second term, we do not change the minimal value and obtain that
Z Z 2
B(θ)2 dµ1 (θ) ≥ B(θ)dµ1 (θ)
≥ ((m1 − m0 − (ε + v0 )I − ε)+ )2 .
Combining this with the estimate for the quadratic above yields
Z
IEθ (Tb(X) − T (θ))2 d(λµ0 + (1 − λ)µ1 )(θ)
which is (4.12).
Finally, since the minimax risk is bigger than any mixture risk, we can
bound it from below by the maximum value of this bound, which is obtained
at λ = (I + 1)/(I + 2) to get (4.13), and by the same argument (4.14).
and the polynomial attaining this error by G∗k . For f (t) = |t|, it is known
[Riv90] that
β∗ = lim 2kδ2k (f ) ∈ (0, ∞),
k→∞
4.7. Lower bounds for estimating the `1 norm via moment
matching 116
Lemma 4.22. For an integer k > 0, there are two probability measures ν0 and
ν1 on [−1, 1] that fulfill the following:
Proof. The idea is to construct the measures via the Hahn-Banach theorem
and the Riesz representation theorem.
First, consider f (t) = |t| as an element of the space C([−1, 1]) equipped with
the supremum norm and define Pk as the space of polynomials of degree up to
k, and Fk := span(Pk ∪ {f }). On Fk , define the functional T (cf + pk ) = cδk ,
which is well-defined because f is not a polynomial.
Claim: kT k = sup{T (g) : g ∈ Fk , kgk∞ ≤ 1} = 1.
kT k ≥ 1: Let G∗k be the best-approximating polynomial to f in Pk . Then,
kf − G∗k k∞ = δk , k(f − G∗k )/δk k∞ = 1, and T ((f − G∗k )/δk ) = 1.
kT k ≤ 1: Suppose g = cf + pk with pk ∈ Pk , kgk∞ = 1 and T (g) > 1.
Then c > 1/δk and
1
kf − (−pk /c)k∞ = < δk ,
c
contradicting the definition of δk .
Now, by the Hahn-Banach theorem, there is a norm-preserving extension
of T to C([−1, 1]) which we again denote by T . By the Riesz representation
theorem, there is a Borel signed measure τ with variation equal to 1 such that
Z 1
T (g) = g(t)dτ (t), for all g ∈ C([−1, 1]).
−1
For θ ∈ IRn ,
β∗2
inf sup IEθ (Tb(X) − T (θ))2 ≥ (1 + o(1)). (4.16)
Tb θ∈IRn 16e2 log n
The last remaining ingredient for the proof are the Hermite polynomials, a
family of orthogonal polynomials with respect to the Gaussian density
1 2
ϕ(y) = √ e−y /2 .
2π
For our purposes, it is enough to define them by the derivatives of the density,
dk
ϕ(y) = (−1)k Hk (y)ϕ(y),
dy k
and to observe that they are orthogonal,
Z Z
2
Hk (y)ϕ(y)dy = k!, Hk (y)Hj (y)ϕ(y)dy = 0, for k 6= j.
Proof. We want to use Theorem 4.21. To construct the prior measures, we scale
the measures from Lemma 4.22 appropriately: Let kn be an even integer that
is to be determined, and ν0 , ν1 the two measures given by Lemma 4.22. Define
g(x) = M x and define the measures µi by dilating νi , µi (A) = νi (g −1 (A)) for
every Borel set A ⊆ [−M, M ] and i ∈ {0, 1}. Hence,
1. µ0 and µ1 are symmetric about 0;
2. tl dµ1 (t) = tl dµ0 (t) for all l ∈ {0, 1, . . . , kn };
R R
R R
3. |t|dµ1 (t) − |t|dµ0 (t) = 2M δkn .
Qn
To get priors for n i.i.d. samples, consider the product priors µni = j=1 µi .
With this, we have
and
M2
IEµn0 (T (θ) − IEµn0 T (θ))2 = IEµn0 (T (θ) − IEµn0 T (θ))2 ≤ ,
n
since each θi ∈ [−M, M ].
4.7. Lower bounds for estimating the `1 norm via moment
matching 118
where the last step used a Stirling type estimate, k! > (k/e)k .
Note that if x = o(1/n), then (1 + x)n − 1 = o(nx) by Taylor’s formula,
so we can choose kn ≥ log n/(log log n) to guarantee that for n large enough,
In → 0. With Theorem 4.21, we have
√
(2M δkn − (M/ n)In )2
inf sup IE(Tb − T (θ))2 ≥
Tb θ∈Θn (M ) (In + 2)2
2
log log n
= β∗2 M 2 (1 + o(1)).
log n
4.7. Lower bounds for estimating the `1 norm via moment
matching 119
√
To prove the lower bound over IRn , take M = log n and kn to be the
smallest integer such that kn ≥ 2e log n and plug this into (4.17), yielding
kn !n
2 3/2 e log n
In ≤ 1 + n − 1,
2e log n
Note that [CL11] also complement these lower bounds with upper bounds
that are based on polynomial approximation, completing the picture of esti-
mating T (θ).
4.8. Problem Set 120
as long as σ 2 ≤ n.
(c) The set of nonnegative vectors of IRd .
σ
√ {0, 1}d
(d) The discrete hypercube 16 n
.
Problem 4.5. Fix β ≥ 5/3, Q > 0 and prove that the minimax rate of esti-
2β
mation over Θ(β, Q) with the k · kL2 ([0,1]) -norm is given by n− 2β+1 .
[Hint:Consider functions of the form
N
C X
fj = √ ωji ϕi
n i=1
Over the past decade or so, matrices have entered the picture of high-dimensional
statistics for several reasons. Perhaps the simplest explanation is that they are
the most natural extension of vectors. While this is true, and we will see exam-
ples where the extension from vectors to matrices is straightforward, matrices
have a much richer structure than vectors allowing “interaction” between their
rows and columns. In particular, while we have been describing simple vectors
in terms of their sparsity, here we can measure the complexity of a matrix by
its rank. This feature was successfully employed in a variety of applications
ranging from multi-task learning to collaborative filtering. This last application
was made popular by the Netflix prize in particular.
In this chapter, we study several statistical problems where the parameter of
interest θ is a matrix rather than a vector. These problems include: multivari-
ate regression, covariance matrix estimation and principal component analysis.
Before getting to these topics, we begin by a quick reminder on matrices and
linear algebra.
121
5.1. Basic facts about matrices 122
|Ax|2 y > Ax
λmax (A) = maxn = maxn = max y > Ax .
x∈IR |x|2 x∈IR |y|2 |x|2
m
x∈S n−1
y∈IR y∈S m−1
Vector norms
The simplest way to treat a matrix is to deal with it as if it were a vector. In
particular, we can extend `q norms to matrices:
X 1/q
|A|q = |aij |q , q > 0.
ij
5.1. Basic facts about matrices 123
The case q = 2 plays a particular role for matrices and |A|2 is called the
Frobenius norm of A and is often denoted by kAkF . It is also the Hilbert-
Schmidt norm associated to the inner product:
hA, Bi = Tr(A> B) = Tr(B > A) .
Spectral norms
Let λ = (λ1 , . . . , λr , 0, . . . , 0) be the singular values of a matrix A. We can
define spectral norms on A as vector norms on the vector λ. In particular, for
any q ∈ [1, ∞],
kAkq = |λ|q ,
is called Schatten q-norm of A. Here again, special cases have special names:
• q = 2: kAk2 = kAkF is the Frobenius norm defined above.
• q = 1: kAk1 = kAk∗ is called the Nuclear norm (or trace norm) of A.
• q = ∞: kAk∞ = λmax (A) = kAkop is called the operator norm (or
spectral norm) of A.
We are going to employ these norms to assess the proximity to our matrix
of interest. While the interpretation of vector norms is clear by extension from
the vector case, the meaning of “kA−Bkop is small” is not as transparent. The
following subsection provides some inequalities (without proofs) that allow a
better reading.
Proof. Note that the last equality of the lemma is obvious since
r
X
A − Ak = λj uj vj>
j=k+1
and the matrices in the sum above are orthogonal to each other.
Thus, it is sufficient to prove that for any matrix B such that rank(B) ≤ k,
it holds
Xr
kA − Bk2F ≥ λ2j .
j=k+1
The model
Throughout this section, we consider the following multivariate linear regres-
sion model:
Y = XΘ∗ + E , (5.1)
where Y ∈ IRn×T is the matrix of observed responses, X is the n × d observed
design matrix (as before), Θ ∈ IRd×T is the matrix of unknown parameters and
E ∼ subGn×T (σ 2 ) is the noise matrix. In this chapter, we will focus on the
prediction task, which consists in estimating XΘ∗ .
As mentioned in the foreword of this chapter, we can view this problem as T
(univariate) linear regression problems Y (j) = Xθ∗,(j) +ε(j) , j = 1, . . . , T , where
Y (j) , θ∗,(j) and ε(j) are the jth column of Y, Θ∗ and E respectively. In particu-
lar, an estimator for XΘ∗ can be obtained by concatenating the estimators for
each of the T problems. This approach is the subject of Problem 5.1.
The columns of Θ∗ correspond to T different regression tasks. Consider the
following example as a motivation. Assume that the Subway headquarters
want to evaluate the effect of d variables (promotions, day of the week, TV
ads,. . . ) on their sales. To that end, they ask each of their T = 40, 000
restaurants to report their sales numbers for the past n = 200 days. As a
result, franchise j returns to headquarters a vector Y(j) ∈ IRn . The d variables
for each of the n days are already known to headquarters and are stored in
a matrix X ∈ IRn×d . In this case, it may be reasonable to assume that the
same subset of variables has an impact of the sales for each of the franchise,
though the magnitude of this impact may differ from franchise to franchise. As
a result, one may assume that the matrix Θ∗ has each of its T columns that
is row sparse and that they share the same sparsity pattern, i.e., Θ∗ is of the
form:
0 0 0 0
• • • •
• • • •
Θ∗ = 0 0 0 0 ,
.. .. .. ..
. . . .
0 0 0 0
• • • •
where • indicates a potentially nonzero entry.
It follows from the result of Problem 5.1 that if each task is performed
individually, one may find an estimator Θ̂ such that
1 kT log(ed)
IEkXΘ̂ − XΘ∗ k2F . σ 2 ,
n n
where k is the number of nonzero coordinates in each column of Θ∗ . We
remember that the term log(ed) corresponds to the additional price to pay
for not knowing where the nonzero components are. However, in this case,
when the number of tasks grows, this should become easier. This fact was
proved in [LPTVDG11]. We will see that we can recover a similar phenomenon
5.2. Multivariate regression 126
when the number of tasks becomes large, though larger than in [LPTVDG11].
Indeed, rather than exploiting sparsity, observe that such a matrix Θ∗ has rank
k. This is the kind of structure that we will be predominantly using in this
chapter.
Rather than assuming that the columns of Θ∗ share the same sparsity
pattern, it may be more appropriate to assume that the matrix Θ∗ is low rank
or approximately so. As a result, while the matrix may not be sparse at all,
the fact that it is low rank still materializes the idea that some structure is
shared across different tasks. In this more general setup, it is assumed that the
columns of Θ∗ live in a lower dimensional space. Going back to the Subway
example this amounts to assuming that while there are 40,000 franchises, there
are only a few canonical profiles for these franchises and that all franchises are
linear combinations of these profiles.
Recall that the threshold for the hard thresholding estimator was chosen to
be the level of the noise with high probability. The singular value thresholding
estimator obeys the same rule, except that the norm in which the magnitude of
the noise is measured is adapted to the matrix case. Specifically, the following
lemma will allow us to control the operator norm of the matrix F .
Lemma 5.3. Let A be a d × T random matrix such that A ∼ subGd×T (σ 2 ).
Then p p
kAkop ≤ 4σ log(12)(d ∨ T ) + 2σ 2 log(1/δ)
with probability 1 − δ.
Proof. This proof follows the same steps as Problem 1.4. Let N1 be a 1/4-
net for S d−1 and N2 be a 1/4-net for S T −1 . It follows from Lemma 1.18
that we can always choose |N1 | ≤ 12d and |N2 | ≤ 12T . Moreover, for any
u ∈ S d−1 , v ∈ S T −1 , it holds
1
u> Av ≤ max x> Av + max u> Av
x∈N1 4 u∈S d−1
1 1
≤ max max x> Ay + max max x> Av + max u> Av
x∈N1 y∈N2 4 x∈N1 v∈S T −1 4 u∈S d−1
1
≤ max max x> Ay + max max u> Av
x∈N1 y∈N2 2 u∈S d−1 v∈S T −1
It yields
kAkop ≤ 2 max max x> Ay
x∈N1 y∈N2
Proof. Assume without loss of generality that the singular values of Θ∗ and y
are arranged in a non increasing order: λ1 ≥ λ2 ≥ . . . and λ̂1 ≥ λ̂2 ≥ . . . .
Define the set S = {j : |λ̂j | > 2τ }.
Observe first that it follows from Lemma 5.3 that kF kop ≤ τ for τ chosen as
in (5.3) on an event A such that IP(A) ≥ 1 − δ. The rest of the proof assumes
that the event A occurred.
Note that it follows from Weyl’s inequality that |λ̂j − λj | ≤ kF kop ≤ τ . It
implies that S ⊂ {j : |λj | > τ }P and S c ⊂ {j : |λj | ≤ 3τ }.
Next define the oracle Θ̄ = j∈S λj uj vj> and note that
Moreover,
Therefore, X
kΘ̂svt − Θ̄k2F ≤ 72|S|τ 2 = 72 τ2 .
j∈S
= 144 rank(Θ∗ )τ 2 .
In the next subsection, we extend our analysis to the case where X does not
necessarily satisfy the assumption ORT.
Penalization by rank
The estimator from this section is the counterpart of the BIC estimator in the
spectral domain. However, we will see that unlike BIC, it can be computed
efficiently.
Let Θ̂rk be any solution to the following minimization problem:
n1 o
min kY − XΘk2F + 2τ 2 rank(Θ) .
Θ∈IRd×T n
This estimator is called estimator by rank penalization with regularization pa-
rameter τ 2 . It enjoys the following property.
Theorem 5.5. Consider the multivariate linear regression model (5.1). Then,
the estimator by rank penalization Θ̂rk with regularization parameter τ 2 , where
τ is defined in (5.3) satisfies
1 σ 2 rank(Θ∗ )
kXΘ̂rk − XΘ∗ k2F ≤ 8 rank(Θ∗ )τ 2 . d ∨ T + log(1/δ) .
n n
with probability 1 − δ.
which is equivalent to
kXΘ̂rk − XΘ∗ k2F ≤ 2hE, XΘ̂rk − XΘ∗ i − 2nτ 2 rank(Θ̂rk ) + 2nτ 2 rank(Θ∗ ) .
where
XΘ̂rk − XΘ∗
U= .
kXΘ̂rk − XΘ∗ kF
Write
XΘ̂rk − XΘ∗ = ΦN ,
where Φ is a n × r, r ≤ d matrix whose columns form orthonormal basis of the
column span of X. The matrix Φ can come from the SVD of X for example:
X = ΦΛΨ> . It yields
ΦN
U=
kN kF
and
It follows from Theorem 5.5 that the estimator by rank penalization enjoys
the same properties as the singular value thresholding estimator even when X
does not satisfy the ORT condition. This is reminiscent of the BIC estimator
which enjoys the same properties as the hard thresholding estimator. However
this analogy does not extend to computational questions. Indeed, while the
rank penalty, just like the sparsity penalty, is not convex, it turns out that
XΘ̂rk can be computed efficiently.
Note first that
1 n1 o
min kY − XΘk2F + 2τ 2 rank(Θ) = min min kY − XΘk2F + 2τ 2 k .
Θ∈IRd×T n k n Θ∈IRd×T
rank(Θ)≤k
min kY − XΘk2F
Θ∈IRd×T
rank(Θ)≤k
5.2. Multivariate regression 131
can be solved efficiently. To that end, let Ȳ = X(X> X)† X> Y denote the orthog-
onal projection of Y onto the image space of X: this is a linear operator from
IRd×T into IRn×T . By the Pythagorean theorem, we get for any Θ ∈ IRd×T ,
kY − XΘk2F = kY − Ȳk2F + kȲ − XΘk2F .
Next consider the SVD of Ȳ:
X
Ȳ = λj uj vj>
j
It holds,
n
> 1 X >
(Xi x)(Xi> y) − IE (Xi> x)(Xi> y) .
x (Σ̂ − Id )y =
n i=1
5.3. Covariance matrix estimation 133
s 2 2
s 2 2
i
= IE exp Z+ − IE[Z+ ] − Z− − IE[Z− ])
4 4
s 2 2
s 2 2
1/2
≤ IE exp Z+ − IE[Z+ ] IE exp − Z− − IE[Z− ] ,
2 2
where in the last inequality, we used Cauchy-Schwarz. Next, since X ∼
subGd (1), we have Z+ , Z− ∼ subG(2), and it follows from Lemma 1.12 that
2 2 2 2
Z+ − IE[Z+ ] ∼ subE(32) , and Z− − IE[Z− ] ∼ subE(32).
t 2d 2 2d 2 1/2
≥ log(144) + log(1/δ) ∨ log(144) + log(1/δ) .
32 n n n n
This concludes our proof.
Theorem 5.7 indicates that for fixed d, the empirical covariance matrix is a
consistent estimator of Σ (in any norm as they are all equivalent in finite dimen-
sion). However, the bound that we got is not satisfactory in high-dimensions
when d n. To overcome this limitation, we can introduce sparsity as we have
done in the case of regression. The most obvious way to do so is to assume
that few of the entries of Σ are non zero and it turns out that in this case
5.4. Principal component analysis 134
thresholding is optimal. There is a long line of work on this subject (see for
example [CZZ10] and [CZ12]).
Once we have a good estimator of Σ, what can we do with it? The key
insight is that Σ contains information about the projection of the vector X
onto any direction u ∈ S d−1 . Indeed, we have that var(X > u) = u> Σu, which
d > u) = u> Σ̂u. Observe that it follows from
can be readily estimated by Var(X
Theorem 5.7 that
d > u) − Var(X > u) = u> (Σ̂ − Σ)u
Var(X
≤ kΣ̂ − Σkop
r
d + log(1/δ) d + log(1/δ)
. kΣkop ∨
n n
with probability 1 − δ.
The above fact is useful in the Markowitz theory of portfolio section for
example [Mar52], where a portfolio of assets is a vector u ∈ IRd such that
|u|1 = 1 and the risk of a portfolio is given by the variance Var(X > u). The
goal is then to maximize reward subject to risk constraints. In most instances,
the empirical covariance matrix is plugged into the formula in place of Σ. (See
Problem 5.4).
Figure 5.1. Projection onto two dimensions of 1, 387 points from gene expression data.
Source: Gene expression blog.
This model is often called the spiked covariance model. By a simple rescaling,
it is equivalent to the following definition.
Definition 5.8. A covariance matrix Σ ∈ IRd×d is said to satisfy the spiked
covariance model if it is of the form
Σ = θvv > + Id ,
5.4. Principal component analysis 136
u> > 2
∠(u1 , x) − λ2 sin2 ∠(u1 , x)
1 Au1 − v1 Av1 ≥ λ1 − λ1 cos
= (λ1 − λ2 ) sin2 ∠(u1 , x) .
where in the last inequality, we used the fact that rank(u1 u> >
1 − v1 v1 ) = 2.
It is straightforward to check that
which concludes the first part of the lemma. Note that we can replace λ1 − λ2
with µ1 − µ2 since the result is completely symmetric in A and B.
It remains to show the second part of the lemma. To that end, observe that
Sparse PCA
In the example of Figure 5.1, it may be desirable to interpret the meaning of
the two directions denoted by PC1 and PC2. We know that they are linear
combinations of the original 500,000 gene expression levels. A natural question
to ask is whether only a subset of these genes could suffice to obtain similar
results. Such a discovery could have potential interesting scientific applications
as it would point to a few genes responsible for disparities between European
populations.
In the case of the spiked covariance model this amounts to having a sparse
v. Beyond interpretability as we just discussed, sparsity should also lead to
statistical stability as in the case of sparse linear regression for example. To
enforce sparsity, we will assume that v in the spiked covariance model is k-
sparse: |v|0 = k. Therefore, a natural candidate to estimate v is given by v̂
defined by
v̂ > Σ̂v̂ = max u> Σ̂u .
u∈S d−1
|u|0 =k
It is easy to check that λkmax (Σ̂) = v̂ > Σ̂v̂ is the largest of all leading eigenvalues
among all k × k sub-matrices of Σ̂ so that the maximum is indeed attained,
though there my be several maximizers. We call λkmax (Σ̂) the k-sparse leading
eigenvalue of Σ̂ and v̂ a k-sparse leading eigenvector.
Theorem 5.11. Let Y ∈ IRd be a random vector such that IE[Y ] = 0, IE[Y Y > ] =
Id and Y ∼ subGd (1). Let X1 , . . . , Xn be n independent copies of sub-Gaussian
random vector X = Σ1/2 Y so that IE[X] = 0, IE[XX > ] = Σ and X ∼ subGd (kΣkop ).
Assume further that Σ = θvv > + Id satisfies the spiked covariance model for
v such that |v|0 = k ≤ d/2. Then, the k-sparse largest eigenvector v̂ of the
empirical covariance matrix satisfies,
r
1 + θ k log(ed/k) + log(1/δ) k log(ed/k) + log(1/δ)
min |εv̂ − v|2 . ∨ .
ε∈{±1} θ n n
with probability 1 − δ.
Since both v̂ and v are k sparse, there exists a (random) set S ⊂ {1, . . . , d}
such that |S| ≤ 2k and {v̂v̂ > − vv > }ij = 0 if (i, j) ∈
/ S 2 . It yields
Where for any d × d matrix M , we defined the matrix M (S) to be the |S| × |S|
sub-matrix of M with rows and columns indexed by S and for any vector
5.4. Principal component analysis 139
Proof. We use the general technique developed in Chapter 4. Using this tech-
nique, we need to provide the a set v1 , . . . , vM ∈ S d−1 of k sparse vectors such
that
It yields
θ
KL(IPu , IPv ) = Tr (uu> − vv > )(Id + θuu> )
2(1 + θ)
θ2
= (1 − (u> v)2 )
2(1 + θ)
θ2
sin2 ∠(u, v)
=
2(1 + θ)
It yields
θ2
sin2 ∠(vi , vj ) .
KL(IPvi , IPvj ) =
2(1 + θ)
Next, note that
k d log M
sin2 ∠(vi , vj ) = 1 − γ 2 ωi> ωj − `2 ≤ γ 2 (k − 1) ≤ Cγ 2 log
≤ 2 ,
θ n k θ n
for Cγ small enough. We conclude that (ii) holds, which completes our proof.
Together with the upper bound of Theorem 5.11, we get the following corol-
lary.
p
Corollary 5.13. Assume that θ ≤ 1 and k log(d/k) ≤ n. Then θ−1 k log(d/k)/n
is the minimax rate of estimation over B0 (k) in the spiked covariance model.
Proof. 1. =⇒ 2.:
X X X
Φ(B) = (−1)|B\C| Ψ(C)
B⊆A B⊆A C⊆B
X X
= Ψ(C) (−1)|B\C|
C⊆A C⊆B⊆A
X X
= Ψ(C) (−1)|H| .
C⊆A H⊆A\C
Proof. Define
HA (x) = log f (xA , x∗Ac )
and X
ΦA (x) = (−1)|A\B| HB (x).
B⊆A
(p + k) log(d/δ)
kΘ̂ − Θ∗ k2F .
n
with probability at least 1 − δ.
for some Θ̃ = Θ∗ + t(Θ − Θ∗ ), t ∈ [0, 1]. Note that essentially by the convexity
of log det, we have
Tr(Θ̃−1 (Θ − Θ∗ )Θ̃−1 (Θ − Θ∗ )) = kΘ̃−1 (Θ − Θ∗ )k2F ≥ λmin (Θ̃−1 )2 kΘ − Θ∗ k2F
and
λmin (Θ̃−1 ) = (λmax (Θ̃))−1 .
If we write ∆ = Θ − Θ∗ , then for k∆kF ≤ 1,
λmax (Θ̃) = kΘ̃kop = kΘ∗ +∆kop ≤ kΘ∗ kop +k∆kop ≤ kΘ∗ kop +k∆kF ≤ kΘ∗ kop +1,
and therefore
l(Θ, Σn ) − l(Θ∗ , Σn ) ≥ Tr((Σn − Σ∗ )(Θ − Θ∗ )) + ckΘ − Θ∗ k2F ,
for c = (kΘ∗ kop + 1)−2 /2, if k∆kF ≤ 1.
This takes care of the case where ∆ is small. To handle the case where it
is large, define g(t) = l(Θ∗ + t∆, Σn ) − l(Θ∗ , Σn ). By the convexity of l in Θ,
g(1) − g(0) g(t) − g(0)
≥ ,
1 t
so that plugging in t = k∆k−1
F gives
1
l(Θ, Σn ) − l(Θ∗ , Σn ) ≥ k∆kF l(Θ∗ +∆, Σn ) − l(Θ∗ , Σn )
k∆kF
1
≥ k∆kF Tr((Σn − Σ∗ ) ∆) + c
k∆kF
= Tr((Σn − Σ∗ )∆) + ck∆kF ,
for k∆kF ≥ 1.
If we now write ∆ = Θ̂ − Θ∗ and assume k∆kF ≥ 1, then by optimality of
Θ̂,
k∆kF ≤ C [Tr((Σ∗ − Σn )∆) + λ(kΘ∗Dc k1 − kΘDc k1 )]
≤ C [Tr((Σ∗ − Σn )∆) + λ(k∆S k1 − k∆S c k1 )] ,
where S = {(i, j) ∈ Dc : Θ∗i,j 6= 0} by triangle inequality. Now, split the error
contributions for the diagonal and off-diagonal elements,
Tr((Σ∗ − Σn )∆) + λ(k∆S k1 − k∆S c k1 )
≤ k(Σ∗ − Σn )D kF k∆D kF + k(Σ∗ − Σn )Dc k∞ k∆Dc k1
+ λ(k∆S k1 − k∆S c k1 ).
√
By Hölder inequality,
p k(Σ∗ −√Σn )D kF ≤ dkΣ∗ − Σn k∞ , and by Lemma 5.18,
kΣ∗ − Σn k∞ ≤ log(d/δ)/ n for n & log(d/δ). Combining these two esti-
mates,
Tr((Σ∗ − Σn )∆) + λ(k∆S k1 − k∆S c k1 )
r r
d log(d/δ) log(d/δ)
. k∆D kF + k∆Dc k1 + λ(k∆S k1 − k∆S c k1 )
n n
5.5. Graphical models 147
p
Setting λ = C log(ep/δ)/n and splitting k∆Dc k1 = k∆S k1 + k∆S c k1 yields
r r
d log(d/δ) log(d/δ)
k∆D kF + k∆Dc k1 + λ(k∆S k1 − k∆S c k1 )
n n
r r
d log(d/δ) log(d/δ)
≤ k∆D kF + k∆S k1
n n
r r
d log(d/δ) k log(d/δ)
≤ k∆D kF + k∆S kF
n n
r
(d + k) log(d/δ)
≤ k∆kF
n
Combining this with a left-hand side of k∆kF yields k∆kF = 0 for n &
(d + k) log(d/δ), a contradiction to k∆kF ≥ 1 where this bound is effective.
Combining it with a left-hand side of k∆k2F gives us
(d + k) log(d/δ)
k∆k2F . ,
n
as desired.
Lower bounds
Here, we will show that the bounds in Theorem 5.19 are optimal up to log
factors. It will again be based on applying Fano’s inequality, Theorem 4.10.
5.5. Graphical models 148
and
β2
kΘj − Θl k2F . (k log(1 + d/(2k)) + d)
n
β2
≤ log(M1 M2 ),
n
which yields the following theorem.
Theorem 5.20. Denote by Bk the set of positive definite matrices with at most
k non-zero off-diagonal entries and assume n & k log(d). Then, if IPΘ denotes
a Gaussian distribution with inverse covariance matrix Θ and mean 0, there
exists a constant c > 0 such that
α2
⊗n ∗ 2 1
inf sup IPΘ∗ kΘ̂ − Θ kF ≥ c (k log(1 + d/(2k)) + d) ≥ − 2α.
∗
Θ̂ Θ ∈Bk n 2
Ising model
Like the Gaussian graphical model, the Ising model is also a model of pairwise
interactions but for random variables that take values in {−1, 1}d .
Before developing our general approach, we describe the main ideas in the
simplest Markov random field: an Ising model without2 external field [VMLC16].
Such models specify the distribution of a random vector Z = (Z (1) , . . . , Z (d) ) ∈
{−1, 1}d as follows
where we used the fact that the diagonal elements W are equal to zero.
2 The presence of an external field does not change our method. It merely introduces
an intercept in the logistic regression problem and comes at the cost of more cumbersome
notation. All arguments below follow, potentially after minor modifications and explicit
computations are left to the reader.
5.5. Graphical models 150
With this representation, it is easy to see that w 7→ `¯n (w) is a concave function.
The (constrained) maximum likelihood estimator (MLE) ŵ ∈ IRd−1 is defined
to be any solution of the following convex optimization problem:
This problem can be solved very efficiently using a variety of methods similar
to the Lasso.
The following lemma follows from [Rig12].
Lemma 5.21. Fix δ ∈ (0, 1). Conditionally on (X1 , . . . , Xn ), the constrained
MLE ŵ defined in (5.16) satisfies with probability at least 1 − δ that
n
r
1X > 2 λ log(2(p − 1)/δ)
(X (ŵ − w)) ≤ 2λe
n i=1 i 2n
Therefore, writing Ỹi = (Yi + 1)/2 ∈ {0, 1}, we get that ŵ is the solution to the
following minimization problem:
where
n
1 X
− 2Ỹi Xi> w + log 1 + exp(2Xi> w)
κ̄n (w) =
n i=1
5.5. Graphical models 151
For any w ∈ IRd−1 , write κ(w) = IE[κ̄n (w)] where here and throughout this
proof, all expectations are implicitly taken conditionally on X1 , . . . , Xn . Ob-
serve that
κ(w) = −IE[Ỹ1 ]X1> w + log 1 + exp(2Xi> w) .
Next, we get from the basic inequality κ̄n (ŵ) ≤ κ̄n (w) for any w ∈ B1 (λ)
that
n
1X
κ(ŵ) − κ(w) ≤ (Ỹi − IE[Ỹi ])Xi> (ŵ − w)
n i=1
n
2λ X
≤ max (Ỹi − IE[Ỹi ])Xi> ej ,
n 1≤j≤p−1 i=1
where in the second inequality, we used Hölder’s inequality and the fact |Xi |∞ ≤
1 for all i ∈ [n]. Together with Hoeffding’s inequality and a union bound, it
yields
r
log(4(p − 1)/δ) δ
κ(ŵ) − κ(w) ≤ 2λ , with probability 1 − .
2n 2
Note that the Hessian of κ is given by
n
1 X 4 exp(2Xi> w)
∇2 κ(w) = Xi Xi> .
n i=1 (1 + exp(2Xi> w))2
4 exp(2Xi> w) 4e2λ
>
≥ ≥ e−2λ
(1 + exp(2Xi w)) 2 (1 + e2λ )2
To that end, we must exploit the covariance structure of the Ising model. The
following lemma is similar to the combination of Lemmas 6 and 7 in [VMLC16].
5.5. Graphical models 152
Lemma 5.22. Fix R > 0, δ ∈ (0, 1) and let Z ∈ IRd be distributed according
to (5.14). The following holds with probability 1−δ/2, uniformly in u ∈ B1 (2λ):
n
r
1X > 2 1 2 log(2p(p − 1)/δ)
(Z u) ≥ |u|∞ exp(−2λ) − 4λ
n i=1 i 2 2n
To control the cross term, let us condition on the neighborhood Z (¬1) to get
p p
X X
2 IE u(1) Z (1) u(j) Z (j) = 2 IE IE u(1) Z (1) |Z (¬1) u(j) Z (j)
j=2 j=2
p
2 X 2
≤ IE IE u(1) Z (1) |Z (¬1) + IE u(j) Z (j) ,
j=2
where we used the fact that 2|ab| ≤ a2 + b2 for all a, b ∈ IR. The above two
displays together yield
2
var(Z > u) ≥ |u|2∞ (1 − sup IE[Z (1) |Z (¬1) = z]
z∈{−1,1}p−1
exp(2e>
1 W z) − 1 exp(2λ) − 1
sup IE[Z (1) |Z (¬1) = z] = sup = .
z∈{−1,1}p−1 z∈{−1,1}p exp(2e>
1 W z) + 1 exp(2λ) +1
5.5. Graphical models 153
Therefore,
|u|2∞ 1
var(Z > u) ≥ ≥ |u|2∞ exp(−λ) .
1 + exp(2λ) 2
Together with (5.17), it yields the desired result.
Combining the above two lemmas immediately yields the following theorem.
Theorem 5.23. Let ŵ be the constrained maximum likelihood estimator (5.16)
and let w∗ = (e>j W)
(¬j)
be the jth row of W with the jth entry removed. Then
with probability 1 − δ we have
r
∗ 2 log(2p2 /δ)
|ŵ − w |∞ ≤ 9λ exp(3λ) .
n
5.6. Problem set 154
Problem 5.1. Using the results of Chapter 2, show that the following holds
for the multivariate regression model (5.1).
1. There exists an estimator Θ̂ ∈ IRd×T such that
1 rT
kXΘ̂ − XΘ∗ k2F . σ 2
n n
with probability .99, where r denotes the rank of X .
2. There exists an estimator Θ̂ ∈ IRd×T such that
1 |Θ∗ |0 log(ed)
kXΘ̂ − XΘ∗ k2F . σ 2 .
n n
with probability .99.
Problem 5.2. Consider the multivariate regression model (5.1) where Y has
SVD: X
Y= λ̂j ûj v̂j> .
j
Let M be defined by
X
M̂ = λ̂j 1I(|λ̂j | > 2τ )ûj v̂j> , τ > 0 .
j
Problem 5.3. Consider the multivariate regression model (5.1) and define Θ̂
be the any solution to the minimization problem
n1 o
min kY − XΘk2F + τ kXΘk1
Θ∈IRd×T n
5.6. Problem set 155
where λ∗1 ≥ λ∗2 ≥ . . . and λ̂1 ≥ λ̂2 ≥ . . . are the singular values
of XΘ∗ and Y respectively and the SVD of Y is given by
X
Y= λ̂j ûj v̂j>
j
when a solution exists for a given. It is the portfolio that has minimum risk
among all portfolios with reward at least λ, provided such portfolios exist.
In practice, the distribution of X is unknown. Assume that we observe n
independent copies X1 , . . . , Xn of X and use them to compute the following
estimators of µ(u) and R(u) respectively:
n
1X >
µ̂(u) = X̄ > u = X u,
n i=1 i
n
> 1 X
R̂(u) = u Σ̂u, Σ̂ = (Xi − X̄)(Xi − X̄)> .
n − 1 i=1
and
1
R̂(u) − R(u) . √ .
n
2. Show that r
log d
R̂(û) − R(û) . ,
n
with probability .99.
3. Define the estimator ũ by:
ũ = argmin R̂(u)
u :µ̂(u)≥λ−ε
find the smallest ε > 0 (up to multiplicative constant) such that we have
R(ũ) ≤ R(u∗ ) with probability .99.
Bibliography
[AS08] Noga Alon and Joel H. Spencer. The probabilistic method. Wiley-
Interscience Series in Discrete Mathematics and Optimization.
John Wiley & Sons, Inc., Hoboken, NJ, third edition, 2008. With
an appendix on the life and work of Paul Erdős.
[AW02] Rudolf Ahlswede and Andreas Winter. Strong converse for iden-
tification via quantum channels. IEEE Transactions on Infor-
mation Theory, 48(3):569–579, 2002.
[Ber09] Dennis S. Bernstein. Matrix mathematics. Princeton University
Press, Princeton, NJ, second edition, 2009. Theory, facts, and
formulas.
[Bil95] Patrick Billingsley. Probability and measure. Wiley Series in
Probability and Mathematical Statistics. John Wiley & Sons
Inc., New York, third edition, 1995. A Wiley-Interscience Pub-
lication.
[BLM13] Stéphane Boucheron, Gábor Lugosi, and Pascal Massart. Con-
centration inequalities. Oxford University Press, Oxford, 2013.
A nonasymptotic theory of independence, With a foreword by
Michel Ledoux.
[BLT16] Pierre C. Bellec, Guillaume Lecué, and Alexandre B. Tsybakov.
Slope meets lasso: Improved oracle bounds and optimality.
arXiv preprint arXiv:1605.08651, 2016.
[Bre15] Guy Bresler. Efficiently learning Ising models on arbitrary
graphs. STOC’15—Proceedings of the 2015 ACM Symposium
on Theory of Computing, pages 771–782, 2015.
[BRT09] Peter J. Bickel, Ya’acov Ritov, and Alexandre B. Tsybakov. Si-
multaneous analysis of Lasso and Dantzig selector. Ann. Statist.,
37(4):1705–1732, 2009.
[BT09] Amir Beck and Marc Teboulle. A fast iterative shrinkage-
thresholding algorithm for linear inverse problems. SIAM J.
Imaging Sci., 2(1):183–202, 2009.
157
Bibliography 158
Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.
Alternative Proxies: