Lectures On Convex Sets
Lectures On Convex Sets
Niels Lauritzen
N IELS L AURITZEN
D EPARTMENT OF M ATHEMATICAL S CIENCES
U NIVERSITY OF A ARHUS
D ENMARK
E MAIL : niels@imf.au.dk
URL: http://home.imf.au.dk/niels/
March 2009
Contents
Preface
Notation
1
Introduction
1.1 Linear inequalities .
1.1.1 Two variables
1.2 Polyhedra . . . . . .
1.3 Exercises . . . . . . .
vii
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Basics
2.1 Convex subsets of R n . . . . . . . . . . .
2.2 The convex hull . . . . . . . . . . . . . .
2.2.1 Intersections of convex subsets .
2.3 Extremal points . . . . . . . . . . . . . .
2.4 The characteristic cone for a convex set
2.5 Convex cones . . . . . . . . . . . . . . .
2.6 Affine independence . . . . . . . . . . .
2.7 Caratheodorys theorem . . . . . . . . .
2.8 The dual cone . . . . . . . . . . . . . . .
2.9 Exercises . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Separation
3.1 Separation of a point from a closed convex set
3.2 Supporting hyperplanes . . . . . . . . . . . . .
3.3 Separation of disjoint convex sets . . . . . . . .
3.4 An application . . . . . . . . . . . . . . . . . . .
3.5 Farkas lemma . . . . . . . . . . . . . . . . . . .
3.6 Exercises . . . . . . . . . . . . . . . . . . . . . .
Polyhedra
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
1
2
2
4
6
.
.
.
.
.
.
.
.
.
.
9
9
11
14
15
16
16
18
20
21
24
.
.
.
.
.
.
27
29
31
33
33
34
37
39
iii
Contents
iv
4.1
.
.
.
.
.
.
.
.
40
41
43
46
48
49
50
54
A Linear (in)dependence
A.1 Linear dependence and linear equations . . . . . . . . . . . . .
A.2 The rank of a matrix . . . . . . . . . . . . . . . . . . . . . . . .
A.3 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
57
57
59
61
B Analysis
B.1 Measuring distances . . . . . . . .
B.2 Sequences . . . . . . . . . . . . . .
B.2.1 Supremum and infimum . .
B.3 Bounded sequences . . . . . . . . .
B.4 Closed subsets . . . . . . . . . . . .
B.5 The interior and boundary of a set
B.6 Continuous functions . . . . . . . .
B.7 The main theorem . . . . . . . . . .
B.8 Exercises . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
63
63
65
68
68
69
70
71
72
74
75
75
77
79
Bibliography
81
Index
83
4.2
4.3
4.4
4.5
4.6
4.7
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Preface
The theory of convex sets is a vibrant and classical field of modern mathematics with rich applications in economics and optimization.
The material in these notes is introductory starting with a small chapter
on linear inequalities and Fourier-Motzkin elimination. The aim is to show
that linear inequalities can be solved and handled with very modest means.
At the same time you get a real feeling for what is going on, instead of plowing through tons of formal definitions before encountering a triangle.
The more geometric aspects of convex sets are developed introducing notions such as extremal points and directions, convex cones and their duals,
polyhedra and separation of convex sets by hyperplanes.
The emphasis is primarily on polyhedra. In a beginning course this seems
to make a lot of sense: examples are readily available and concrete computations are feasible with the double description method ([7], [3]).
The main theoretical result is the theorem of Minkowski and Weyl on
the structure of polyhedra. We stop short of introducing general faces of
polyhedra.
I am grateful to Markus Kiderlen and Jesper Funch Thomsen for very
useful comments on these notes.
Notation
Z denotes the set of integers . . . , 2, 1, 0, 1, 2, . . . and R the set of real
numbers.
R n denotes the set of all n-tuples (or vectors) {( x1 , . . . , xn ) | x1 , . . . , xn
R } of real numbers. This is a vector space over R you can add vectors and multiply them by a real number. The zero vector (0, . . . , 0)
R n will be denoted 0.
Let u, v R n . The inequality u v means that holds for every
coordinate. For example (1, 2, 3) (1, 3, 4), since 1 1, 2 3 and
3 4. But (1, 2, 3) (1, 2, 2) is not true, since 3 2.
When x R n , b R m and A is an m n matrix, the m inequalities,
Ax b, are called a system of linear inequalities. If b = 0 this system is
called homogeneous.
Let u, v R n . Viewing u and v as n 1 matrices, the matrix product
ut v is nothing but the usual inner product of u and v. In this setting,
ut u = |u|2 , where |u| is the usual length of u.
vii
Chapter 1
Introduction
You probably agree that it is quite easy to solve the equation
2x = 4.
(1.1)
(1.2)
x + y + 2z = 9
of linear equations in x, y and z. Using the first equation 2x + y + z = 7 we
solve for x and get
x = (7 y z)/2.
(1.3)
This may be substituted into the remaining two equations in (1.2) and we get
the simpler system
3y + z = 9
y + 3z = 11
of linear equations in y and z. Again using the first equation in this system
we get
y = (9 z)/3
(1.4)
to end up with the simple equation
8z = 24.
Chapter 1. Introduction
y 3
y 0
(1.6)
or simply 0 x 2. Here you see the real difference between linear equations and linear inequalities. When you reverse = you get =, but when you
reverse after multiplying by 1, you get . This is why solving linear
inequalities is more involved than solving linear equations.
(1.5) to
0 x
x 6 2y
2y x
x 3+y
along with the inequality 0 y which does not involve x. Again, just like in
one variable, this system can be reduced just two inequalities
x min(6 2y, 3 + y)
max(0, 2 y) x
(1.7)
(1.8)
You can solve (1.7) in x and y if and only if you can solve (1.8) in y. If
you think about it for a while you will realize that (1.8) is equivalent to the
following four inequalities
0
0
2 y
2 y
6
3
6
3
2y
+ y
2y
+ y
where the inequality y 0 from before is attached. This system can be reduced to
0 y 3
Through a lot of labor we have proved that two numbers x and y solve the
system (1.5) if and only if
0 y 3
max(0, 2 y) x min(6 2y, 3 + y)
If you phrase things a bit more geometrically, we have proved that the projection of the solutions to (1.5) on the y-axis is the interval [0, 3]. In other words,
if x, y solve (1.5), then y [0, 3] and if y [0, 3], there exists x R such that
x, y form a solution to (1.5). This is the context for the next section.
Chapter 1. Introduction
1.2 Polyhedra
Let us introduce precise definitions. A linear inequality in n variables x1 , . . . , xn
is an inequality of the form
a1 x1 + + an xn b,
where a1 , . . . , an , b R.
DEFINITION 1.2.1
The set of solutions
P=
to a system
( x1 , . . . , x n ) R n
a11 x1 + +
a1n xn
..
.
b1 )
am1 x1 + + amn xn bm
a11 x1 + +
a1n xn
..
.
b1
am1 x1 + + amn xn bm
of finitely many linear inequalities (here aij and bi are real numbers) is called a polyhedron. A bounded polyhedron is called a polytope.
A polyhedron is an extremely important special case of a convex subset
of R n . We will return to the definition of a convex subset in the next chapter.
The proof of the following important theorem may look intimidating at
first. If this is so, then take a look at 1.1.1 once again. Do not get fooled by the
slick presentation here. In its purest form the result goes back to a paper by
Fourier1 from 1826 (see [2]). It is also known as Fourier-Motzkin elimination,
simply because you are eliminating the variable x1 and because Motzkin2
rediscovered it in his dissertation Beitrage zur Theorie der linearen Ungleichungen with Ostrowski3 in Basel, 1933 (not knowing the classical paper
by Fourier). The main result in the dissertation of Motzkin was published
much later in [6].
THEOREM 1.2.2
Consider the projection : R n R n1 given by
( x1 , . . . , x n ) = ( x2 , . . . , x n ) .
If P R n is a polyhedron, then ( P) R n1 is a polyhedron.
1 Jean
2 Theodore Samuel
1.2. Polyhedra
a1n xn
..
.
b1
am1 x1 + + amn xn bm
We divide the m inequalities into
G = {i | ai1 > 0}
Z = {i | ai1 = 0}
L = {i | ai1 < 0}
aj2 x2 + + ajn xn + bj x1 ,
if j L, where aik = aik /ai1 and bi = bi /ai1 for k = 2, . . . , n. So the inequalities in L and G are equivalent to the two inequalities
max ai2 x2 + + ain xn + bi i L x1
min aj2 x2 + + ajn xn + bj j G .
Now ( x2 , . . . , xn ) ( P) if and only if there exists x1 such that ( x1 , . . . , xn )
P. This happens if and only if ( x2 , . . . , xn ) satisfies the inequalities in Z and
max ai2 x2 + + ain xn + bi i L min aj2 x2 + + ajn xn + bj j G
Chapter 1. Introduction
1.3 Exercises
(1) Let
P=
x
3 3x
( x, y, z) R
x
x
y z
y z
+ 3y z
y + 3z
0 )
1
2
3
x + y
y +
z
z
z
z
0
0
0
1
1
1
1.3. Exercises
(ii) Let P denote the polyhedron from Exercise 1. You can see that
1 1
(0, 0, 0), (1, , ) P
2 2
have values 0 and 1 on their first coordinates, but what is the minimal first coordinate of a point in P.
(4) A vitamin pill P is produced using two ingredients M1 and M2 . The pill
needs to satisfy two requirements for the vital vitamins V1 and V2 . It
must contain at least 6 mg and at most 15 mg of V1 and at least 5 mg and
at most 12 mg of V2 . The ingredient M1 contains 3 mg of V1 and 2 mg of
V2 per gram. The ingredient M2 contains 2 mg of V1 and 3 mg of V2 per
gram.
We want a vitamin pill of minimal weight satisfying the requirements.
How many grams of M1 and M2 should we mix?
Chapter 2
Basics
The definition of a convex subset is quite elementary and profoundly important. It is surprising that such a simple definition can be so far reaching.
[u, v] S
for every u, v S.
EXAMPLE 2.1.2
Two subsets S1 , S2 R2 are sketched below. Here S2 is convex, but S1 is not.
9
Chapter 2. Basics
10
S1
S2
These non-polyhedral convex sets are usually much more complicated than
their polyhedral cousins especially when you want to count the number of
integral points1 inside them. Counting the number N (r) of integral points
inside a circle of radius r is a classical and very difficult problem going back
to Gauss2 . Gauss studied the error term E(r) = | N (r) r2 | and proved that
E(r) 2 2r.
Counting integral points in polyhedral convex sets is difficult but theoretically much better understood. For example if P is a convex polygon in
the plane with integral vertices, then the number of integral points inside P
is given by the formula of Pick3 from 1899:
| P Z2 | = Area( P) +
1
B( P) + 1,
2
with coordinates in Z.
Friedrich Gauss (17771855), German mathematician. Probably the greatest mathematician that ever lived.
3 Georg Alexander Pick (18591942), Austrian mathematician.
2 Carl
11
gon P:
| P Z2 | =
55 1
+ 7 + 1 = 32.
2
2
The polygon contains 32 integral points. You should inspect the drawing to
check that this is true!
v1
v
v3
Chapter 2. Basics
12
We have marked the vertices v1 , v2 and v3 . Notice that T coincides with the
points on line segments from v2 to v, where v is a point on [v1 , v3 ] i.e.
T = {(1 )v2 + ((1 )v1 + v3 ) | [0, 1], [0, 1]}
Clearly (1 ) 0, (1 ) 0, 0 and
(1 ) + (1 ) + = 1.
It is not too hard to check that (see Exercise 2)
T = {1 v1 + 2 v2 + 3 v3 | 1 , 2 , 3 0, 1 + 2 + 3 = 1}.
With this example in mind we define the convex hull of a finite set of vectors.
DEFINITION 2.2.1
Let v1 , . . . , vm R n . Then we let
conv({v1 , . . . , vm }) := {1 v1 + + m vm | 1 , . . . , m 0, 1 + + m = 1}.
We will occasionally use the notation
conv(v1 , . . . , vm )
for conv({v1 , . . . , vm }).
EXAMPLE 2.2.2
To get a feeling for convex hulls, it is important to play around with (lots of)
examples in the plane. Below you see a finite subset of points in the plane.
To the right you have its convex hull.
13
v = 1 v1 + + m v m ,
where 1 , . . . , m , 1 , . . . , m 0 and
1 + + m = 1 + + m = 1.
Then (for 0 1)
u + (1 )v = (1 + (1 )1 )v1 + + (m + (1 )m )vm
where
(1 + (1 )1 ) + + (m + (1 )m ) =
(1 + + m ) + (1 )(1 + + m ) = + (1 ) = 1.
DEFINITION 2.2.3
If X R n , then
conv( X ) =
conv(v1 , . . . , vm ).
m 1
v1 ,...,vm X
Chapter 2. Basics
14
THEOREM 2.2.4
Let X R n . Then conv( X ) is the smallest convex subset containing X.
The theorem follows from the following proposition (see Exercise 4).
PROPOSITION 2.2.5
If S R n is a convex subset and v1 , . . . , vm S, then
conv(v1 , . . . , vm ) S.
Proof. We must show that
1 v1 + + m1 vm1 + m vm S,
where v1 , . . . , vm S, 1 , . . . , m 0 and
1 + + m1 + m = 1.
For m = 2 this is the definition of convexity. The general case is proved using
induction on m. For this we may assume that m 6= 1. Then the identity
1 v1 + + m 1 v m 1 + m v m =
1
m 1
( 1 + + m 1 )
v1 + +
v
( 1 + + m 1 )
( 1 + + m 1 ) m 1
+ m vm
and the convexity of S proves the induction step. Notice that the induction
step is the assumption that we already know conv(v1 , . . . , vm1 ) S for m
1 vectors v1 , . . . , vm1 S.
i I
Xi = { x R n | x Xi , for every i I }.
15
{ X1 , . . . , X m } = ( X i ) i I
with I = {1, . . . , m}. With this out of the way we can state the following.
PROPOSITION 2.2.6
The intersection of a family of convex subsets of R n is a convex subset.
The proof of this proposition is left to the reader as an exercise in the definition of the intersection of subsets. This leads us to the following modern
formulation of conv( X ):
PROPOSITION 2.2.7
The convex hull conv( X ) of a subset X R n equals the convex subset
\
C,
C IX
or
z=y
Chapter 2. Basics
16
17
DEFINITION 2.5.2
cone(v1 , . . . , vm ) := {1 v1 + + m vm | 1 , . . . , m 0}.
Clearly cone(v1 , . . . , vm ) is a cone. Such a cone is called finitely generated.
There is an intimate relation between finitely generated cones and convex
hulls. This is the content of the following lemma.
LEMMA 2.5.3
v conv(v1 , . . . , vm )
v
1
cone
v1
1
vm
,...,
1
( x1 , y1 ) , ( x2 , y2 ) , ( x3 , y3 )
in the plane. Lemma 2.5.3 says that a given point ( x, y) T if and only if
x
y cone
1
x3 !
x2
x1
y1 , y2 , y3
1
1
1
(2.1)
You can solve this problem using linear algebra! Testing (2.1) amounts to
solving the system
1
x1 x2 x3
x
y1 y2 y3 2 = y
1 1 1
3
1
(2.2)
Chapter 2. Basics
18
v1
1
vm
,...,
1
R n +1
is linearly independent.
As a first basic example notice that 1, 1 R are affinely independent (but
certainly not linearly independent).
LEMMA 2.6.2
Let v1 , . . . , vm R n . Then the following conditions are equivalent.
(i) v1 , . . . , vm are affinely independent.
(ii) If
1 v1 + + m v m = 0
and 1 + + m = 0, then 1 = = m = 0.
(iii)
v2 v1 , . . . , v m v1
are linearly independent in R n .
Proof. Proving (i ) = (ii ) is the definition of affine independence. For
(ii) = (iii), assume for 2 , . . . , m R that
2 (v2 v1 ) + + m (vm v1 ) = 0.
Then
1 v1 + + m = 0
with 2 = 2 , . . . , m = m and
1 = (2 + + m ).
In particular, 1 + + m = 0 and it follows that 1 = = m = 0 and
thereby 2 = = m = 0. For (iii ) = (i ) assume that
!
!
v1
vm
1
+ + m
= 0.
1
1
19
Then
0 = 1 v1 + + m v m = 1 v1 + + m v m ( 1 + + m ) v1
= 2 ( v2 v1 ) + + m ( v m v1 )
Chapter 2. Basics
20
v cone(v1 , . . . , vm )
(2.3)
= (1 1 )v1 + + (m m )vm .
(2.4)
Let
= min{ 0 | i i 0, for every i = 1, . . . , m}
o
n
i
| i > 0, i = 1, . . . , m .
= min
i
When you insert into (2.4), you discover that v also lies in the subcone generated by a proper subset of {v1 , . . . , vm }. Now keep repeating this procedure
until the proper subset consists of linearly independent vectors. Basically we
are varying in (2.4) ensuring non-negative coefficients for v1 , . . . , vm until
the first time we reach a zero coefficient in front of some v j . This (or these)
v j is (are) deleted from the generating set. Eventually we end up with a linearly independent subset of vectors from {v1 , . . . , vm }.
5 Constantin
21
A special case of the following corollary is: if a point in the plane is in the
convex hull of 17364732 points, then it is in the convex hull of at most 3 of
these points. When you play around with points in the plane, this seems very
obvious. But in higher dimensions you need a formal proof of the natural
generalization of this!
COROLLARY 2.7.2
Let v1 , . . . , vm R n . If
v conv(v1 , . . . , vm )
cone
v1
1
vm
,...,
1
u
1
cone
1
uk
,...,
1
where
n
u1
1
uk
,...,
1
v1
1
vm
,...,
1
{ x R n | t x 0}
{ x R n | t x 0}.
Chapter 2. Basics
22
DEFINITION 2.8.1
If C R n is a convex cone, we call
C = { R n | t x 0, for every x C }
the dual cone of C.
The subset C R n is clearly a convex cone. One of the main results of these
notes is that C is finitely generated if C is finitely generated. If C is finitely
generated, then
C = cone(v1 , . . . , vm )
for suitable v1 , . . . , vm R n . Therefore
C = { R n | t v1 0, . . . , t vm 0}.
(2.5)
The notation in (2.5) seems to hide the basic nature of the dual cone. Let us
unravel it. Suppose that
a11
a1m
.
.. , . . . , vm = ...
v1 =
an1
anm
and
x1
.
.
=
. .
xn
Then (2.5) merely says that C is the set of solutions to the inequalities
a11 x1 + + an1 xn 0
..
.
(2.6)
a1m x1 + + anm xn 0.
The main result on finitely generated convex cones says that there always
exists finitely many solutions u1 , . . . , u N from which any other solution to
(2.6) can be constructed as
1 u 1 + + N u N ,
where 1 , . . . , N 0. This is the statement that C is finitely generated in
down to earth terms. Looking at it this way, I am sure you see that this is a
non-trivial result. If not, try to prove it from scratch!
23
EXAMPLE 2.8.2
In the picture above we have sketched a finitely generated cone C along with
its dual cone C . If you look closer at the drawing, you will see that
!
!
!
!
2
1
1
2
C = cone
,
and C = cone
,
.
1
2
2
1
Notice also that C encodes the fact that C is the intersection of the two affine
half planes
n
x
y
R2
1
2
!t
x
y
and
x
y
R2
2
1
!t
x
y
o
0 .
Chapter 2. Basics
24
2.9 Exercises
(1) Draw the half plane H = {( x, y)t R2 | a x + b y c} R2 for a = b =
c = 1. Show, without drawing, that H is convex for every a, b, c. Prove in
general that the half space
H = {( x1 , . . . , xn )t R n | a1 x1 + + an xn c} R n
is convex, where a1 , . . . , an , c R.
(2) Let v1 , v2 , v3 R n . Show that
2.9. Exercises
25
This means that the extremal points in a convex hull like X consists of the
indispensable vectors in spanning the convex hull.
(10) What are the extremal points of the convex subset
C = {( x, y) R2 | x2 + y2 1}
of the plane R2 ? Can you prove it?
(11) What is the characteristic cone of a bounded convex subset?
(12) Can you give an example of an unbounded convex set C with ccone(C ) =
{0}?
(13) Draw a few examples of convex subsets in the plane R2 along with their
characteristic cones, extremal points and extremal directions.
(14) Prove Lemma 2.5.3
(15) Give an example of a cone that is not finitely generated.
(16) Prove that you can have no more than m + 1 affinely independent vectors
in R m .
!
7
4
19
8
1
8
1
+
8
1
2
1
+
4
2
2
1
+
2
2
3
of 4 vectors in R2 .
(i) Write v as a convex compbination of 3 of the 4 vectors.
(ii) Can you write v as the convex combination of 2 of the vectors?
(18) Let
!
!
2
1
C = cone
,
.
1
2
(i) Show that
C = cone
!
!
1
2
,
.
2
1
Chapter 2. Basics
26
(ii) Suppose that
C = cone
!
!
a
b
,
,
c
d
where
a b
c d
Chapter 3
Separation
A linear function f : R n R is given by f (v) = t v for R n . For every
R we have an affine hyperplane given by
H = { v R n | f ( v ) = }.
We will usually omit affine in front of hyperplane. Such a hyperplane divides R n into two (affine) half spaces given by
H = { v R n | f ( v ) }
H = { v R n | f ( v ) } .
H> = { v R n | f ( v ) > } .
Separation by half spaces shows the important result that (closed) convex
sets are solutions to systems of (perhaps infinitely many) linear inequalities.
Let me refer you to Appendix B for refreshing the necessary concepts from
analysis like infimum, supremum, convergent sequences and whatnot.
The most basic and probably most important separation result is strict
separation of a closed convex set C from a point x 6 C. We need a small
preliminary result about closed (and convex) sets.
27
Chapter 3. Separation
28
LEMMA 3.0.1
Let F R n be a closed subset. Then there exists x0 F such that
n
o
| x0 | = inf | x|x F .
o
n
= inf | x| x F .
1
1
1
( x0 + y0 ) = x0 + y0 F.
2
2
2
29
where x0 C. If 0 6 C, then
o
n
| x0 | = inf | x| x C ,
x0t z > | x0 |2 /2
for every z C.
Proof. First notice that x0 exists and is unique by Lemma 3.0.1. We argue by
contradiction. Suppose that z C and
x0t z | x0 |2 /2
(3.1)
| x0 |2 | z |2
for every 0 < 1. Letting 0, this implies that x0 = 0 contradicting
our assumption that 0 6 C.
If you study Lemma 3.1.1 closer, you will discover that we have separated
{0} from C with the hyperplane
1
{z R n | x0t z = | x0 |2 }.
2
Chapter 3. Separation
30
x0
C
0
xt0 z = |x0 |2 /2
There is nothing special about the point 0 here. The general result says
the following.
THEOREM 3.1.2
Let C be a closed convex subset of R n with v 6 C and let x0 be the unique point in
C closest to v. Then
1
( x0 v ) t ( z v ) > | x0 v |2
2
for every z C: the hyperplane H = { x R n | t x = } with = x0 v and
= ( x0 v)t v + | x0 v|2 /2 separates {v} strictly from C.
Proof. Let C = C v = { x v | x C }. Then C is closed and convex and
0 6 C . The point closest to 0 in C is x0 v. Now the result follows from
Lemma 3.1.1 applied to C .
With this result we get one of the key properties of closed convex sets.
THEOREM 3.1.3
A closed convex set C R n is the intersection of the half spaces containing it.
Proof. We let J denote the set of all half spaces
H = { x R n | t x }
with C H . One inclusion is easy:
C
H J
H .
In the degenerate case, where C = R n we have J = and the above intersection is R n . Suppose that there exists
x
H J
H \ C.
(3.2)
31
{ x R n | t x 0}.
(1 ) xn + yn z.
Since z is the limit of a convergent sequence with vectors in S, we have shown
that z S.
Chapter 3. Separation
32
DEFINITION 3.2.2
A supporting hyperplane for a convex set C R n at a boundary point z C is
hyperplane H with z H and C H . Here H is called a supporting half space
for C at z.
Below you see an example of a convex subset C R2 along with supporting hyperplanes at two of its boundary points u and v. Can you spot any
difference between u and v?
THEOREM 3.2.3
Let C R n be a convex set and z C. Then there exists a supporting hyperplane
for C at z.
Proof. The fact that z C means that there exists a sequence of points (zn )
with zn 6 C, such that zn z (z can be approximated outside of C). Proposition 3.2.1 says that C is a convex subset. Therefore we may use Lemma 3.0.1
to conclude that C contains a unique point closest to any point. For each zn
we let xn C denote the point closest to zn and put
un =
zn xn
.
|zn xn |
1
|zn xn |
2
(3.3)
for every x C. Since (un ) is a bounded sequence, it has a convergent subsequence. Let u be the limit of this convergent subsequence. Then (3.3) shows
that H = { x R n | ut x = ut z} is a supporting hyperplane for C at z, since
|zn xn | |zn z| 0 as n .
33
2 = sup{t y | y C2 }
and
H = {u R n | t u = 1 }
is the desired hyperplane.
The separation in the theorem does not have to be proper (example?).
However, if C1 6= or C2 6= then the separation is proper (why?).
3.4 An application
We give the following application, which is a classical result [4] due to Gordan2 dating back to 1873.
THEOREM 3.4.1
Let A be an m n matrix. Then precisely one of the following two conditions holds
(i) There exists an n-vector x, such that
Ax < 0.
2 Paul
Chapter 3. Separation
34
(ii) There exists a non-zero m-vector y 0 such that
yt A = 0
Proof. Define the convex subsets
C1 = { Ax | x R n }
and
C2 = {y R m | y < 0}
C2 { x R n | t x }.
(3.4)
(3.5)
(3.6)
35
from Fourier-Motzkin elimination. Using separation in this case, only complicates matters and hides the polyhedral nature of the convex subset in
(3.6).
In teaching convex sets last year, I tried to convince the authors of a certain engineering textbook, that they really had to prove, that a finitely generated cone (like the one in (3.6)) is a closed subset of R n . After 3 or 4 email
notes with murky responses, I gave up.
The key insight is the following little result, which is the finite analogue
of Corollary 3.1.4.
LEMMA 3.5.1
A finitely generated cone
C = cone(v1 , . . . , vr ) R n
is a finite interesection of half spaces i.e. there exists an m n matrix A, such that
C = {v R n | Av 0}.
Proof. Let B denote the n r-matrix with v1 , . . . , vr as columns. Consider the
polyhedron
P = {( x, y) Rr +n | y = Bx, x 0}
y Bx 0 )
( x, y) Rr +n Bx y 0
x
0
Chapter 3. Separation
36
and
yt b < 0.
Proof. The properties (i ) and (ii ) cannot be true at the same time. Suppose
that they both hold. Then we get that
yt b = yt ( Ax) = (yt A) x 0
since yt A 0 and x 0. This contradicts yt b < 0. The real surprise is the
existence of y if Ax = b cannot be solved with x 0. Let v1 , . . . , vm denote
the m columns in A. Then the key observation is that Ax = b is solvable with
x 0 if and only if
b C = cone(v1 , . . . , vm ).
So if Ax = b, x 0 is non-solvable we must have b 6 C. Lemma 3.5.1 shows
that b 6 C implies you can find y R n with yt b < 0 and yt A 0 simply by
using the description of C as a finite intersection of half planes.
3.6. Exercises
37
3.6 Exercises
(1) Give an example of a non-proper separation of convex subsets.
(2) Let
B1 = {( x, y) | x2 + y2 1}
B2 = {( x, y) | ( x 2)2 + y2 1}
(a) Show that B1 and B2 are closed convex subsets of R2 .
(b) Find a hyperplane properly separating B1 and B2 .
(c) Can you separate B1 and B2 strictly?
(d) Put B1 = B1 \ {(1, 0)} and B2 = B2 \ {(1, 0)}. Show that B1 and B2 are
convex subsets. Can you separate B1 from B2 strictly? What about B1
and B2 ?
(3) Let S be the square with vertices (0, 0), (1, 0), (0, 1) and (1, 1) and P =
(2, 0).
(i) Find the set of hyperplanes through (1, 21 ), which separate S from P.
(ii) Find the set of hyperplanes through (1, 0), which separate S from P.
(iii) Find the set of hyperplanes through ( 23 , 1), which separate S from P.
(4) Let C1 , C2 R n be convex subsets. Prove that
C1 C2 := { x y | x C1 , y C2 }
is a convex subset.
(5) Take another look at the proof of Theorem 1.2.2. Show that
( P ) = { x R n 1 | A x 0}
if P = { x R n | Ax 0}, where A and A are matrices with n and n 1
columns respectively.
(6) Show using Farkas that
x
2 0 1
y =
1 1
2
z
!
1
0
Chapter 3. Separation
38
and
C2 { x R n | t x }
(10) Find all the supporting hyperplanes of the triangle with vertices (0, 0), (0, 2)
and (1, 0).
Chapter 4
Polyhedra
You already know from linear algebra that the set of solutions to a system
a11 x1 + + an1 xn = 0
..
.
(4.1)
a1m x1 + + anm xn = 0
of (homogeneous) linear equations can be generated from a set of basic solutions v1 , . . . , vr R n , where r n. This simply means that the set of solutions
to (4.1) is
{1 v1 + + r vr | i R } = cone(v1 , . . . , vr ).
Things change dramatically when we replace = with and ask for a set of
basic solutions to a set of linear inequalities
a11 x1 + + an1 xn 0
..
.
(4.2)
a1m x1 + + anm xn 0.
The main result in this chapter is that we still have a set of basic solutions
v1 , . . . , vr . However here the set of solutions to (4.2) is
{1 v1 + + r vr | i R and i 0} = cone(v1 , . . . , vr )
(4.3)
Chapter 4. Polyhedra
40
0
y
0
z 0
(4.4)
1
0
0 !
0 , 1 , 0
0
0
1
(4.5)
are no longer (basic) solutions. The essence of our next result is to describe
this change.
(4.6)
in the three variables x, y and z. Here the solution set is the set of vectors
with 3x + 4y + 5z = 0 along with the non-negative multiples of just one
vector ( x0 , y0 , z0 ) with
3x0 + 4y0 + 5z0 < 0.
Concretely the solution set to (4.6) is
4
0
4
0
1 !
cone 3 , 3 , 5 , 5 , 1 .
0
0
4
4
1
1 Hermann
2 Hermann
(4.7)
41
The double description method is a systematic way of updating the solution set when we add further inequalities.
x Rn
a1t x
atm x
0 o
..
.
(4.8)
C = cone(v | v V )
for V = {v1 , . . . , v N }. Then
C { x R n | at x 0} = cone(w | w W ),
where (the inequality at x 0 with a R n is added to (4.8))
W ={v V | at v < 0}
(4.9)
Chapter 4. Polyhedra
42
EXAMPLE 4.1.2
Now we can write up what happens to the solutions to (4.4) when we add
the extra inequality x y + z 0. It is simply a matter of computing the set
W of new generators in (4.9). Here
1
0
0 o
n 1
a = 1 and V = 0 , 1 , 0 .
1
0
0
1
Therefore
W=
0
1
0 o
1 , 1 , 1 .
0
0
1
EXAMPLE 4.1.3
Suppose on the other hand that the inequality x y z 0 was added to
(4.4). Then
0
1
1 o
n 0
W = 1 , 0 , 1 , 0 .
0
1
0
1
All of these generators are basic solutions. You cannot leave out a single
of them. Let us add the inequality x 2y + z 0, so that we wish to solve
z
x y z
x 2y +z
0
0
0
0
0
(4.10)
Here we apply Lemma 4.1.1 with a = (1, 2, 1)t and V = W above and the
new generators are (in transposed form)
= (0, 1, 0)
= (1, 1, 0)
(0, 1, 0) + 2(0, 0, 1) = (0, 1, 2)
2(0, 1, 0) + 2(1, 0, 1) = (2, 2, 2)
43
(4.11)
{ x R d | A I x = 0} = {w | R }.
If v = u1 + u2 for u1 , u2 C \ {0}, then we must have I (u1 ) = I (u2 ) = I.
Therefore v, u1 and u2 are all proportional to w. We must have v = u1 or
v = u2 for some > 0 proving that v is an extremal ray.
Chapter 4. Polyhedra
44
DEFINITION 4.2.2
Two extremal rays u and v in
C = { x R d | Ax 0}
x Rn
a1t x
0 o
..
.
atm x 0
(4.13)
45
other hand that w is extremal. This means that AK has rank n 1. By (4.13)
this shows that the rank of A I J has to be n 2 proving that u and v are
adjacent.
We will now revisit our previous example and weed out in the generators
using Lemma 4.2.3.
EXAMPLE 4.2.4
z
x y z
0
0
0
0
(4.14)
Here a1 = (1, 0, 0), a2 = (0, 1, 0), a3 = (0, 0, 1), a4 = (1, 1, 1). The
extremal rays are
V = {(0, 1, 0), (1, 1, 0), (0, 0, 1), (1, 0, 1)}
We add the inequality x 2y + z 0 and form the matrix A with the extra
row a5 := a = (1, 2, 1). The extremal rays are divided into two groups
V = { v | a t v < 0} { v | a t v > 0}
corresponding to
V = {(0, 1, 0), (1, 1, 0)} {(0, 0, 1), (1, 0, 1)}.
You can check that
I ((0, 1, 0)) = {1, 3}
From this you see that (1, 1, 0) is not adjacent to (0, 0, 1) and that (0, 1, 0) is not
adjacent to (1, 0, 1). These two pairs correspond exactly to the superfluous
rays encountered in Example 4.1.3. So Lemma 4.2.3 tells us that we only
need to add the vectors
Chapter 4. Polyhedra
46
Ha .
(4.15)
aC
Ha = C.
aC
EXAMPLE 4.3.1
In Example 4.1.3 we showed that the solutions of
z
x y z
x 2y +z
0
0
0
0
0
(4.16)
47
0
0
0
0
We solve this system of inequalities using Lemma 4.1.1. The set of solutions
to
y
0
x +y
0
y +2z 0
is generated by the extremal rays V = {(2, 2, 1), (1, 0, 0), (0, 0, 1)}. We
add the fourth inequality 3x + 2y + z 0 (a = (3, 2, 1)) and split V according
to the sign of at v:
V = {(2, 2, 1)} {(1, 0, 0), (0, 0, 1)}.
This gives the new extremal rays
3(1, 0, 0) + 3(2, 2, 1) = (3, 6, 3)
and the extremal rays are {(1, 0, 0), (0, 0, 1), (1, 2, 1), (1, 1, 1)} showing that the solutions of (4.16) really are the solutions of
0
z 0
x 2y +z 0
x y z 0.
We needed four inequalities and not five! Of course we could have spotted
this from the beginning noting that the inequality y 0 is a consequence
of the inequalities x 0, x 2y + z 0 and x y z 0.
Chapter 4. Polyhedra
48
(4.17)
am1 x1 + + amn xn bm
of linear inequalities is called a polyhedron. It may come as a surprise to you,
but we have already done all the work for studying the structure of polyhedra. The magnificent trick is to adjoin an extra variable xn+1 and rewrite
(4.17) into the homogeneous system
(4.18)
xn+1 0.
( x1 , . . . , xn ) solves (4.17)
( x1 , . . . , xn , 1) solves (4.18).
cone
u1
0
,...,
ur
0
v1
,
1
vs
,...,
1
!!
(4.19)
49
x 2
Adjoining the extra variable y this lifts to the system
xy 0
x + 2y 0
y 0
(4.20)
On the other hand, if P is the sum of a cone and a polytope as in (4.20), then
we define P R n+1 to be the cone generated by
!
!
!
!
u1
ur
v1
vs
,...,
,
,...,
0
0
1
1
Chapter 4. Polyhedra
50
If
P = cone
1
1
,...,
N
N
P=
R
1 x + 1 z 0, . . . , tN x + N z 0 .
z
Therefore
But an element x R n belongs to P if and only if ( x, 1) P.
o
n
P = x R n 1t x + 1 0, . . . , tN x + N 0
o
n
n t
= x R 1 x 1 , . . . , tN x N
is a polyhedron.
P = { x R n | Ax b} =
x Rn
a1t x b1
..
.
atm x bm
51
P=
x
y
1
0 o
x
2
R 2 1
1 .
y
1
2
1
We will find the first extremal point and leave the computation of the other
extremal points to the reader. First we try and see if we can find z P with
!
1 1
A z = { a1 , a2 } =
.
2 1
If z = ( x, y)t this leads to solving
x y = 0
2x y = 1,
giving ( x, y) = (1/3, 1/3). Since 1/3 + 2 (1/3) = 1/3 < 1 we see that
z = (1/3, 1/3) P. This shows that z is an extremal point in P. Notice that
the rank of Az is 2.
DEFINITION 4.6.3
A subset
L = {u + tv | t R }
with u, v R n and v 6= 0 is called a line in R n .
THEOREM 4.6.4
Let P = { x R n | Ax b} 6= . The following conditions are equivalent
(i) P contains an extremal point.
(ii) The characteristic cone
ccone( P) = { x R n | Ax 0}
does not contain a line.
Chapter 4. Polyhedra
52
(iii) P does not contain a line.
53
1
1
(z + u) + (z u)
2
2
Chapter 4. Polyhedra
54
4.7 Exercises
(1) Verify that the set of solutions to (4.6) is as described in (4.7).
(2) Find the set of solutions to the system
x +
z 0
y + z 0
z 0
1/4
1/2
1/4
1 o
n
conv 1/4 , 1/4 , 1/2 , 1 R3
1/2
1/4
1/4
1
as a polyhedron (an intersection of halfspaces).
(4) Convert the inequalities
x + y 1
x + y 1
x 2y 2
to a set of 4 homogeneous inqualities by adjoining an extra variable z.
Show that the original inequalities are unsolvable using this.
(5) Is the set of solutions to
x + 2y z 1
x y z 2
2x y z 1
y + z 1
x y + z 0
x + y
1
bounded in R3 ?
(6) Let P1 and P2 be polytopes in R n . Show that
P1 + P2 = {u + v | u P1 , v P2 }
is a polytope. Show that the sum of two polyhedra is a polyhedron.
4.7. Exercises
55
x K
Appendix A
Linear (in)dependence
The concept of linear (in)dependence is often a stumbling block in introductory courses on linear algebra. When presented as a sterile definition in an
abstract vector space it can be hard to grasp. I hope to show here that it
is simply a fancy way of restating a quite obvious fact about solving linear
equations.
A.1
(A.1)
x+y+z = 0
a21 x1 + + a2n xn = 0
..
.
am1 x1 + + amn xn = 0
57
58
a21 x1 + + a2n xn = 0
..
.
(A.2)
am1 x1 + + amn xn = 0
of linear equations always has a non-zero solution if m < n.
Proof. The induction is on m the number of equations. For m = 1 we have
1 linear equation
a1 x1 + + a n x n = 0
with n variables where n > m = 1. If ai = 0 for some i = 1, . . . , n then clearly
xi = 1 and x j = 0 for j 6= i is a non-zero solution. Assume otherwise that
ai 6= 0 for every i = 1, . . . , m. In this case x1 = 1, x2 = a1 /a2 , x3 = =
xn = 0 is a non-zero solution.
If every ai1 = 0 for i = 1, . . . , m, then x1 = 1, x2 = = xn = 0 is a
non-zero solution in (A.2). Assume therefore that a11 6= 0 and substitute
x1 =
1
( a12 x2 a1n xn )
a11
59
( a22
(A.3)
1
( a12 a2 a1n an ), a2 , . . . , an )
a11
A.2
A system
a11 x1 + + a1n xn = 0
a21 x1 + + a2n xn = 0
..
.
(A.4)
am1 x1 + + amn xn = 0
of linear equations can be conveniently presented in the matrix form
x1
0
. .
. .
A
. = . ,
xn
0
where A is the m n matrix
a11 a1n
.
..
..
..
.
.
.
am1 amn
We need to attach a very important invariant to A called the rank of the matrix. In the context of systems of linear equations the rank is very easy to
understand. Let us shorten our notation a bit and let
Li ( x) = ai1 x1 + + ain xn .
60
Then the solutions to (A.4) are
S = { x R n | L1 ( x) = 0, . . . , Lm ( x) = 0}.
Suppose that m = 3. If one of the equations say L3 ( x) is expressible by the
other equations say as
L3 ( x) = L1 ( x) + L2 ( x),
with , R, then we dont need the equation L3 in S. This is because
L1 ( x ) = 0
L2 ( x ) = 0
L3 ( x ) = 0
L1 ( x ) = 0
L2 ( x ) = 0
S = { x R n | A I x = 0},
where A I is an m n- matrix consisting of a subset of the rows in A with
m = the rank of A. Now the result follows by applying Theorem A.1.1.
A.3. Exercises
A.3
61
Exercises
5
6
!
0
.
0
1
( a12 a2 a1n an ), a2 , . . . , an )
a11
1 2 3
4 5 6
14 19 24 .
6 9 12
Appendix B
Analysis
In this appendix we give a very brief overview of the basic concepts of introductory mathematical analysis. Focus is directed at building things from
scratch with applications to convex sets. We have not formally constructed
the real numbers.
| x| =
x12 + + x2n .
Our first result about the length is the following lemma called the inequality
of Cauchy-Schwarz. It was discovered by Cauchy1 in 1821 and rediscovered
by Schwarz2 in 1888.
LEMMA B.1.1
For x = ( x1 , . . . , xn )t R n and y = (y1 , . . . , yn )t R n the inequality
63
Appendix B. Analysis
64
Proof. For n = 2 you can explicitly verify that
(B.1)
This implies as you can check that there exists R such that x1 = y1 and
x2 = y2 .
The formula in (B.1) generalizes for n > 2 by induction (Exercise 1) to
(B.2)
( x1 y2 y1 x2 ) + + ( x n 1 y n y n 1 x n ) ,
where the last sum is over the squares of the 2 2 minors in the matrix
A=
x1 x2 x n 1 x n
y1 y2 y n 1 y n
R3
| u |2 | v |2 = | u t v |2 + | u v |2 .
One of the truly fundamental properties of the length of a vector is the
triangle inequality (also inspired by the one in 2 dimensions).
THEOREM B.1.2
For two vectors x, y R n the inequality
| x + y| | x | + |y|
holds.
B.2. Sequences
65
| x + y |2 = ( x + y ) t ( x + y )
| x y |.
From Theorem B.1.2 you get formally
| x z| = | x y + y z| | x y| + |y z|
for x, y, z R n . This is the triangle inequality for distance saying that the
shorter way is always along the diagonal instead of the other two sides in a
triangle:
y
|x y|
|y z|
x
|x z|
B.2 Sequences
Limits appear in connection with (infinite) sequences of vectors in R n . We
need to formalize this.
DEFINITION B.2.1
A sequence in R n is a function f : {1, 2, . . . } R n . A subsequence f I of f is f
restricted to an infinite subset I {1, 2, . . . }.
Appendix B. Analysis
66
(B.3)
2, 4, 6, 8, 10, . . .
(B.4)
2, 6, 4, 8, 10, . . .
(B.5)
1 1 1
, , , ...
2 3 4
(B.6)
1,
x2
x3
x4
x5
x6
(B.3)
(B.4)
1
2
2
4
3
6
4
8
5
10
6
12
(B.5)
(B.6)
2
1
10
12
1
2
1
3
1
4
1
5
1
6
B.2. Sequences
67
> 0 N N n N : | x xn | .
Such a sequence is called convergent.
This is a very formal (but necessary!) way of expressing that . . .
the bigger n gets, the closer xn is to x.
You can see that (B.3) and (B.4) are not convergent, whereas (B.6) converges to 0. To practice the formal definition of convergence you should (Exercise 4) prove the following proposition.
PROPOSITION B.2.6
Let ( xn ) and (yn ) be sequences in R n . Then the following hold.
(i) If xn x and xn x , then x = x .
(ii) If xn x and yn y, then
xn + yn x + y
and
xn yn xy.
Before moving on with the more interesting aspects of convergent sequences we need to recall the very soul of the real numbers.
Appendix B. Analysis
68
B.2.1
69
COROLLARY B.3.1
A bounded sequence of real numbers has a convergent subsequence.
Proof. This is a consequence of Lemma B.2.4 and Lemma B.2.8.
We want to generalize this result to R m for m > 1. Surprisingly this is not
so hard once we use the puzzling properties of infinite sets. First we need to
define bounded subsets here.
A subset S R m is called bounded if there exists R > 0 such that | x| R
for every x S. This is a very natural definition. You want your set S to be
contained in vectors of length bounded by R.
THEOREM B.3.2
A bounded sequence ( xn ) in R m has a convergent subsequence.
Proof. Let the sequence be given by
xn = ( x1n , . . . , xmn ) R m .
The m sequences of coordinates ( x1n ), . . . , ( xmn ) are all bounded sequences
of real numbers. So the first one ( x1n ) has a convergent subsequence ( x1ni ).
Nothing is lost in replacing ( xn ) with its subsequence ( xni ). Once we do this
we know that the first coordinate converges! Move on to the sequence given
by the second coordinate and repeat the procedure. Eventually we end with
a convergent subsequence of the original sequence.
Appendix B. Analysis
70
The points of S are simply the points you can reach with convergent sequences from S. Therefore the following result must be true.
PROPOSITION B.4.3
Let S R n . Then S is closed.
Proof. Consider a convergent sequence (ym ) S with ym y. We wish to
prove that y S. By definition there exists for each ym a convergent sequence
( xm,n ) S with
xm,n ym .
For each m we pick xm := xm,n for n big enough such that |ym xm | < 1/m.
We claim that xm y. This follows from the inequality
| y x m | = | y y m + y m x m | | y y m | + | y m x m |,
using that ym y and |ym xm | being small for m 0.
The following proposition comes in very handy.
PROPOSITION B.4.4
Let F1 , . . . , Fm R n be finitely many closed subsets. Then
F := F1 Fm R n
is a closed subset.
Proof. Let ( xn ) F denote a convergent sequence with xn x. We must
prove that x F. Again distributing infinitely many elements in finitely
many boxes implies that one box must contain infinitely many elements.
Here this means that at least one of the sets
N i = {n N | xn Fi }, i = 1, . . . , m
must be infinite. If N k inifinite then { x j | j N k } is a convergent (why?)
subsequence of ( xn ) with elements in Fk . But Fk is closed so that x Fk F.
71
DEFINITION B.5.1
Let S R n . Then the interior S of S is
R n \ R n \ S.
The boundary S is
S R n \ S.
|| x| |y|| | x y|
for every x, y R n . This shows that
| f ( x) f ( xn )| | x xn |
proving that f ( xn ) f ( x) if xn x. Therefore f ( x) = | x| is a continuous
function.
The following result is very useful for proving that certain subsets are
closed.
LEMMA B.6.3
If f : R m R n is continuous then
f 1 ( F ) = { x R n | f ( x ) F } R m
is a closed subset, where F is a closed subset of R n .
Proof. If ( xn ) f 1 ( F ) with xn x, then f ( xn ) f ( x) by the continuity of
f . As F is closed we must have f ( x) F. Therefore x f 1 ( F ).
Appendix B. Analysis
72
f (y) = sup{ f ( x) | x C }.
In more boiled down terms, this corollary says that a real continuous
function on a compact set assumes its minimum and its maximum. As an
example let
f ( x, y) = x18 y113 + 3x cos( x) + ex sin(y).
A special case of the corollary is that there exists points ( x0 , y0 ) and ( x1 , y1 )
in B, where
B = {( x, y) | x2 + y2 117}
such that
f ( x0 , y0 ) f ( x, y)
f ( x1 , y1 ) f ( x, y)
(B.7)
73
for every ( x, y) B. You may say that this is clear arguing that if ( x0 , y0 ) does
not satisfy (B.7), there must exist ( x, y) B with f ( x, y) < f ( x0 , y0 ). Then
put ( x0 , y0 ) := ( x, y) and keep going until (B.7) is satisfied. This argument
is intuitive. I guess that all we have done is to write it down precisely in the
language coming from centuries of mathematical distillation.
Appendix B. Analysis
74
B.8 Exercises
(1) Use induction to prove the formula in (B.2).
(2)
|| x| |y|| | x y|
for every x, y R n .
(10) Give an example of a subset S R and a continuous function f : S R,
such that f (S) is not bounded.
Appendix C
xR
n
Ax
b
Ax b
x
0
(C.1)
C.1
Extremal points
The determination of extremal points of polyhedra in standard form is a direct (though somewhat laborious) translation of Proposition 4.6.1.
THEOREM C.1.1
Let A be an m n matrix af rank m. There is a one to one correspondence between
extremal points in
P = { x R n | Ax = b, x 0}
and m linearly independent columns B in A with B1 b 0. The extremal point
corresponding to B is the vector with zero entries except at the coordinates corresponding the columns of B. Here the entries are B1 b.
Proof. Write P as in (C.1):
n
xR
n
Axb
75
76
where
A = A ,
I
P=
! x
x
1 2 3
y
y =
4 5 6
z
z
!
o
1
, x, y, z 0 .
3
! 1
1
3
2 3
5 6
! 1
1
3
!
1/3
,
1/3
!
1
.
1/3
1 3
4 6
! 1
1
3
!
1/2
,
1/6
From this you see that P only has the two extremal points
x1
1/3
y1 = 1/3
z1
0
and
x2
1/2
y2 = 0 .
z2
1/6
P=
! x
x
1 2 3
y
y =
4 5 6
z
z
!
o
1
, x, y, z 0 ,
1
77
1
1
!
1/2
,
1/2
C.2
Extremal directions
The next step is to compute the extremal infinite directions for a polyhedron
in standard form.
THEOREM C.2.1
Let A be an m n matrix of rank m. Then the extremal infinite directions for
P = { x R n | Ax = b, x 0}
are in correspondence with ( B, a j ), where B is a subset of m linearly independent
columns in A, a j a column not in B with B1 a j 0. The extremal direction corresponding to ( B, a j ) is the vector with entry 1 on the j-th coordinate, B1 a j on the
coordinates corresponding to B and zero elsewhere.
Proof. Again let
P=
where
x Rn
A x b
A = A ,
I
78
EXAMPLE C.2.2
The following example comes from the 2004 exam. Consider the polyhedron
( x
P = y R3
z
x + 2y 3z = 1
3x + y 2z = 1
x
0
y
0
z 0
! 1
2
1
! 1
1
3
7
5 ,
1
!
7
.
5
and
5
7
71
57
1
7 .
5
Using Theorem C.1.1 you can check that the only extremal point of P is
1
5
2
5 .
0
1 , 2 , 3 0 .
C.3. Exercises
C.3
79
Exercises
(
y
5
P=
z R
u
v
x
y 2z u
x +2y +4z 2u +v
x
y
z
u
v
=
=
1
1
0
0
0
0
0
Bibliography
[1] Julius Farkas, Theorie der einfachen Ungleichungen, J. Reine Angew. Math.
124 (1901), 127.
[2] Joseph Fourier, Solution dune question perticuli`ere du calcul des inegalites,
Nouveau Bulletin des Sciences par la Societe philomatique de Paris
(1826), 317319.
[3] Komei Fukuda and Alain Prodon, Double description method revisited,
Combinatorics and computer science (Brest, 1995), Lecture Notes in Comput. Sci., vol. 1120, Springer, Berlin, 1996, pp. 91111.
[4] Paul Gordan, Ueber die Auflosung linearer Gleichungen mit reellen Coefficienten, Math. Ann. 6 (1873), no. 1, 2328.
[5] Hermann Minkowski, Allgemeine Lehrsatze uber
81
Index
Az , 58
H< , 31
H> , 31
H , 31
H , 31
I ( x) indices of equality (cone),
50
S , for subset S R n , 81
ccone(C ), 19
cone(v1 , . . . , vm ), 20
conv( X ), 16
conv({v1 , . . . , vm }), 14
conv(v1 , . . . , vm ), 14
ext(C ), 18
S, for subset S R n , 80
S, for subset S R n , 81
u v for u, v R n , vii
u v, 74
ut v for u, v R n , vii
Index
84
convex subset
closure of, 36
definition, 11
intersection of half spaces, 35
cross product, 74
distance in R n , 75
double description method, 46
dual cone, 25
extremal direction, 19
in polyhedron of standard form,
89
extremal point, 18
in polyhedra, 59
in polyhedron, 58
in polyhedron of standard form,
87
extremal ray, 19
adjacent, 51
condition, 50
Farkas lemma, 39
Farkas, Gyula, 39
finitely generated cone, 20
Fourier, J., 5
Fourier-Motzkin elimination, 5
Gauss, Carl Friedrich, 12
Gordans theorem, 38
Gordan, Paul, 38
half space, 31
half space, 25
affine, 31
supporting, 36
homogeneous system, vii
hyperplane, 25
affine, 31
supporting, 36
independence
affine, 21
infimum, 78
integral points
in polygon, 12
inside circle, 12
interior, 81
intersection of half spaces, 35
intersection of subsets, 17
length function
is continuous, 82
length of a vector, 73
limit concept, 73
line in R n , 60
line segment, 11
linear inequalities
homogeneous, vii
linear dependence, 66
linear equations, 66
linear function, 31
linear independence, 65
linear inequalites
homogeneous, vii
linear inequalities, vii, 2
solving using elimination, 3
magnificent trick, 56
mathematical analysis, 73
Minkowski, Hermann, 46, 57
Minkowski-Weyl theorem, 57
Motzkin, T. S., 5
Ostrowski, A., 5
Picks formula, 12
Pick, Georg Alexander, 12
polyhedral convex sets, 12
polyhedron, 5
in standard form, 87
Index
of standard form
extremal directions in, 89
extremal points in, 87
polyheral cone, 41
polytope, 5
Pythagorean formula, 73
rank of a matrix, 68
recession cone, 19
Schwarz, Hermann Amandus, 73
separation of convex subsets, 31
disjoint convex sets, 37
from a point, 33
proper, 31
strict, 31
sequence, 76
convergent, 77
decreasing, 77
increasing, 77
simplex, 22
simplicial cone, 23
standard form of polyhedron, 87
subsequence, 76
supporting half space, 36
supporting hyperplane, 36
existence of, 37
supremum, 78
tetrahedron, 22
triangle inequality, 74
vector product, 74
Weyl, Hermann, 46, 57
85