GENERALIZATIONS
OF
THE SINGULAR
VALUE
AND
QR
DECOMPOSITIONS*
BART DE MOOR AND PAUL VAN DOOREN$
We dedicate thispaper to Gene Golub, a true source
of
inspirationfor
ourwork, but alsoagenuinefriend, onthe occasion
of
his60th birthdayAbstract. This paper discusses multimatrix generalizations of twowell-knownorthogonalrank factorizations ofa matrix: the generalized singular valuedecomposition and thegeneralized QR-(or URV-) decomposition. These generalizations can be obtained for any number ofmatrices of compatible dimensions. This paper discussesin detail the structure of these generalizations and theirmutual relations and givesaconstructiveprooffor the generalized QR-decompositions.
Key words, singularvalue decomposition, QR-factorization, URV-decomposition, complete orthogonaldecomposition
AM$(MOS) subject classifications. 15A09, 15A18, 15A21, 15A24,65F20
1. Introduction.
In
this paper,wepresent multimatrixgeneralizations ofsomewell-knownorthogonalrank factorizations.
We
showhow the ideaofaQR-decompo-sition
(QRD),
a URV-decomposition(URVD),
and a singular value decomposition(SVD)
for one matrix can be generalized to any number of matrices. Whilegener-alizations of the
SVD
for any number ofmatrices have been derived in[9],
one ofthe main contributionsofthispaperis the constructive derivationofageneralization for the
QRD
(or URVD)
for any number ofmatrices of compatible dimensions. The idea is to reduce the set ofmatricesA1, A2,...,
Ak
to a simpler form using unitarytransformationsonly. Hereby, weavoidexplicitproducts andinverses ofthe matrices that are involved.
We
show that these generalized QR-decompositions(GQRD)
canbeconsidered as apreliminaryreduction for anygeneralized singular value
decompo-sition
(GSVD).
The reason is thatthere is a certainone-to-one relationbetweenthe structure ofaGQRD
and the "corresponding"GSVD,
which is explained in detail below.This paper is organized as follows.
In
2,
we provide a summary of orthogonalrank
factorizations
for one matrix.We
briefly review theSVD,
theQRD,
and theURVD
asspecial cases.In
3,
wegiveasurvey o1existinggeneralizations of theSVD
and
QRD
for two or three matrices.In
4,
wesummarize the results onGSVDs
foranynumber ofmatricesofcompatible dimensions. Section
5,
which containsthemain newcontributionofthis paper, describes ageneralization oftheQRD
and theURVD
for any number of matrices.We
derive a constructive, inductive proofwhich showsthata
GQRD
can be usedas apreliminaryreduction for acorrespondingGSVD.
In
6,
weanalyzeindetail thestructureof theGQRDs
andGSVDs
and show thatthere is aone-to-one relation between the two generalizations. Thisrelation is elaboratedin more detail in
7,
where weillustrate howaGQRD
can be used as apreliminaryStep
inthederivationofacorrespondingGSVD.
Receivedbythe editorsApril 16, 1991; accepted forpublication (inrevisedform) October 18,
1991. This researchwas partially supported bytheBelgian Programon Interuniversity Attraction Poles andtheEuropeanCommunityResearchProgram ESPRIT, BRA3280.
Departmentof Electrical Engineering, Katholieke UniversiteitLeuven, B-3001Leuven,Belgium (demoor@esat.kuleuven.ac.be). Thisauthoris aResearchAssociate of the Belgian National Fund for ScientificResearch(NFWO).
Coordinated Science Laboratory, University of Illinois, Urbana, Illinois 61801 (vdoorenOuicsl.csl.uiuc.edu).
While all results inthis paper are stated for complex matrices, they canbe spe-cializedto thereal casewithout much difficulty. Thiscanbe done inmuch the same way as with the
SVD
for complex and real matrices.In
particular, it suffices to re-state most results using the term real orthonormalinsteadof unitaryand to replacea superscript "."
(which
denotes thecomplex conjugate transpose ofamatrix)
by a superscript"t" (which
is the transpose ofamatrix).
2. Orthogonal rankfactorizations.
Any
matrixA
EC
mn canbe factorizedas
where
R
C
nxn is upper trapezoidal andH
is a real n x n permutation matrix that permutes the columns ofA
so that the firstra
rank(A)
columns are linearly independent. The matrixQ
C
mxm isunitary andcanbe partitionedasIFa m
ra
).
If we partition
R
accordingly asR
(Rll
R12),
whereRll
C
rxra is uppertriangularandnonsingular, weobtain
A=QI(RI Rle)H
whichis sometimes cMled the
QR-factorization
ofA.
Ifwe rewrite(1)
asQ’A-(
R
weseethat
Q
isanorthogonal transformationthatcompressestherows ofA.
There-fore,
it is called a row compression.A
similar construction exists, of course, for acolumncompression.
A
complete orthogonalfactorization
ofanmx matrixA
is any factoriation of the form(2)
A
U
(
T
0V*
where
T
israx
ra
square nonsingular ndra
rank(A).
One
particular case istheSVD,
which has become an important tool in the nalysis and numericM solution of numerous problems, especially since the development of numericMly robustalgo-rithms by Golub ndhiscoworkers
[15], [16],
[17].
TheSVD
is complete orthogonMfactorizationwhere thematrix
T
is diagonalwith positive diagonM elements:A
UEV*.
Here U
C
mxm andV
C
nxnare unitary and E
mxn
is oftheform(71 0 0 0
0 a2 0 0
0 0
rra
00 0 0 0
In this paper, we use the convention that zero blocks may be "empty" matrices, i.e., certain block dimensions may be 0.
GENERALIZED SVD AND QR 995 The positive numbers 0-1
_
0"2_
_
0-ra>
0 are called the singular values ofA,
whilethe columns ofU
andV
aretheleft
and right singular vectors.In
applications where m>>
n,it isoftenagoodideato use theQRD
ofthematrix as a preliminary step in the computation ofitsSVD.
TheSVD
ofA
isobtained viathe
SVD
ofits triangular factor asA
QR-
Q(UrE.V?)
(QUr)ErV?.
This idea of combining the
QRD
and theSVD
of the triangular matrix, in order to compute theSVD
ofthe full matrix, is mentioned in[22,
p.119]
and morefully
analyzedin
[3].
In
[18]
the methodisreferredtoasR-bidiagonalization.Its
flopcount is(mn2+n3),
ascomparedto(2mn2-2/3n3)
forabidiagonalization ofthefullmatrix.Hence,
whenever m_>
5/3n,
it is more advantageousto use the R-bidiagonalization algorithm.There exist still other complete orthogonal factorizations ofthe form
(2)
whereonly
T
is required to be triangular(upper
orlower) (see,
e.g.,[18]).
Sucha factoriza-tionwas called aURV-decompositionin[27].
Here
where
U
EC
m mV
C
gularuppertriangular.
are unitarymatrices and
R
C
raxr is squarenonsin-It
is well known that the QR-factorization of a singular matrixA
and of itstranspose A* canbeused for findingthe image and kernelof
A
(URV-decompositions
actually give both at
once).
In
this paper, we try to extend these ideas to several matrices.Suppose
we have a sequence ofmatricesAi,
1,...,
k,
and we want toknowthekernels
(or
nullspaces)
of eachpartialproductA1. A2...
Aj.
We
could form these products andcompute QR-decompositions ofeach ofthem. That can, infact,
be avoided, as shown below.
Let
us take the "special" exampleAi
A,
1, 2, 3,with 0 1 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0
It
iswellknown that the null spaces ofA
infactgive the Jordan structureofA,
and thisstructure is alreadyobviousfrom the form ofA.
But
letusreconstructit from a sequence ofQR-decompositions(in
factweneed hereRQ-decompositions ofA).
The first oneis, ofcourse, acolumn compression ofA1,
for whichwe usethe permutationof columns 2 and 4
(denoted
by thematrixP24):
0 0 0 0 0 0 1 0 1 0 0 0 0 0 0 0 1 0 0 0
The separation line here indicatesthat the first two columns of
P24
(i.e.,
el andea)
span thekernelof
A
A.
For
the kernelofA
2butapply theinverseof the orthogonal transform
P24 (which
is againP24)
tothe rows ofA2
A:
0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0Since
A1A2
(AIP24)(P24A2),
it is clear that the kernel ofAIA2
is also the kernelofthe bottom part of
P24A2.
The following column compression ofP24A2
actually yields the kernel of bothA2
and theproductAA2.
Perform indeedthe orthogonal transformationP24P35:
P24A2P24P35
0 0 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0We
see thatthekernelofA2
comprisesthefirst twocolumns ofP24P35
(i.e.,
el ande4as
before)
and the kernel ofA1A2
comprises the firstfour columns ofP24P35,
i.e., e2, e4, and e5.An
additional step ofthis procedure finally shows that the kernel ofthe product
A1A2A3
(AIP24)(P24A2P24P35)(P35P24A3)
is thatofthe bottom part of thematrix 0 1 0 0 0 0 0 0 0 1P35P24A3--
0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 which is a zero matrix.Hence
the kernel ofA
3is the whole space, as expected.
The interestingpart ofthis simple example"is the fact that we have not formed the
intermediate products to get their corresponding kernels. The case treated here of equal matrices
Ai
is asimple one(and
could besolvedusing theresults of[19]),
but in the nextfewsections we show how thiscanalso be done forarbitrarysequencesofmatrices. The key idea is that at each step wedo anumber of QR-factorizations on
the blocks ofapartitioned matrix
(column
blocks in ourcase).
This then induces a new partitioning onthe rows ofthis matrix,on the columns of thenext matrix, andso on.
3. Generalizations for two or three matrices.
In
the last decade or so, several generalizations for theSVD
have been derived. The motivation is basically the necessity to avoidthe explicit formationof products and matrix quotients inthecomputation of the
SVD
of products and quotients of matrices.Let
A
andB
benonsingular square matrices and assume that weneed the
SVD
ofAB-*
USV*.
2It
is well known that the explicit calculation ofB
-1,
followed by the computationof the product, may result in loss of numerical precision
(digit
cancellation),
evenbefore any factorization is
attempted!
The reason is the finite machine precision of Thenotation B-* refers tothe complexconjugatetransposeof theinverseofthematrixB.any calculator.
Therefore,
it seems more appropriate to come up with an implicit combined factorizationofA
andB
separately, such asA
UD1X
-1,
(3)
B=X-*D2V*,
where
U
andV
are unitaryandX
nonsingular. The matricesD1
andD2
arereal but"sparse" (quasi-diagonal),
and theproductDID
isdiagonalwith positivediagonal elements. ThenwefindAB-*
UDIX-1XDtV
U(DDt)v
*.
A
factorization as in(3)
is always possible for two square nonsingular matrices.In
fact,
it is always possible for two matricesAE
C
mn andB
E
C
nv(as
long as the numberof columns ofA
isthesame asthe numberofrowsofB,
which werefertoas acompatibility
condition).
In
general, thematricesA
andB
mayevenberank deficient. The combined factorization(3)
is called the quotient singular value decomposition(QSVD)
andwas firstsuggestedin[32]
andrefined in[23]
(it
wasoriginally calledthegeneralized
SVD,
but wehave suggested astandardizednomenclature in[6]).
A
similar idea might be exploited for theSVD
ofthe product of two matricesAB
USV*,
viatheso-called product singular value decomposition(PSVD)
A
UD1
X-l,
(4)
B
XD2V*,
sothat
AB
U(DID2)V*,
which is anSVD
ofAB.
The combined factorization(4)
was proposed in
[13]
as aformalization ofideas in[21].
In
thegeneral case, for twocompatiblematrices
A
andB
(which
mayberankdeficient),
thePSVD
of(4)
alwaysexistsandprovidesthe
SVD
ofAB
without theexplicit construction ofthe product. Similarly, ifA
andB
are compatible, theQSVD
always
exists.However,
it does not always deliver theSVD
ofABt
whenB
is rank deficient(Bt
is the pseudoinverseof
B).
Another generalization, this time for three matrices, is the restricted singular
value decomposition
(RSVD).
It
was proposedn
[35],
and numerous applications were reviewedin[7].
It
was soon found that allofthesegeneralizedSVDs
for twoor three matricesarespecialcasesofageneraltheorem,
presentedin[9].
Themainresultis that there exist
GSVDs
for any number ofmatricesAI,A.,...,Ak
of compatible dimensions. The general structure of theseGSVDs
wasfurtheranalyzed
in{10].
Thedimensions ofthe blocks that occur in any
GSVD
can be expressed asranks of the matricesinvolved and ascertain products and concatenationsof these.We
present a summary ofthe resultsbelow.As
for generalizations of theQRD,
it is mainly Paige{25]
who pointed out the importance of generalizedQRDs
for two matrices as a basic conceptual and math-ematical tool. The motivation is that in some applications, we need theQRD
ofaproduct oftwo matrices
AB
whereA
mx
andB
nxp.
For
general matricesA
andB
such a computation avoids forming the product explicitly, and transformsA
andB
separatelyto obtain the desired results. Paige[25]
refers to such a factor-ization as a productQR
factorization.
Similarly, in some applications we need theQR-factorization of
AB
-
whereB
is square andnonsingular.A
general numerically robust algorithm would not compute theinverse ofB
northe product explicitly, but would transformA
andB
separately. Paige[25]
proposed calling such a combineddecompositionoftwo matricesageneralized
QR
factorization,
following[20].
We
pro-pose here toreservethename generalizedQRD
for the completeset ofgeneralizations of the QR-decompositions,which aredevelopedin thispaper.We
also propose anovelnomenclature,
as wedidfor the generalizations oftheSVD
in[6].
Stoer
[28]
appears tobethefirst to havegivenareliable computationofthistype of generalized QR-factorization for two matrices(see
[14]).
Computational methods for producingthetwo types of generalizedQR
factorizations for twomatrices, as de-scribedabove,
haveappearedregularlyin theliteratureas(intermediate)
stepsinthesolutionofsomeproblems.
In
this paper, wederiveaconstructiveproofofgeneraliza-tionsofthe
QRD
for any number ofmatrices.As
weseebelow,
ourgeneralizedQRDs
can also beconsideredthe appropriate generalization oftheURVD
ofa matrix.4. Generalizedsingularvaluedecompositions.
In
thissection, wepresentageneral theoremthatcanbeconsideredthe appropriategeneralization forany number
ofmatrices of the
SVD
ofonematrix.It
contains the existing generalizations of theSVD
for two matrices(i.e.,
thePSVD
and theQSVD)
and three matrices(i.e.,
theRSVD)
asspecial cases.A
constructiveproofcan befoundin[9].
THEOREM 4.1
(generalized
singular value decompositions for kmatrices).
Con-sideraset
of
k matrices with compatible dimensions:A1
(no
nl),A2 (nl
n2),...,
Ak-1
(nk-2
nk-1),Ak (nk-1
nk).
Then there exist--Unitarymatrices
UI
(no
no)
andVk (nk
nk).
--Matrices
Dj,
j 1,2,...,
(k
1)
of
theform
nj_ nj where
(6)
(7)
rj_ rj 2 rj 2 2 rj_ rj rj J rj J nj-1 rj-1 rj 2 3 J nj -rj [rj rj rj rjI
0 0 0 0 0 0 0 0 0 0I
0 0 0 0 0 0 0 0 0 0I
0 0 0 0 0 0I
0 0 0 0 0ro
O,
A
matrixSk
of
theform
rk rk- rk
r
2r_l
rk
rk Jrank(Aj).
rjE
rj i=1 kr r r
rk nk rko
o
o
o
0 0 0 0 0 0S
0 0 0 0 0 0 0 0 0 0S
0 0 0 0 0 nk-l rk-l rk\
0 0where
(8)
rkE r
rank(A)
i--1
and the
rk rk
matricesS
are diagonalwith positivediagonal elements. Expressionsfor
the integers rj aregiven in6
aNonsingularmatrices
Xj
(n
n)
andZi,
j1,
2,...,(k-
1)
whereZ
is eitherZj
X*
orZ
Xi
(i.e.,
both choices are alwayspossible),
such that thegiven matrices can be
factorized
asA1
U1D1X 1,
A2
ZID2X
1,
A3-
Z2D3X
1,
Ai-
Zi-IDiX-1,
Ak
Zk- Sk
V;
Observe that the matrices
Dj
in(5)
andSk
in(7)
are generally not diagonal.Theironlynonzero
blocks, however,
are diagonalblockmatrices.We
propose tolabelthemasquasi-diagonalmatrices. Thematrices
Dj,
j 1,..., k-1arequasi-diagonal,their only nonzero blocks being identity matrices. The matrix
Sk
is quasi-diagonalandits nonzeroblocksarediagonalmatriceswith positivediagonal elements. Observe
thatwealwaystakethelastfactorinevery factorizationastheinverseofanonsingular matrix, which is only a matter ofconvention
(another
convention would result in a modified definition of the matricesZi). As
for the name of a certainGSVD,
we propose to adopt the followingconvention(see
also[9]).
DEFINITION 4.2
(the
nomenclatureforGSVDs).
If k 1 in Theorem4.1,
thenthe correspondingfactorizationofthe matrix
A1
willbe called the(ordinary)
singular value decomposition. If for a matrix pairAi, Ai+l,
1<_
_<
k- 1 in Theorem 4.1,we have
Zi
Xi, then the factorization ofthe pair is said to be ofP
type.If,
on the otherhand,
for a matrix pairAi, Ai+I,
1<_
_<
k- 1 in Theorem 4.1, we haveZi
X-*,
then the factorizationofthe pair is said to be ofQ
type. The nameofaGSVD
ofthe matricesAi,
1, 2,... k>
1 as in Theorem 4.1, is then obtained by simplyenumerating the different factorizationtypes.Let
usgive some examples.Example. Considertwo matrices
A
(no
n)
andA2
(n
n2).
Then,
we havetwo possible
GSVDs"
A1
P
typeQ
typeU1D1X
UID1X
X S:
V
X*
S: V;
The
P-type
factorization is called thePSVD
(see
[8]
and referencestherein),
while theQ-type
factorizationis called theQSVD.
Example.
Let
uswrite aPQQP-SVD
for five matrices"A1
U1DIX
1,
A2
XID2Xf
1,
A3
Xf*
D3X
1,
A4
Xf D4X
1,
A
X4SV.
We
also introducea notation using powers thatsymbolize acertainrepetitionofaletter orofasequence of letters:
p3Q2-SVD
PPPQQ-SVD,
(PQ)2Q3(PPQ)2-SVD
PQPQQQQPPQPPQ-SVD.
Despite thefact thatthereare 2k-1 different sequencesof letters
P
andQ
atlevelk
>
1, not all ofthesesequences correspondto differentGSVDs.
The reasonfor this isthat,
for instance, theQP-SVD
of(A
1,
A
2,
A
3)
canbe obtainedfrom thePQ-SVD
of((A3)
*,
(A2)
*,
(A1)*).
Similarly, theP2(Qp)3-SVD
of(A1,...
,A
9)
isessentiallythe same as the(PQ)3p2-SVD
of((A9)*,..., (A1)*).
The number ofdifferentfactoriza-(2k-1
2k/2
(2k-1
(ktions for k matrices is, in
fact,
/ for k even and /:2-1)/2)
for k odd.A
possible way to visualize Theorem 4.1 is to builda tree with all differentfac-torizationsfor
1,
2, 3, etc matrices asfollows:O
P
Q
p2
pQ
Q2
p3
pQ
pQp
pQ
QpQ
Q3
5. Generalized
URVDs.
In
thissection, wederiveageneralization for several matrices, oftheURVD
ofone matrix.We
proceed inseveral stages. First, we show how k matrices can be reduced to blocktria.ngular
matrices using unitarytransfor-mations only.
Next,
weshow how the block triangular factors can betriangularized furthertotriangular factors.THEOREM 5.1 Given k complex matrices
AI
(no
hi),
A2
(hi
n2),
Ak
(nk-1
nk),
there always exist unitarymatricesQo,
Q1,..., Qk
such thatwhere
Ti
is a block lower triangular orblock upper triangularmatrix(both
cases arealways
possible)
with thefollowing structures:--Lower
block triangular(denoted
by a superscriptl)
(9)
T/t
2 i--1r+l
ri ri r ri ri_Ti,1
0 0 0 0 2 ri_*
Ti,2
0 0 0r_
T,
0--Upper
block triangular(denoted
by asuperscriptu)"
(10)
T/u
2 i-1r+l
ri ri r ri ri-1*
0 2Ti,2
*
*
0 ri_ 0 0T,
0 ri_where
Ti,j
j1,...,
arefull
columnrankmatricesand each represents anonzeroblock. The block dimensions coincide with those
of
Theorem4.1.In
particular,r
no,r
+1nullity(Ai)
ni ri, andE
jrank(A)
r r j--1 ri_ hi-1. j=lOur
proof ofTheorem 5.1 is inductive:We
obtain the required factorization ofAi
fromthat ofAi-1.
Proof.
The inductionis initialized for 1 as follows. First, take thecasewhereT1
is to be lower block triangular.Use
a unitary column compression matrixQ1
toreduce thematrix
A
towhere and
T=AIQ=r
(TI,1
0),
r
rank(T1,1)
rank(A),
rl2nullity(A1)
n
rl,r
o.
The casewhere
T1
is requiredto beupper block triangularis similar:T*
AIQ1
=r
(T1,1
0).
Observethatwe havetaken
Qo
Ino.
Now,
we canstartourinduction.Assume
that wehave the requiredfactorizationfor the first i-1 matrices
=QoAQ,
where the matrices
Tj,j
1,...
,i- 1 have the block structure as in Theorem 5.1.We
now want to find a unitary matrix Qi such thatTi
Q_IAQi
is either lower or upper block triangular. First, consider the case whereTi
is to be lower block triangular. The matrixQ-IA
canbe partitioned accordingtothe dimensionsof the block columns ofT_I
as(11)
ni ri_ 2 r,
Q_
n
ri_It
isalwayspossible to constructaunitarymatrixQ
to compress the columns of eachofthe block rowsto the left as
(12)
1-1 /and where the subblocks
Ti,j
are of full columnrank,
denoted byr,
r
+1nullity(A).
Hereto,
we first compress the first block row of(11)
to the left with unitarycolumntransformationsappliedtothe fullmatrix. Thenweproceedwith thesecond blockrowin thedeflatedmatrix(i.e.,
withoutmodifyingthe previousblockcolumn).
By
repeating this procedure times, wefind therequired form(12).
Obviously,
1_<
l-l,
i.(13)
r ri_l,The construction of
Ti
when it is required to be upper block triangular is similar.Construct
a unitary matrixQ
that compresses the columns of the block rows ofQ-IAi
tothe right. The onlydifference isthatwe nowstartfrom thebottomto find that(14)
r+l
2 ri /’i ri 0Til
*
ri-1 2 0 0T2
*
T
Q_IAiQi
ri-1 0 0Ti,i
l’i-We
can nowapplyanadditional(block)
column permutation to theright ofthematrixT/
so asto findthe matrix of(10).
This completes theproof.We
now demonstrate that the matricesT,j
can always be further reduced tointhe casewhen
Ti
is lower block triangular.Here,
R,j
is alower triangularmatrixSimilarly, we can always reduce
Ti,j
tointhecasewhere
Ti
isupper blocktriangular.Here
R
.
is anuppertriangularmatrix 3In
order todemonstrate this, weneed the following result.LEMMA
5.2.Let
PI,...,
Pk
bek givencomplexmatriceswherePi
hasdimensions pi-1 pi, pi-1>_
pi andrank(Pi)
pi. Then there always exist unitary matricesQ0, Q1,...,Qk
such thati--1
where
Ri
is eitherof
theform
(15)
Pi
Ri-Pi-l-pi
)pi
(
0R
with
R
a lowertriangular matrix, or(16)
R-
pip-pi Pi
with
R
upper triangular.For
every1,...,k,
both choices,(15)
and(16),
arealwayspossible.
Proof.
Again, the proofis by induction, but now for decreasingindex i.For
theinitialization, start with k and obtain a QR-decomposition of
Pk
with either an upperor a lower triangular factorasrequired. This defines the unitary matrixQk-.
We
takeQkIpk. Hence,
wefind--Lower
triangular:(0)
--Upper
triangular:Pk
lk-lRk
Qk-1
We
cannowstartthe.
inductionfor k 1, k2,...,
1.Therefore,
assumethat wehave the requiredfactorizations for the matricesPk, Pk-1,..-,
Pi+"
R
k(*k_ Pk
kR
k-1l*k_
2Pk-
l
k-1,Then,
ifRi
istobe lower triangular, obtain aQR-decomposition ofthe product PiQias
(o)
PiQi Qi-lRi Qi-1
R
sothat
R
Q-IPQ"
If
Ri
is required tobe uppertriangular, obtain a QR-decompositionasPiQi Qi-IRi Qi-1
sothat
R
Q-I
PQi.Thiscompletes the construction.
We
now repeatedly applyLemma
5.2 on the full rank blocks inthe matricesTi
in(9)
and(10).
First, weapplyLemma
5.2 to the sequence ofk subblocksNext,
weapplyit tothe sequence of the k 1 subblocksIn
general,we applyLemma
5.2 ktimes to the k sequences of subblocksTj,j,Tj+I,j,...,Tk,j
for j1,...,k.
In
applyingLemma
5.2tothe jth of these sequences,we canfindasequenceofunitary matricesQJ] QJ]
[J]Udk-j+l and matricesRi,j
such thatTi,j
o[J]
i-jRi,j’i-j+l
i=j,...,k, whereor
We
nowdefine the unitarymatrices(i
for 0,...,k,
which areblock diagonalwith blockswith
Qk+l]
I.
Next
wedefineQi_
TiOi,
i=O,...,k.Then,
it canbe verifiedthat
for the lower triangular case we obtain2
r+l
r r r ri_l ,1 0 0 0 2Ri,2
0 0 I’i_ ri_l*
*
Ri,i
0and for the upper triangularcase wefind that
(18)
i--1r+l
r r r ri_l ,1*
*
0 9.R,9.
0 ri_ 0 0R,
0 ri_Ifwe nowcombine
(9)-(17)
and(10)-(18),
weobtaina combined factorizationofthe formHence,
wehave proved the followingtheorem.THEOREM
5.3(generalized
URVDs).
Givenk complexmatricesA1
(no
xnl),
A2
(n,
xn9),
Ak
(nk-
xnk),
there always exist unitary matricesQo,Q,...,Qk
such that
where is a lower triangular or upper triangular matrix
(both
cases are alwayspossible)
with the following structures:--Lower
triangular(denoted
by a superscriptl)"
where
2
r+l
r r r
11
Ri,1
0 0 02
,li
ri-1*
Ri,2
0 0and
R,j
is a square nonsingular lower triangularmatrix.--Upper
triangular(denoted
by a superscriptu)"
i--1
ri+l
r r r ri_l ,1*
*
0 2Ri,2
0 ri_ 0 0Ri,i
0 ri_ whereand
R
is a square nonsingular upper triangular matrix. The block dimensions co-incide with thoseof
Theorem 4.1.As
for the nomenclature of these generalizedURVDs,
we propose the followingdefinition.
DEFINITION 5.4
(nomenclature
for generalizedURV).
The nameofageneralizedURVD
of kmatricesofcompatible dimensionsisgeneratedbyenumerating the lettersL
(for lower)
andU
(for
upper),
according to the lower or upper triangularity ofthe matricesT,
1,...,
k inthe decompositionofTheorem 5.3.For
k matrices, thereare 2k different sequences with two letters.For
instance, for k 3, there are eight generalizedURVD
(LLL, LLU,
LUL, LUU, ULL,
ULU,
UUL,
VVV).
Remarks. The decompositions in Theorems 5.1 and 5.3 both use column and
row compressions of a matrix as a cornerstone for the rank determination of the individualblocks.
As
already pointed out in2,
the rank determination can bedone via anordinarySVD
(OSVD),
buta moreeconomicalmethodusestheQRD
asinitialstep, since typically the matrices involved here have many more columns than rows or vice versa.
A
further alternative would be to replace theOSVD
ofthe triangularmatrix resulting from the initial
QRD
by a rank-revealingQRD.
Since the time ofthe initial paper drawing attention to this
[5],
much progress has been made in thisarea, and we only want to stress here that such alternatives can only benefit our decomposition.
The overall complexity of this
GQRD
is easilyseen to be comparableto that of performing twoQRDs
ofeach matrixA
involved.For
eachA
we indeed applytheleft transformation
Q-I
derivedfrom theprevious matrix andthenapplya "special"compression
Q
of the resulting matrix while respecting its block structure. Bothsteps have a complexity comparable to a
QRD
ofa matrix of the same dimensions.For
parallel machines we can check that the "block" algorithms[18]
for one-sidedorthogonal transformations such as the
QRD
can also be applied to the present de-composition, and thattheywillyield satisfactory speedups. The mainreasonfor this is that the two-sided orthogonal transforms applied to eachAi
are done separately, and hencethey canessentially be considered one-sidedfor parallellization purposes.6.
On
the structure of theGSVD
and theGQRD.
In
this section, we first point out how for eachGSVD
there are twogeneralizedURVDs,
and weclarifythe correspondence between the two types of generalized decompositions.
Next,
we J in Theorems 4.1 and 5.1 give asummary ofexpressions for the block dimensions rin terms of the ranks ofthe matrices
A
1,...,Ak
and concatenations and productsthereof. These expressions werederived in
[10].
Recall the nomenclature for the generalized
URVDs
(Definition
5.4)
and theGSVDs
(Definition
4.2).
The relationshipbetweenthese two definitionsis asfollows.A
pairofidenticalletters,
i.e.,L-L
orU-U,
thatoccursinthe factorization ofAi,
Ai+l
corresponds to a
P-type
factorizationofthe pair.A
pair ofalternatingletters,
i.e.,L-U
orU-L,
thatoccursin the factorizationofA,
Ai+l
corresponds to aQ-type
fac-torization of the pair.
As
an example, for aPQP-SVD
of four matrices, there are twopossible corresponding generalizedURVDs,
namelyan LLUL-decompositionandaUULU-decomposition.
As
withtheGSVD,
we canalso introduce theconventionto use powers of(a
sequenceof)
letters.For
instance, for ap3Q2-SVD,
there are twoGURVs,
namely, anLnUL-URV
and aU4LU-URV.
j 4
Let
usfirst considertheWe
nowderive expressions forthe block dimensionsr.
case ofa
GSVD
that consists only ofP-type
factorizations.Denote
the rank oftheproduct ofthe matrices
A,A+I,...,
Aj
with_<
j byri(i+l)...(j-1)j
rank(AA+l
Aj_IAj).
THEOREM
6.1(on
the structure ofPk-I-SVD, Lk-URV,
andUk-URV).
Con-sider any
of
thefactorizations
abovefor
the matricesA1,
A2,...,
Ak.
Then,
the block dimensionsrji
that appearin Theorems4.1,
5.1, and5.3 are given by:(19)
(20)
--r( rj 1)(2)...(j), rj ri(i+l)...(j r(i-1)(i)...(j), with rj riif
j.Next,
consider the case of aGSVD
that only consists ofQ-type
factorizations.Denote
the rank ofthe block bidiagonal matrix(21)
A
0 0 0 0 0Ai*+l A+e
0 0 0 0 0Ai*+3
Ai+4
0 0 0 0 0 0A_
3Aj_2
0 0 0A_
Aj
(by
rili+ll...lj_llj).
THEOREM
6.2(on
the structure ofQk-I-SVD,
(LU)k/2-URV
(k even),
(UL)
k/2-URV
(k even),
(UL)(k-1)/2U-URV
(k odd),
and(LU)(k-1)/2L-URV
(k odd)).
Con-sider any
of
the abovefactorizations
for
the matricesA1,
Ae,..., Ak.
Then,
Ifj-iiseven,
2
it}+2
r}+4
j--2 jril...Ij ril...Ij-1 -t-
(r
+
rj+...
+
rj)
-t- -t- -t-+
rj -t-rj;Ifj-i is
odd,
4 Recall that the subscript refersto the ith matrix, while the superscript j refers to the jth
For
thegeneral case, we needamixture of the twoprecedingnotations for block bidiagonal matrices, the blocks ofwhich canbe products of matrices, such as(A
A.
..A3_1
0 0(A3
Ai4-1)Ai4
A-I
00
A,
Ajwhere 1
_< io
<
il
<
i2
<
i3
<
<
it
_<
j_<
k. Theirrankis denoted byFor
instance, the rank of thematrixA2A3 0 0
A?
A5A6A7
00
(A8A9)*
AlO
is represented by r(2)(3)141(5)(6)(7)1(8)(9)l(10).
THEOREM
6.3(on
the structure of aGSVD
and aGURV).
r(io)(io+l)...(i-l)li...(i.-1)l...li...j can be derived asfollows:
i=1,2,.,
l+l"
1. Calculate the following+
1 integersBy,
The rank 2
r}O
sj rj-[-rj
2ro+l
rO+2
rl
sj+
+...+
2. Depending on even orodd there are two cases:
--I
even:t
odd:Observe that Theorems 6.1 and 6.2 are special cases of Theorem 6.3. While interms ofdifferences Theorem 6.1 provides adirect expression of the dimensions rj
of ranks of products, Theorems 6.2 and 6.3 do soonly implicitly. This is illustrated inthe following examples.
Example.
Let
us determinethe blockdimensions of the quasi-diagonalmatrix$4 in aQPP-SVD
ofthe matricesA1, A2,
A3,
A4
(which
arealso the block dimensions ofan
LUUU-
or aULLL-decomposition). From
Theorem 6.2 wefindthatand
r(1)1(2)(3)(4 rl
+
s42,
sothatr42
r11(2)(3)(4 rl.Finally, since r4
r
+
r
+
r43+
r44,
wefindthatr
rl+
r(2)(3)(4) r11(2)(3)(4).Observe that this last relation can be interpreted geometrically as the dimension of
theintersectionbetweenthe row spacesof
A1
andA2A3A4:
r41
dimspanrow(A1)
+
dimspanrow(A2A3A4
dimspanro
A2A3A3
Example. Consider the determination of
r, r,
r53,
r,
r
in aPQ3-SVD
of five matricesA1,A2, A3, A4,
A5
withTheorem6.3,
which coincides withthe structure ofa
UULUL-URV
or anLLULU-URV
(see
Table1).
TABLE r415 r2J314[5 r(1)(2)J3J4J5
8
r+r +r
+r
48
r55
8 =-r5+
r5+
r5 85 r5 1+}_2 8 r r 82 r3 3__r4 858
r55
3s
--r+
r5These relationscanbe used to set up aset ofequationsfortheunknowns
r,
r,
r53,
r5,
r55,
usingTheorem 6.3 as 1 1 1 1 1r
r5 0 0 0 0 1r52
r415
r4 1 1 1 0 1r53
r3j4j5 r3]4 0 0 1 0 1r54
r2131415
r2J314
0 1 1 0 1r55
r(1)(2)J3J4J5 r(1)(2)J3J4 the solutionofwhich is1010 BARTDE MOOR AND PAULVANDOOREN
r
r(1)(2)131415 r(1)(2)1314r2131415
--
r21314,r
r2131415
r21314
r415
+
ra,r54
r5r31415
-
r314,r55
r415
r4.7.
A
further block diagonalization of theGQRD. In
thissection, wenotethat afurther block diagonalization ofa
GQRD
can be interpretedas apreliminary step towards the correspondingGSVD. We
proceedin two stages. First, we observe that each upper or lower triangular matrix in the generalizedURVD
of Theorem 5.3 can be block diagonalized.Next,
we show how these block diagonalizations canbe propagated backward through the
GQRD.
The first step is the factorization of the upper and lower triangular matricesTi
of Theorem 5.3 into an upper or lower triangularmatrixandablock diagonalmatrix.For
lowertriangularmatricesi
/,
we can obtain afactorization ofthe formwhere
2 i--1
ri_l ri_l ri_ ri_l
I
0 0 0 ri_ 2 ri_ *I
0 0 ?i-1 * * *I
2r+l
/’i /’i ri_l ,1 0 0 0 2Ri
2 0 0 0 0Ri,i
0i--Since the diagonal blocks
Ri,j
are of full columnrank,
such a factorization isal-ways possible.
In
a similar way, for upper iangular matricesi
/u,
we find a factorizationof the formwith
Ui
anuppertriangular blockmatrixwithidentitymatricesonthe blockdiagonal:2 i--1
ri--1 i--1 ri--1 i--1
I
ri_ 2 0I
ri_ 0 0 0I
ri_ 2r+l
ri ?i l’i ri-1 ,1 0 0 0Ri,2
0 0r_
0 0Ri,i
0Now
suppose that wehave done thisfor all matricesi,
1,..., kin aGQRD
of Theorem 5.3.We
show how we can propagate a further block diagonalization backward throughtheGQRD,
inaway thatiscompletelyconsistent withthecorre-sponding
GSVD
of Theorem4.1.To
simplify the notation, wesimply replaceTi
byTi
andDi
byD
inthe following.First, assumethat
Tk
is lowerblock triangular.It
thenfollows from theprevious sectionthatwe can factorizeTk
asTk
LD.
Dependingonwhether
T-I
isupper orlower triangular, wehave twocases".T-I
T_
lower triangular.In
thiscase,
theproductTk_IL
k islowertrian-gularas
well,
andwe can obtain asimilardecompositionTk
t-lLa
La_Da_,
where
La_
is againlower triangularandD_
has the same diagonal blocksRi,
asTk-1
Tkl
upper triangular.In
this case, the productT%L*
is uppertriangular, andwe can obtain afactorization
-Uk-lDk
T_I
L;*
where
U-I
is upper triangularandDk-1
hasthe samediagonal blocksR,y
asTkl.
It
is easily verified that whenTk
is upper triangular, similar conclusions can be obtained.In
generM,
letT
be lower triangular and assume thatit is factorized asT
L
D Z.
Assume
thatT_
is lowertriangular. ThenT-I
canbe fctored asT_
L_D_L 1.
IfT_
is uppertriangular, it can be factoredasTiL
U_
D_
1Li,
where
U-I
is uppertriangular. Thecases withT
uppertriangulararesimilar. Table2 summarizesall possibilities.
Ti Lowertriangular
Ti LiDZ
T
UppertriangularT
UiD Z
TABLE 2
Ti- Lowertriangular Ti-1 Uppertriangular Ti-1 Lowertriangular Ti_ Uppertriangular
Ti-1 Li-1 Di-1
U
Ti-i Vi-1Di_U
-1Example.
Let
us applythis result to a sequenceof four matricesA1,A2,A3,A4
with compatibledimensions. Iftherequired sequence isULUU,
thenA1
QoTQ*
Qo(UD1L:)Q
(QoU1)DI(QIL2)*
A:
QIT2Q
QI(L:DU)Q
(QIL2)D2(Q2Ua)*,
A3
Q2TQ
Q2(U3D3ui)Q
(Q2U3)D3(Q3U4)
-1Note
thatU1
Ino.
Thisfollows immediately fromtheblockstructureofU
for 1.Observe that the relationships between the common factors in theleft-hand side of these expressions conform with the requirements for a
QQP-SVD.
Only
the middlefactors
D,
1, 2, 3,
4 are notquasi-diagonal.8. Conclusions.
In
this paper, aconstructive proofwas givenofamultimatrixgeneralization of the concept of rank factorization. The connection ofthis new
de-composition with the analogous
GSVD
was also shown. The blockstructure of both generalizations and the ranks of theindividualdiagonal blocksinboth decompositionswere indeed shown to be identical.
As
is shown in a forthcoming paper, the spacesspanned by certainblock columns of the orthogonaltransformation matrices
Q
are,in
fact,
identicaltothoseoftheGSVD.
The difference lies only inaparticular choiceofbasisvectorsfor thesespaces. Theconsequencesof theseconnections arestillunder
investigation.
We
mentionthe following results here:Updating the above decompositiontoyield the
GSVD
requiresnonorthogonaltransformation. Theseupdatingtransformations canbe chosenblock triangularwith
diagonal block sizescompatiblewiththe indexsets derived in Theorem4.1.
A
modified orthogonal decomposition can be defined where the compoundmatrix is not triangularized but diagonalized. This new factorization is a variant of
the above decomposition where now a special coordinate system is chosen for each
of theindividual orthogonal transformations
Q.
The resultis an orthogonaldecom-position ofthe type ofTheorem 5.3 where now the generalized singular values can be extracted from the diagonal elements of some triangular blocks. The orthogo-hal updating needed to obtain this new decomposition can be done with techniques
described in
[2].
A
geometric interpretation can be givenofthe bases obtainedfromthe trans-formation matricesQ
in Theorem 5.1.As
particular examples of these spaces we retrieve the followingwell-knownconcepts.(a)
For
the caseA
(A-I)
theGQRD
infactreconstructs the nested null spacesofthe matrices
(A
aI)
,
which reveal theJordanstructureofthe matrixA
at theeigenvalue a(see
also theexample in2).
(b)
For
the casesA2
(A-aB)
andA2+1
B
the decomposition reconstructs thenested null spacesof thesequences
[B-.I(A
aB)]
and[(A
aB)B-1]
,
whichrevealthe Kronecker structureof thepencil
AB- A
atthe generalized eigenvaluec
(see
[30]
and[31]).
(c)
For
the casesA
D
andA
C.
A
-.
B,i 1,..., the decomposition reconstructs theinvertibility subspaces ofthe discrete timesystem
Xk+l
Axk
+
Buk,
YkCxk
+
Duk.
These are in fact also the spaces constructed by the structure algorithm of Sil-verman
[29],
andthey play arole inseveralkey problems of geometrical systems theory[34].
Other applications of
GSVDs
have beendescribed in[7], [8], [11], [13],
and[35],
while applications of the generalized QR-decompositions are described in
[25]
and Acknowledgment.Gene
Golub’s hospitality inthesummer of 1989 liesat thegeneratingthe
MATLAB
codefortheGSVD
and theGQRD
basedonthe constructiveproofs of the
GSVD
in[9]
and oftheGQRD
ofTheorem 5.3ofthis paper.REFERENCES
[1] A. BJRCK,Solvinglinearleastsquaresproblems byGram-Schmidt orthogonalization,BIT,7
(1967),pp. 1-21.
[2] A. BOJANCZYK, M. EWERBRING, F. LUK, AND P. VAN DOOREN, An accurateproductSVD
algorithm, in Proc. Second Internat. Sympos. on SVD and Signal Processing, Kingston,
RI, pp.217-228, June1990; SignalProc.,to appear.
[3] T. CHAN, Animproved algorithmforcomputing the singular valuedecomposition,ACMTrans.
Math. Software, 8(1982),pp. 72-83.
[4]
,
Algorithm581: Animproved algorithmforcomputingthe singular value decomposition,ACM Trans. Math. Software,8(1982), pp. 84-88.
[5]
,
RankrevealingQRfactorizations,LinearAlgebra Appl.,88/89 (1987),pp. 67-82. [6] B. DE MOOR AND G. H. GOLUB, Generalized singular value decompositions: A proposalfor astandardizednomenclature,ESAT-SISTA Report 1989-10,Departmentof Electrical
Engineering,Katholieke Universiteit Leuven,Leuven,Belgium, April 1989.
[7]
,
The restricted singular value decomposition: Properties and applications, SIAM J.MatrixAnal.Appl., 12 (1991), pp. 401-425.
[8] B. DE MOOR, On the structure and geometry ofthe product singular value decomposition, LinearAlgebra Appl., 168(1992), pp. 95-136.
[9] B. DE MOOR AND H. ZHA, A tree of generalizations ofthe singular value decomposition,
ESAT-SISTA Report 1990-11, Departmentof Electrical Engineering, Katholieke
Univer-siteitLeuven, Leuven, Belgium, June1990; LinearAlgebra Appl.,to appear.
[10] B. DE MOOR, On the structure ofgeneralized singular value decompositions, ESAT-SISTA Report 1990-12, Department of Electrical Engineering, Katholieke Universiteit Leuven, Leuven,Belgium, June1990.
[11]
,
Generalizations oftheOSVD: Structure, properties, applications, in Proc. SecondIn-ternat. Workshop on SVDandSignal Processing, University of Rhode Island, Kingston, RI, June1990,pp. 209-216; SignalProc.,to appear.
[12]
,
A history ofthe singular value decomposition, ESAT-SISTA Report, Department of Electrical Engineering, Katholieke UniversiteitLeuven, Leuven,Belgium, February1991. [13] K. V.FERNANDOANDS. J. HAMMARLING, Aproductinducedsingular valuedecompositionfortwo matrices and balanced realisation, in Linear Algebrain Signal Systems andControl, B. N.Datta, C. R. Johnson,M. A. Kaashoek, R. Plemmons,andE. Sontag, eds., Society for Industrial andApplied Mathematics,Philadelphia,PA, 1988,pp. 128-140.
[14] P. E. (ILL, W. MURRAY, AND M.
n.
WRIGHT, Practical Optimization, Academic Press,London, 1981.
[15] G H. GOLUBANDW.KAHAN, Calculating the singular values and pseudo-inverseofamatrix,
SIAMJ. Numer. Anal.,2(1965), pp. 205-224.
[16] P A. BUSINGERANDG.H. GOLUB,Algorithm358: Singular value decompositionofacomplex
matrix,Comm.Assoc. Comp. Mach., 12 (1969),pp. 564-565.
[17] G H. GOLUB AND C. REINSCH, Singular value decomposition and least squares solutions,
Numer. Math., 14 (1970),pp. 403-420.
[18] ( H. GOLUBAND C. VAN LOAN, MatrixComputations, Second Edition,The JohnsHopkins
UniversityPress,Baltimore, MD,1989.
[19] G GOLUBANDJ. WILKINSON,Ill-conditioned eigensystems andthecomputationoftheJordan
canonicalform, SIAMRev., 18(1976),pp. 578-619.
[20] S HAMMARLING, The numerical solution ofthe general Gauss-Markov linearmodel, NAG
Tech. ReportTR2/85,NumericalAlgorithms GroupLimited,Oxford, 1985.
[21] M. T. HEATH, A. J. LAUB, C. C. PAIGE, AND R. C. WARD, Computing the singularvalue decomposition of a product of two matrices, SIAM J. Sci. Statist. Comput., 7 (1986), pp. 1147-1159.
[22] R. J. HANSONANDC. L.LAWSON,Extensions andapplicationsofthe Householder algorithms forsolving linear least squaresproblems, Math. Comp.,23 (1969),pp. 787-812.
[23] C.C. PAIGEANDM.A. SAUNDERS, Toward8ageneralized singular value decomposition,SIAM J. Numer.Anal.,18 (1981), pp. 398-405.
[24] C. C. PAIGE, Computing thegeneralized singular value decomposition, SIAM J. Sci. Statist.
Comput., 7(1986),pp. 1126-1146.
CoxandS. Hammarling,eds., OxfordUniversityPress, Oxford,U.K., 1990, pp. 73-91. [26] G. W. STEWART, A methodfor computing the generalized singular value decomposition, in
Proc. MatrixPencils, PiteHavsbad, 1982, B.KagstrSmandA. Ruhe,eds., Lecture Notes inMathematics973, Springer-Verlag, Berlin, NewYork, 1983.
[27]
,
An updating algorithmforsubspace tracking, UMIACS-TR-90-86, CS-TR2494,Com-puterScience Tech.ReportSeries, University ofMaryland, CollegePark,MD,July 1990. [28] J. STOER, Onthe numerical solution ofconstrainedleast-squares problems, SIAM J. Numer.
Anal.,8(1971),pp. 382-411.
[29] L. M. SILVERMAN, Discrete Riccati Equations: AlternativeAlgorithms, AsymptoticProperties andSystem TheoryInterpretations, inControland DynamicalSystems, AcademicPress, New York,1976.
[30] P. VANDOOREN, ThecomputationofKronecker’s canonicalform ofasingular pencil, Linear
AlgebraAppl., 27(1979), pp. 103-140.
[31]
,
Reducing subspaces: Definitions, properties and algorithms, in Proc. MatrixPencils,Lecture NotesinMathematics973, Springer-Verlag, Berlin,NewYork, 1983,pp. 58-73. [32] C. F. VAN LOAN, Generalizingthe singular value decomposition, SIAM J. Numer. Anal., 13
(1976),pp. 76-83.
[33] J. H. WILKINSON, The Algebraic Eigenvalue Problem, The Clarendon Press, Oxford, U.K., 1965.
[34] M. WONHAM, Linear Multivariable Control, a Geometric Approach, Springer-Verlag, New York, 1974.
[35] H. ZHA, The restrictedSVDformatrixtriplets and rank determinationofmatrices, Scientific
Report89-2, Konrad-ZuseZentrumfiirInformationstechnik, Berlin,Germany, 1989;SIAM
J.Matrix Anal. Appl., 12 (1991),pp. 172-194.
[36]