• No results found

On the nonuniqueness of the factorization factors in the product singular value decomposition

N/A
N/A
Protected

Academic year: 2021

Share "On the nonuniqueness of the factorization factors in the product singular value decomposition"

Copied!
13
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

www.elsevier.com/locate/laa

On the nonuniqueness of the factorization factors in the product singular value decomposition

Delin Chu a, , Bart De Moor b,1

a

Department of Mathematics, National University of Singapore, Lower Kent Road, Singapore 119260, Singapore

b

Department of Electrical Engineering (ESAT), Research Group SISTA, Katholieke Universiteit Leuven, Kardinaal Mercierlaan 94, B-3001 Leuven, Belgium

Received 18 December 1998; accepted 4 April 2000 Submitted by V. Mehrmann

Abstract

The product singular value decomposition (PSVD) of two matrices is revisited in this pa- per. The nonuniqueness of the factorization factors in the PSVD is characterized in a way different from that in existing work. © 2000 Elsevier Science Inc. All rights reserved.

AMS classification: 65F15; 65H15

Keywords: Product singular value decomposition; Nonuniqueness; Factorization factors

This work is supported by several institutions:

1. The Flemish Government:

(a) Concerted Research Action GOA-MIPS (Model-based Information Processing Systems), (b) the FWO (Fund for Scientific Research—Flanders) project G.0292.95: Matrix algorithms and

differential geometry for adaptive signal processing, system identification and control,

(c) the FWO project G.0256.97: Numerical Algorithms for Subspace System Identification, Extension to Special Cases,

(d) the FWO Research Communities: ICCoS (Identification and Control of Complex Systems) and Advanced Numerical Methods for Mathematical Modelling.

2. The Belgian State, Prime Minister’s Office—Federal Office for Scientific, Technical and Cultural Affairs: Interuniversity Poles of Attraction Programme (IUAP P4-02 (1997–2001): Modeling, Iden- tification, Simulation and Control of Complex Systems; and IUAP P4-24 (1997–2001): Intelligent Mechatronic Systems (IMechS)).

Corresponding author. Fax: +65-7795452.

E-mail addresses: matchudl@math.nus.edu.sg (D. Chu), Bart.Demoor@esat.kuleuven.ac.be (B. De Moor).

1

Tel.: +32-16-32-1970; fax: +32-16-32-1709.

0024-3795/00/$ - see front matter



2000 Elsevier Science Inc. All rights reserved.

PII: S 0 0 2 4 - 3 7 9 5 ( 0 0 ) 0 0 1 3 6 - 1

(2)

1. Introduction

The product singular value decomposition (PSVD) is a generalization for two matrices of the (ordinary) singular value decomposition (SVD) for one matrix. The explicit formulation of the PSVD was given for the first time by Fernando and Hammarling [7], who called it the SVD. In this paper, unless noted, we always denote

r a = rank(A), r b = rank(B), r ab = rank AB T 

, X −T = (X −1 ) T for any given matrices A, B of appropriate dimensions and nonsingular matrix X.

Let us first state the following:

Theorem 1 (The PSVD Theorem). Given matrices A ∈ R m×n and B ∈ R p×n . Then there exist orthogonal matrices U ∈ R m×m , V ∈ R p×p and a nonsingular matrix X ∈ R n×n such that

U AX =

r ab r a − r ab r b − r ab n + r ab − r a − r b

r ab S 0 0 0

r a − r ab 0 I 0 0

m − r a 0 0 0 0

, (1)

V BX −T =

r ab r a − r ab r b − r ab n + r ab − r a − r b

r ab S 0 0 0

r b − r ab 0 0 I 0

p − r b 0 0 0 0

,

where

S = diag{σ 1 I i

1

, σ 2 I i

2

, . . . , σ k I i

k

}, σ 1 > σ 2 > · · · > σ k > 0, X k j =1

i j = r ab .

Observe that

U AB T  V T =

r ab r b − r ab p − r b

r ab SS T 0 0

r a − r ab 0 0 0

m − r a 0 0 0

, hence, SS T contains the singular values of AB T .

Algorithmic ideas to implement the PSVD in a numerically reliable way can be

found in [2,7]. Applications include the orthogonal Procrustes problem [1], com-

puting balancing transformations for state space systems [7,9], and computing the

Kalman decomposition of a linear system [8]. The PSVD could also be applied in

the computation of approximate intersections between subspaces in the stochastic

realization problem [6], as an alternative to canonical correlation analysis.

(3)

The structure and geomety of the PSVD have been studied in [4]. In particular, the nonuniqueness of the factorization factors in the PSVD has been analyzed in detail.

In this paper, we revisit the PSVD. Our purpose is to characterize the nonuniqueness of the factors in the PSVD in a way different from that in [4].

2. Main result

Before we state our main result, we need a technical lemma.

Lemma 2. Given A ∈ R m×n , B ∈ R p×n . Then there exist orthogonal matrices U a ∈ R m×m , V b ∈ R p×p and Q ab ∈ R n×n such that

U a AQ ab =

r ab r a − r ab r b − r ab n + r ab − r a − r b

r ab A 11 A 12 0 0

r a − r ab 0 A 22 0 0

m − r a 0 0 0 0

, (2)

V b BQ ab =

r ab r a − r ab r b − r ab n + r ab − r a − r b

r ab B 11 0 B 13 B 14

r b − r ab 0 0 B 23 0

p − r b 0 0 0 0

,

where A 11 , A 22 , B 11 , B 23 are nonsingular.

Proof. See Appendix A. 

Based on Lemma 2, we can prove Theorem 1. In fact, we have:

Corollary 3. Given A ∈ R m×n and B ∈ R p×n . Let orthogonal matrices U a , V b and Q ab be defined as in Lemma 2. Assume that the SVD of A 11 B 11 T is

U 11 A 11 B 11 T V 11 T = S 2 , (3)

where U 11 , V 11 are orthogonal and S is defined as in Theorem 1. Define U =

 U 11

I m−r

ab



U a , V =

 V 11

I p−r

ab

 V b ,

X = Q ab

 

 

 

A −1 11 U 11 T S −A −1 11 A 12 A −1 22 0 0

0 A −1 22 0 0

B 13 T V 11 T S −1 0 B 23 T 0 B 14 T V 11 T S −1 0 0 I

 

 

 

.

(4)

Then U AX and V BX −T are in the form (1).

Corollary 3 provides an alternative and very simple way to characterize the PSVD, which is different from the derivation in [4,5,7]. We are now in the position to present our main result.

Theorem 4 (Main Result). Given A ∈ R m×n and B ∈ R p×n . Let orthogonal matri- ces U a , V b , Q ab , U 11 and V 11 be defined as in Lemma 2 and Corollary 3. Assume that U ∈ R m×m and V ∈ R p×p are orthogonal and X ∈ R n×n is nonsingular. Then U AX and V BX −T are in the form (1) if and only if

U =

r ab r a − r ab m − r a r ab P 11 U 11

r a − r ab P 22

m − r a P 33

U a ,

V =

r ab r b − r ab p − r b

r ab P 11 V 11

r b − r ab W 22

p − r b W 33

V b , (4)

X =Q ab

 

 

 

A −1 11 U 11 T SP 11 T −A −1 11 A 12 A −1 22 P 22 T 0 0

0 A −1 22 P 22 T 0 0

B 13 T V 11 T S −1 P 11 T X 32 B 23 T W 22 T X 34

B 14 T V 11 T S −1 P 11 T X 42 0 X 44

 

 

 

,

where P 11 , P 22 , P 33 , W 22 and W 33 are arbitrary orthogonal matrices, P 11 the block-diagonal:

P 11 =

 

 

i 1 i 2 · · · i k

i 1 P 11 (1) i 2 P 11 (2)

.. . . . .

i k P 11 (k)

 

  , (5)

X 32 , X 34 and X 42 the real and arbitrary, and X 44 is arbitrary real nonsingular matrix.

Obviously, Theorem 4 gives a complete characterization of the nonuniqueness

property of the factorization factors in the PSVD.

(5)

3. The proof of Theorem 4

In order to prove Theorem 4, we need two preliminary lemmas, in which the first one is well known and the second one is a direct consequence of QR and QL factorizations.

Lemma 5. The SVD of AB T is given by

 U 11

I m−r

ab

 U a

 AB T

 V 11

I p−r

ab

 V b

 T

=

 S 2 0

0 0

 . Furthermore, if orthogonal matrices U and V satisfy

U AB T V T =

 S 2 0

0 0

 , then

U =

r ab r a − r ab m − r a r ab P 11 U 11

r a − r ab P 22 P 23

m − r a P 32 P 33

U a ,

(6)

V =

r ab r b − r ab p − r b r ab P 11 V 11

r b − r ab W 22 W 23

p − r b W 32 W 33

V b ,

where P 11 ,

 P 22 P 23

P 32 P 33

 and

 W 22 W 23

W 32 W 33



are orthogonal, and P 11 is of the block-diagonal form (5).

Proof. The proof is trivial. 

Lemma 6. Let X ∈ R n×n be nonsingular. Then there exists an orthogonal matrix Q such that

Q T X =

 

r ab r a − r ab r b − r ab n + r ab − r a − r b

r ab L 11 L 12 0 0

r a − r ab 0 L 22 0 0

r b − r ab L 31 L 32 L 33 L 34

n + r ab − r a − r b L 41 L 42 L 43 L 44

 

,

(7)

where

(6)

L 11 , L 22 and

 L 33 L 34

L 43 L 44



are nonsingular.

Proof. We partition X to be

X =  r ab r a − r ab r b − r ab n + r ab − r a − r b

n X 1 X 2 X 3 X 4

 .

Let ˜ Q be such that

Q ˜ T 

X 3 X 4

 =

 

r b − r ab n + r ab − r a − r b

r ab 0 0

r a − r ab 0 0

r b − r ab L 33 L 34

n + r ab − r a − r b L 43 L 44

 

, and denote

Q ˜ T 

X 1 X 2

 =

 

r ab r a − r ab

r ab ˜L 11 ˜L 12

r a − r ab ˜L 21 ˜L 22 r b − r ab L 31 L 32

n + r ab − r a − r b L 41 L 42

 

.

Now, there exists an orthogonal matrix ˆ Q satisfying

Q ˆ T  ˜ L 11 ˜L 12

˜L 21 ˜L 22



=

 r ab r a − r ab

r ab L 11 L 12

r a − r ab 0 L 22

 . Set

Q = ˜ Q

 Q ˆ I

 .

Then Q is such that (7) holds.  Now we prove Theorem 4.

Proof of Theorem 4. First we prove the sufficiency, then prove the necessity.

Sufficiency: Assume that U ∈ R m×m and V ∈ R p×p are orthogonal, X ∈ R n×n is nonsingular, and U AX and V BX −T are in the form (1). Then U AB T V T is the SVD of AB T , and hence, by Lemma 5, U and V are of the form (6). Note that

 0 P 32 P 33

 U a AX = 0, 

0 W 32 W 33

 V b BX −T = 0.

This implies that

 0 P 32 P 33

 U a AQ ab = 0, 

0 W 32 W 33

 V b BQ ab = 0.

(7)

So,

P 32 = 0, W 32 = 0. (8)

Consider that

 P 22 P 23

P 32 P 33

 and

 W 22 W 23

W 32 W 33



are orthogonal. Hence, we also have

P 23 = 0, W 23 = 0. (9)

By (8) and (9), we have that U and V are in the form (7).

According to Lemma 6, there exists an orthogonal matrix Q such that Q T X is of the form (7). Set

Q = 

Q 1 Q 2 Q 3 Q 4

 .

Then A 

Q 3 Q 4

 = 0. (10)

But, let us partition Q ab in Lemma 2 into

Q ab =  r ab r a − r ab r b − r ab n + r ab − r a − r b Q ab1 Q ab2 Q ab3 Q ab4 

. Then we know

A 

Q ab3 Q ab4 

= 0.

Hence,

 Q 3 Q 4

 = 

Q ab3 Q ab4  "

Q ˜ 33 Q ˜ 34

Q ˜ 43 Q ˜ 44

# ,

 Q 1 Q 2

 = 

Q ab1 Q ab2  "

Q ˜ 11 Q ˜ 12

Q ˜ 21 Q ˜ 22

#

with "

Q ˜ 11 Q ˜ 12

Q ˜ 21 Q ˜ 22

# and

"

Q ˜ 33 Q ˜ 34

Q ˜ 43 Q ˜ 44

#

orthogonal. Since we also have BQ 2 = 0, BQ ab2 = 0, so,

Q ˜ 21 = 0, Q ˜ 12 = 0.

Hence, we have

(8)

Q = Q ab ×

 

r ab r a − r ab r b − r ab n + r ab − r a − r b

r ab Q ˜ 11

r a − r ab Q ˜ 22

r b − r ab Q ˜ 33 Q ˜ 34

n + r ab − r a − r b Q ˜ 43 Q ˜ 44

 

with

Q ˜ 11 , Q ˜ 22 and

"

Q ˜ 33 Q ˜ 34

Q ˜ 43 Q ˜ 44

#

orthogonal. We can write X to be

X =Q ab

 

Q ˜ 11

Q ˜ 22

Q ˜ 33 Q ˜ 34

Q ˜ 43 Q ˜ 44

 

 

L 11 L 12 0 0

0 L 22 0 0

L 31 L 32 L 33 L 34

L 41 L 42 L 43 L 44

 

=Q ab ×

 

r ab r a − r ab r b − r ab n + r ab − r a − r b

r ab X 11 X 12 0 0

r a − r ab 0 X 22 0 0

r b − r ab X 31 X 32 X 33 X 34

n + r ab − r a − r b X 41 X 42 X 43 X 44

 

.

Obviously,

X 11 , X 22 and

 X 33 X 34

X 43 X 44



are nonsingular. Now we have

P 11 U 11

P 22

P 33

A 11 A 12 0 0

0 A 22 0 0

0 0 0 0

 

X 11 X 12 0 0

0 X 22 0 0

X 31 X 32 X 33 X 34

X 41 X 42 X 43 X 44

 

=

S 0 0 0

0 I 0 0

0 0 0 0

 (11)

and

P 11 V 11

W 22

W 33

B 11 0 B 13 B 14

0 0 B 23 0

0 0 0 0

(9)

=

S 0 0 0

0 0 I 0

0 0 0 0

 

X 11 X 12 0 0

0 X 22 0 0

X 31 X 32 X 33 X 34

X 41 X 42 X 43 X 44

 

T

. (12)

A simple calculation yields that (11) and (12) hold if and only if P 11 U 11 A 11 X 11 = S,

P 22 A 22 X 22 = I, (13)

P 11 U 11 (A 11 X 12 + A 12 X 22 ) = 0 and

P 11 V 11 B 11 = SX T 11 , P 11 V 11 B 13 = SX T 31 ,

W 22 B 23 = X T 33 , (14)

P 11 V 11 B 14 = SX T 41 , X T 43 = 0.

Equivalently, (11) and (12) hold if and only if X 11 = A −1 11 U 11 T SP 11 T ,

X 22 = A −1 22 P 22 T ,

X 12 = −A −1 11 A 12 A −1 22 P 22 T ,

X 33 = B 23 T W 22 T , (15)

X 31 = B 13 T V 11 T P 11 T S −1 = B 13 T V 11 T S −1 P 11 T , X 41 = B 14 T V 11 T P 11 T S −1 = B 14 T V 11 T S −1 P 11 T , X 43 = 0.

Moreover, (15) also implies that X 32 , X 42 , X 34 and X 44 are arbitrarily, and X 44 is nonsingular because

 X 33 X 34

X 43 X 44



is nonsingular and X 43 = 0. Therefore, X is also in the form (4). Up to now, we have completed the proof of sufficiency in Theorem 4.

Necessity: Let orthogonal matrices U ∈ R m×m , V ∈ R p×p and nonsingular ma- trix X ∈ R n×n are in the form (4). Then a simple calculation gives that

U AX =

P 11 SP 11 T 0 0 0 0 P 22 P 22 T 0 0

0 0 0 0

 =

S 0 0 0

0 I 0 0

0 0 0 0

 (16)

(10)

and

S 0 0 0

0 0 I 0

0 0 0 0

 

 

A −1 11 U 11 T SP 11 T −A −1 11 A 12 A −1 22 P 22 T 0 0

0 A −1 22 P 22 T 0 0

B 13 T V 11 T S −1 P 11 T X 32 B 23 T W 22 T X 34

B 14 T V 11 T S −1 P 11 T X 42 0 X 44

 

 

T

=

SP 11 SU 11 A −T 11 0 SP 11 S −1 V 11 B 13 SP 11 S −1 V 11 B 14

0 0 W 22 B 23 0

0 0 0 0

=

P 11 V 11 B 11 0 P 11 V 11 B 13 P 11 V 11 B 14

0 0 W 22 B 23 0

0 0 0 0

=

P 11 V 11

W 22

W 33

B 11 0 B 13 B 14

0 0 B 23 0

0 0 0 0

 , (17)

equivalently, we have

V BX −T =

S 0 0 0

0 0 I 0

0 0 0 0

 . (18)

In (17), we have used the following equalities:

SP 11 = P 11 S, S T = S, V 11 B 11 = S 2 U 11 A −T 11 . Hence, the “necessity” follows directly from (16) and (18). 

4. Conclusion

In this paper, the nonuniqueness of the factorization factors in the PSVD has been characterized in a way different from that in [4].

Appendix A

Before we prove Lemma 2, we need to recall the QR factorization with column pivoting and URV decomposition [1], which will be the building blocks of our con- structive proof of Lemma 2.

It is well known that any matrix A ∈ R m×n can be factorized as

(11)

U A =

 R 1 R 2

0 0



P, (A.1)

where U and P are orthogonal matrix and permutation matrix, respectively, R 1 is nonsingular and upper triangular. The factorization (A.1) is called the QR factoriza- tion of A with column pivoting.

If we continue to squeeze [ R

1

R

2

] into upper triangular form by applying a se- quence of Householder transformations, then we have the following URV decompo- sition of A, i.e., we get an orthogonal matrix V such that

U AV =

 R 0

0 0



(A.2) with R nonsingular and upper triangular.

Now we are ready to present a constructive proof for Lemma 2.

Proof of Lemma 2. We prove Lemma 2 constructively by following six steps:

Step 1: Compute the U RV decomposition of A:

ˆU 1 A ˆ Q 1 =

" r a n − r a r a A (1) 1 0 m − r a 0 0

# ,

where A (1) 1 is nonsingular.

Step 2: Compute the QR factorization of B ˆ Q 1 with column pivoting:

ˆV 1 (B ˆ Q 1 ) =

" r a n − r a

r b B 1 (1) B 3 (1) p − r b 0 0

# , where 

B

1(1)

B

3(1)

 is of full row rank. Note that r ab = rank AB T 

= rank 

A (1) 1 B 1 (1)

 = rank  B 1 (1)

 . Step 3: Compute the QR factorization of (B 1 (1) ) T :

B 1 (1) Q ˆ 2 = h r ab r a − r ab

r b B 1 (2) 0 i with B 1 (2) of full column rank. Set

A (1) 1 Q ˆ 2 = h r ab r a − r ab

r a A (2) 1 A (2) 2 i

.

Step 4: Compute the QR factorizations of A (2) 1 and B 1 (2) with column pivoting:

ˆU 3

h

A (2) 1 A (2) 2 i =

 r ab r a − r ab r ab A 11 A 12

r a − r ab 0 A 22



,

(12)

ˆV 3

h

B 1 (2) B 3 (1) i =

" r ab n − r a r ab B 11 B 13 (3) r b − r ab 0 B 23 (3)

# ,

where A 11 , A 22 and B 11 are nonsingular. We also have that B 23 (3) is of full row rank.

Step 5: Compute the QR factorization of (B 23 (3) ) T with pivoting:

"

B 13 (3) B 23 (3)

# Q ˆ 4 =

 r b − r ab n + r ab − r a − r b

r ab B 13 B 14

r b − r ab B 23 0

 ,

where B 23 is nonsingular.

Step 6: Set U a =  ˆ U 3

I m−r

a

 ˆU 1 ,

V b =  ˆ V 3

I p−r

b

 ˆV 1 ,

Q ab = ˆ Q 1

 Q ˆ 2

I n−r

a

  I r

a

Q ˆ 4

 . Now we have

U a AQ ab =

r ab r a − r ab r b − r ab n + r ab − r a − r b

r ab A 11 A 12 0 0

r a − r ab 0 A 22 0 0

m − r a 0 0 0 0

,

V b BQ ab =

r ab r a − r ab r b − r ab n + r ab − r a − r b

r ab B 11 0 B 13 B 14

r b − r ab 0 0 B 23 0

p − r b 0 0 0 0

with A 11 , A 22 , B 11 and B 23 nonsingular. Therefore, Lemma 2 follows.  In general, the size of A 11 (i.e., the size of B 11 ) is much smaller than those of A and B. Moreover, the condensed form (2) can be computed via numerically sta- ble ways. Hence, similar to [3], the condensed form (2) can be considered to be an efficient preprocessing algorithm for computing the PSVD of matrix pair (A, B).

This preprocessing algorithm will reduce the complexity of the Kogbetliantz-type

algorithm in [2]. Therefore, the PSVD of (A, B) can be computed in the following

two phases:

(13)

• Reduce matrix pair (A, B) to the condensed form (2).

• Compute the PSVD of matrix pair (A 11 , B 11 ) using the Kogbetliantz-type algo- rithm in [2].

References

[1] G.H. Golub, C.F. Van Loan, Matrix Computations, third ed., Johns Hopkins University Press, Balti- more, MD, 1996.

[2] M.T. Heat, A.J. Laub, C.C. Paige, R.C. Ward, Computing the singular value decomposition of a product of two matrices, SIAM J. Sci. Statist. Comput. 7 (1986) 1147–1159.

[3] Z. Bai, H. Zha, A new preprocessing algorithm for the computation of the generalized singular value decomposition, product of two matrices, SIAM J. Sci. Comput. 14 (1993) 1007–1012.

[4] B. De Moor, On the structure and geometry of the product singular value decomposition, Linear Algebra Appl. 168 (1992) 95–136.

[5] B. De Moor, P. Van Dooren, Generalizations of the QR and singular value decomposition, SIAM J.

Matrix Anal. Appl. 13 (1992) 993–1014.

[6] K.S. Arun, S.Y. Kung, Generalized principal component analysis and its applications in approxi- mate stochastic realization, in: U.B. Desai (Ed.), Modelling and Application of Stochastic Processes, Kluwer Acdemic Publishers, Dordrecht, 1986, pp. 75–104.

[7] K.V. Fernando, S.J. Hammarling, A product induced singular value decomposition for two matrices and balanced realisation, NAG Technical Report TR8/87, 1987.

[8] K.V. Fernando, The Kalman reachability/observability canonical form and the SVD, NAG Techni- cal Report TR8/87, 1987.

[9] A.J. Laub, M.T. Heath, C.C. Paige, R.C. Ward, Computation of system balancing transformations and

other applications of simultaneous diagonalization algorithms, IEEE Trans. Automat. Control AC-32

(2) (1987).

Referenties

GERELATEERDE DOCUMENTEN

In all cases enlarged dipole lengths for large separations and augmented data sets using so called circulated data significantly increase the information content..

Key words: Generalized singular value decomposition (GSVD), partial GSVD, Jacobi–Davidson, subspace method, augmented matrix, correction equation, (inexact) accelerated Newton,

system suffered of a high plasma background radiation signal caused by strong AI and All lines that reached the output slit of the monochromator by multiple

Our method is still fundamentally pencil based; however, rather than using a single pencil and computing all of its generalized eigenvectors, we use many different pencils and in

• We derive necessary and sufficient conditions for the uniqueness of a number of simultaneous matrix decompositions: (1) simultaneous diagonalization by equivalence or congruence,

Zha, Analysis and numerical solution of fixed rank matrix approximation with general weighting matrices, ESAT-SISTA Report 1989-22, 28 pp., October 1989,

Future research topics will include: further investigation of the properties of the SVD in the extended max algebra, development of efficient algorithms to

A Simultaneous Generalized Schur Decomposition (SGSD) approach for computing a third-order Canonical Polyadic Decomposition (CPD) with a partial symmetry was proposed in [20]1. It