• No results found

In this work, we consider the decomposition of a third-order I × J × K tensor T into a sum of multilinear rank (M

N/A
N/A
Protected

Academic year: 2021

Share "In this work, we consider the decomposition of a third-order I × J × K tensor T into a sum of multilinear rank (M"

Copied!
2
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

Algebraic and Optimization-based Algorithms for Decomposing Tensors into Block Terms

Nico Vervliet, Ignat Domanov, Lieven De Lathauwer Department of Electrical Engineering (ESAT)—Stadius, KU Leuven

https://homes.esat.kuleuven.be/~nvervlie Abstract

Tensors, or multiway arrays of numerical values, and their decompositions have been applied suc- cessfully in a myriad of applications in, a.o., signal processing, data analysis and machine learning [1]. Key is the ability to use decompositions to extract simple components under mild uniqueness conditions. The simplest decomposition, the canonical polyadic decomposition (CPD), writes a tensor as a sum of rank-1 tensors. However, in some applications more complex terms are required and a block term decomposition (BTD) may be more suitable.

In this work, we consider the decomposition of a third-order I × J × K tensor T into a sum of multilinear rank (M

r

, N

r

, ·) terms:

T =

R

X

r=1

D

r

·

1

A

r

·

2

B

r

(1)

in which D

r

, A

r

and B

r

have dimensions M

r

× N

r

× K, I × M

r

and J × N

r

, resp., ·

n

denotes the mode-n product, i.e., T = D ·

n

A ⇔ T

(n)

= AD

(n)

, and the subscript

(n)

denotes the mode-n unfolding; see [1]. Alternatively, one can see this as the joint block diagonalization of the frontal slices T

k

, k = 1, . . . , K:

T

k

= A · blockdiag(D

1,k

, . . . , D

R,k

) · B

T

, (2) in which A = [A

1

, . . . , A

R

] and B = [B

1

, . . . , B

R

]. This problem has been studied for some special cases; see, e.g., [2, 3]. We tackle the general BTD, which is often solved via a nonconvex optimization problem which has many local optima. For the CPD and the decomposition into multilinear rank (L

r

, L

r

, 1) terms (LL1), algebraic methods often provide suitable initializations.

Here, we aim to derive an algebraic and an optimization-based method for problem (1).

Algebraic algorithm We show that the decomposition (1) can be cast as a coupled simultaneous eigenvalue decomposition (CS-EVD) problem under mild assumptions. Let T ∈ K

I×J ×K

(K can be R or C) admit a decomposition into a sum of R multilinear rank (M

r

, N

r

, ·) terms (Eq. (1)) and assume P

R

r=1

M

r

≤ I and P

R

r=1

N

r

≤ J . (Using compression by multilinear singular value decomposition [4], we can assume w.l.o.g. that these are equalities.) Let A and B have full column rank. Then, under some mild conditions, the decomposition in Eq. (1) is unique (see [5]), and can be reduced to a CS-EVD as follows. Let M ∈ K

IJ K×(I2+J2)

be defined as

M =

T

T1

⊗ I

I

−I

J

⊗ T

1

.. . .. . T

TK

⊗ I

I

−I

J

⊗ T

K

 . (3)

We can show that null(M) =: N has dimension R and can be partitioned as X Y



in which X ∈ K

I

2×R

and Y ∈ K

J2×R

. Let X

r

∈ K

I×I

(Y

r

∈ K

J ×J

) be the reshaped rth column x

r

(y

r

),

1

(2)

r = 1, . . . , R, of X (Y). We can show that A and B can be recovered from the CS-EVD problem X

r

= A · blockdiag(g

1r

I

M1

, . . . , g

Rr

I

MR

) · A

−1

, r = 1, . . . , R, (4) Y

r

= B

−T

· blockdiag(g

1r

I

N1

, . . . , g

Rr

I

NR

) · B

T

, r = 1, . . . , R. (5) Hence, we have two sets of simultaneous EVD problems that are coupled through the eigenvalues g

rs

, r = 1, . . . , R, s = 1, . . . , R, which result in the factors A and B in Eq. (2). The core tensors D

r

can be recovered from Eq. (2) as A and B are invertible by assumption. Note that R, M

r

and N

r

can be determined automatically from the dimension of the null space and the eigenvalues g

rs

. To reduce the computational cost of computing the null space N of the IJ K ×(I

2

+J

2

) matrix M, an iterative EVD algorithm is used to compute N as the subspace corresponding to the zero eigenvalues of M

H

M. The bottleneck computation is solving M

H

Mx = y for x every iteration. We show that the QR factorization of M = QR can be computed efficiently by exploiting the Kronecker and block structures using ideas from the block-QR and quasi-QR factorizations [6]. This way, M

H

Mx = y can be written as R

H

Rx = y which can be solved using forward and backward substitution as R is upper triangular. The cost of computing N then is O(J

4

+ I

2

J

2

+ I

3

) flop per iteration for the EVD step and O(J

6

+ IJ

4

K + I

2

J

3

K) flop for the QR step.

Optimization-based algorithm The CS-EVD problem (Eqs. (4)–(5)) is a coupled LL1 decom- position: by stacking X

r

and Y

r

as frontal slices in X and Y, resp., we have (using CPD notation):

X = JA, A

−T

, GP

1

K, Y = JB

−T

, B, GP

2

K, (6)

in which G collects the eigenvalues in Eqs. (4)–(5) and P

1

= blockdiag(1

TM

1

, . . . , 1

TM

R

) and P

2

= blockdiag(1

TN1

, . . . , 1

TN

R

). We propose a single-step optimization-based approach that exploits this and computes the null space and the decomposition simultaneously by solving

A,B,G

min 1 2

K

X

k=1

JA, T

Tk

A

−T

, GP

1

K − JT

k

B

−T

, B, GP

2

K

2

F

. (7)

This single-step approach has several advantages: increased robustness to noise, reduced accumu- lation of numerical error, and the possibility to exploit structure in both M and N simultaneously.

The problem can be solved efficiently using Gauss–Newton in combination with standard CPD techniques and parametric constraints; see [7]. To avoid zero solutions, we can set G = I.

[1] N. D. Sidiropoulos, L. De Lathauwer, X. Fu, K. Huang, E. E. Papalexakis, and C. Faloutsos. Tensor decomposition for signal processing and machine learning. IEEE Trans. Signal Process., 65 (13): pp.

3551–3582, 2017

[2] L. De Lathauwer. Decompositions of a higher-order tensor in block terms—Part II: Definitions and uniqueness. SIAM J. Matrix Anal. Appl., 30 (3): pp. 1033–1066, 2008

[3] G. Chabriel, M. Kleinsteuber, E. Moreau, H. Shen, P. Tichavsk´ y, and A. Yeredor. Joint matrices decompositions and blind source separation. A survey of methods, identification and applications. IEEE Signal Process. Mag., 31: pp. 34-43, 2014

[4] L. De Lathauwer, B. De Moor, and J. Vandewalle. A multilinear singular value decomposition. SIAM J. Matrix Anal. Appl., 21 (4): pp. 1253-1278, 2000

[5] I. Domanov, N. Vervliet, and L. De Lathauwer. Decomposition of a tensor into multilinear rank (M

r

, N

r

, ·) terms. Internal Report 18-51, ESAT-STADIUS, KU Leuven (Leuven, Belgium), 2018 [6] G. Stewart. Error analysis of the quasi-Gram–Schmidt algorithm. SIAM J. Matrix Analysis Applications,

27: pp. 493-506, 2005

[7] N. Vervliet, and L. De Lathauwer. Numerical optimization-based algorithms for data fusion. Chapter 4 of Data Fusion Methodology and Applications, Cocchi M., ed., vol 33 of Data Handling in Science and Technology, Elsevier, pp. 81-128

2

Referenties

GERELATEERDE DOCUMENTEN

multilinear algebra, higher-order tensor, singular value decomposition, canonical polyadic decomposition, block term decomposition, blind signal separation, exponential polynomial..

IEEE Trans. Signal Processing, vol. Cichocki, “Canonical Polyadic De- composition based on a single mode blind source separation,” IEEE Signal Processing Letters, vol.

We illustrate that by simultaneously taking the tensor nature and the block-Toeplitz structure of the problem into account new uniqueness results and nu- merical methods for computing

Citation/Reference Domanov I., De Lathauwer L., ``Canonical polyadic decomposition of third-order tensors: relaxed uniqueness conditions and algebraic algorithm'', Linear Algebra

IEEE Trans. Signal Processing, vol. Cichocki, “Canonical Polyadic De- composition based on a single mode blind source separation,” IEEE Signal Processing Letters, vol.

De Lathauwer, Blind Signal Separation via Tensor Decomposition with Vandermonde Factor – Part II: Chan- nels with Coherent or Incoherent Multipath and Small Delay Spread,

and [7] that, as in the matrix case, some configurations of ML singular values are not possible but, nevertheless, at least for n × · · · × n tensors the set of ML singular values has

In this section we 1) prove Theorems 11 and 12; 2) show that the assump- tions in Theorem 14 are more restrictive than the assumptions in Theorem 13, which implies the statement