• No results found

Multiple Invariance ESPRIT: A Coupled Matrix-Tensor Factorization Approach

N/A
N/A
Protected

Academic year: 2021

Share "Multiple Invariance ESPRIT: A Coupled Matrix-Tensor Factorization Approach"

Copied!
13
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

Multiple Invariance ESPRIT: A Coupled

Matrix-Tensor Factorization Approach

Mikael Sørensen and Lieven De Lathauwer, Fellow, IEEE

Abstract—The Canonical Polyadic Decomposition (CPD) of higher-order tensors has proven to be an important tool for array processing. CPD approaches have so far assumed regular array geometries such as uniform linear arrays. How-ever, in the case of sparse arrays such as Nonuniform Linear Arrays (NLAs), the CPD approach is not suitable anymore. Using the Coupled Matrix-Tensor Factorization (CMTF) we propose in this paper a multiple invariance ESPRIT method for both one- and multi-dimensional NLA processing. We obtain a multiresolution ESPRIT method for sparse arrays with multiple baselines. The CMTF framework also yields a new uniqueness condition that is relaxed compared to the CPD approach. It also leads to an eigenvalue decomposition based algorithm that is guaranteed to reduce the multi-source NLA problem into decoupled single-source NLA problems in the noiseless case. Finally, we present a new polynomial rooting procedure for the latter problem, which again is guaranteed to find the solution in the noiseless case.

Index Terms—array processing, direction-of-arrival estima-tion, tensor, coupled decomposiestima-tion, sparse array, multiple baseline, multiresolution, nonuniform sampling.

I. Introduction

The connection between Uniform Linear Arrays (ULAs) and the Canonical Polyadic Decomposition (CPD) is well-known, see [17] for a discussion. In order to increase the aperture or to deal with malfunctioning sensors in ULAs, Nonuniform Linear Arrays (NLAs) have been used, see [10], [27], [5], [12], [25], [13] and references therein. Because of the nonuniform sampling, the CPD framework is not suitable for NLA processing. Many of the developed methods for NLA make the assumption that the impinging signals are uncorrelated (orthogonal). This assumption often makes it possible to transform the original NLA problem into a uniformly sampled (multidimensional) harmonic retrieval problem. As our contribution, we present a link between the Coupled Matrix-Tensor Factorization (CMTF) model [20], [21] and NLA processing. A key feature of the CMTF approach is that it works directly on the data (e.g. no data imputation/interpolation or signal statistics are used and orthogonality is not assumed). Interestingly, the CMTF approach can be interpreted as a Multiple In-variance ESPRIT (MI-ESPRIT) method [23] for one- and M. Sørensen and L. De Lathauwer are with KU Leuven - E.E. Dept. (ESAT) - STADIUS Center for Dynamical Systems, Signal Processing and Data Analytics, Kasteelpark Arenberg 10, B-3001 Leuven-Heverlee, Belgium, the Group Science, Engineering and Technology, KU Leuven Kulak, E. Sabbelaan 53, 8500 Kortrijk, Belgium, and iMinds Medi-cal IT, Kasteelpark Arenberg 10, B-3001 Leuven-Heverlee, Belgium, {Mikael.Sorensen, Lieven.DeLathauwer}@kuleuven-kulak.be.

multi-dimensional NLA. We obtain a new uniqueness condition and derive an algorithm for NLA processing. In particular, we adapt some of the ULA results in [18] to NLAs.

A related problem concerns Direction-Of-Arrival (DOA) estimation based on sparse arrays embedded with multiple baselines (e.g. [6], [29], [8], [1]). The multi-ple baseline property can make sparse arrays ambiguity-free when several of the involved shift-invariance struc-tures are exploited together. However, the existing re-sults have so far been limited to dual shift-invariance structures and only CPD based methods have been considered. We explain that the CMTF model provides an algebraic framework for sparse array processing in the case of multiple baselines. This CMTF approach can be interpreted as a multiresolution ESPRIT method [8].

Even though we focus on sensor array processing, the results presented in this paper may also be used in, e.g., system identification and parametric spectral analysis.

The paper is organized as follows. The rest of the introduction will present the notation. Section II reviews the necessary algebraic prerequisites. Section III presents an algebraic CMTF framework for 1D NLA processing including a relaxed uniqueness condition. An algebraic algorithm is developed in Section IV. A nice feature of the CMTF approach is that it can easily be generalized to the multidimensional case. This is illustrated in Sec-tion V for two-dimensional (2D) arrays composed of superimposed NLAs (e.g. L-shaped arrays with missing sensors). Numerical experiments are reported in section VI. Section VII concludes the paper.

Notation: Vectors, matrices and tensors are denoted by

lower case boldface, upper case boldface and upper case calligraphic letters, respectively. The rth column vector and transpose of a matrix A are denoted by ar and AT,

respectively. The symbols⊗ and " denote the Kronecker and Khatri-Rao product, defined as

A⊗B :=     a11B a12B . . . a21B a22B . . . .. . ... . ..    , A"B := [a1⊗ b1 a2⊗ b2. . . ] , in which (A)mn = amn. The outer product of three

vectors a, b and c is denoted by a◦ b ◦ c, such that (a◦ b ◦ c)ijk=aibjck. The range, kernel, Frobenius norm,

real part and imaginary part of matrix A are denoted by range (A), ker (A),$A$F, Re{A} and Im{A}. From the

(2)

unit number, i.e., i = √−1.

Matlab index notation will be used for submatrices of a given matrix. For example, A(1:k,:) represents the submatrix of A consisting of the rows from 1 to k of A. Let A ∈ CI×R, then A = A (2 : I, :) ∈ C(I−1)×R and A =

A(1 : I− 1, :) ∈ C(I−1)×R, i.e., A and A are obtained by

deleting the top and bottom row of A, respectively. Let Cm,n ∈ Cmn×mn denote the commutation matrix

property with Cm,n(a⊗ b) = b ⊗ a, where a ∈ Cm and b ∈ Cn. We will also make use of the linear operator Sn(X) = C2,n

'I nX InX (

∈ C2(n−1)×p, where X∈ Cn×p and where

In∈ Cn×ndenotes the identity matrix.

The cardinality of a set S is denoted by card (S). The binomial coefficient is denoted by Ck

m = k!(mm!−k)!. The

k-th compound matrix of A∈ CI×R is denoted by C

k(A)

CCkI×CkR. It is the matrix containing the determinants of all

k×k submatrices of A, arranged with the submatrix index sets in lexicographic order, see [3] and references therein for a discussion. The modulus and argument of a ∈ C are denoted by |a| and arg(a), respectively. The greatest common divisor of M, N∈ N is denoted by GCD(M, N).

II. Algebraic Prerequisites

A. Canonical Polyadic Decomposition (CPD)

Consider the third-order tensorX ∈ CI×J×K. We say that

X is a rank-1 tensor if it is equal to the outer product of non-zero vectors a ∈ CI, b ∈ CJ and c ∈ CK such

that xijk = aibjck. A Polyadic Decomposition (PD) is a

decomposition ofX into a sum of rank-1 terms: CI×J×K) X =

R

)

r=1

ar◦ br◦ cr. (1)

The rank of a tensorX is equal to the minimal number of rank-1 tensors that yield X in a linear combination. Assume that the rank of X is R, then (1) is called the CPD of X. Let us stack the vectors {ar}, {br} and {cr}

into the matrices A = [a1, . . . , aR], B = [b1, . . . , bR] and C =[c1, . . . , cR]. The matrices A, B and C will be referred

to as the factor matrices of the (C)PD ofX in (1).

1) Matrix Representation: Let X(i··) ∈ CJ×K denote the

matrix such that *X(i··)+

jk = xijk, then X

(i··) = BD

i(A) CT

and we obtain the following matrix representation of (1): CIJ×K) X :='X(1··)T . . . X(I··)T(T=(A" B) CT. (2)

2) Uniqueness: The uniqueness properties of the CPD

have been intensely studied in recent years, see [3], [4], [16] and references therein. In this paper we consider the case where one factor matrix has full column rank. For this case the following sufficient uniqueness condition has been obtained.

Theorem II.1. Consider the PD of X ∈ CI×J×K in (1). If

 

CC2has full column rank,(A)" C2(B) has full column rank,

(3)

then the rank of X is R and the CPD of X is unique [7],

[2], [3]. Generically1, condition (3) is satisfied if 2R(R− 1) ≤ I(I− 1)J(J − 1) and R ≤ K [2], [22].

The uniqueness condition (3) is fairly easy to check and yet very powerful. Moreover, under condition (3) the CPD can be computed by algebraic means [2]. More precisely, the CPD problem was reduced to a Simul-taneous matrix Diagonalization (SD) problem. The SD problem can in the exact case be solved via a standard Generalized EigenValue Decomposition (GEVD).

B. Coupled CPD and Coupled Matrix-Tensor Factorization

We say that a collection of tensors X(n)∈ CIn×Jn×K, n {1, . . . , N}, admits an R-term coupled PD if each tensor X(n) can be written as [20]: X(n)= R ) r=1 a(n)r ◦ b(n)r ◦ cr, n∈ {1, . . . , N}, (4)

with factor matrices A(n) = 'a(n)1 , . . . , a(n)R(, B(n) = '

b(n)1 , . . . , b(n)R(and C = [c1, . . . , cR]. We define the coupled

rank of {X(n)} as the minimal number of coupled

rank-1 tensors a(n)r ◦ b(n)r ◦ cr that yield {X(n)} in a linear

combination. Assume that the coupled rank of {X(n)} is R, then (4) will be called the coupled CPD of {X(n)}.

The coupled (C)PD of {X(n)} given by (4) has the

following matrix representation

X ='X(1)T, . . . , X(N)T(T= FCT∈ C(0Nn=1InJn)×K, (5) where F = [(A(1)"B(1))T, . . . , (A(N)"B(N))T]T∈ C(0N

n=1InJn)×R. The coupled rank-1 tensors in (4) can be arbitrarily permuted and the vectors within the same coupled rank-1 tensor can be arbitrarily scaled provided that the overall coupled rank-1 term remains the same. We say that the coupled CPD is unique when it is only subject to these trivial indeterminacies. Uniqueness conditions for the coupled CPD were derived in [20]. Theorem II.2 extends Theorem II.1 to coupled CPD.

Theorem II.2. Consider the coupled PD ofX(n)∈ CIn×Jn×K,

n∈ {1, . . . , N} in (4). Define G =     C2*A(1)+" C2 * B(1)+ .. . C2*A(N)+" C 2 * B(N)+    ∈ C *0N n=1C2InC 2 Jn + ×C2 R. If  

CGhas full column rank,has full column rank, (6) then the coupled rank of{X(n)} is R and the coupled CPD of

{X(n)} is unique [20].

1A tensor decomposition property is called generic if it holds with probability one when the entries of the factor matrices are drawn from absolutely continuous probability density measures.

(3)

As in ordinary CPD, the uniqueness condition (6) is fairly easy to check and yet very powerful. An extension of the SD method in [2] to the coupled CPD has been developed in [21]. Again, in the noise-free case the SD problem for coupled CPD can be solved via a GEVD.

A special case of (4) is the CMTF of a matrix X∈ CI×K

and a set of tensors X(n)∈ CIn×Jn×K, n∈ {1, . . . , N}, given by 1 X =0Rr=1ar◦ cr, (br= 1), X(n)=0R r=1a (n) r ◦ b(n)r ◦ cr , n∈ {1, . . . , N}. (7) III. New uniqueness condition for NLA via CMTF Assume that R sources are impinging on an I element NLA. Let us additionally assume that the sources are located in the far-field, that the narrowband assumption holds and that K snapshots are available such that

sr∈ CKdenotes the signal vector associated with the rth

source. In this section we also assume that the sources are located in the “xy”-plane such that only the azimuth angles are of interest. Let us choose the first sensor in the array as the reference sensor. Under these assumptions the observed data matrix admits the factorization

X = AST∈ CI×K, A = [a1 . . . aR] , S = [s1 . . . sR] , (8) in which ar='1 e−iωcτr2 . . . e−iωcτrI(T=21 xn2 r . . . xnrI 3T ∈ CI, (9)

where τri=nidxcos(θr)/c corresponds to the propagation

delay associated with the ith sensor and the rth source, ωcis the carrier frequency, c is the propagation speed, dx

is the nominal unit measure along the sensor axis (“x-axis”), nidxis the nominal distance between the reference

sensor (n1= 0) and the ith sensor in the NLA in which ni∈ N, and xr=e−iωcdxcos(θr)/cin which 0≤ θr< π denotes

the azimuth angle associated with the rth source. Clearly

ar is an incomplete Vandermonde vector where not all powers appear. This is in contrast to the ULA case where all powers are present, e.g., ar= [1 xr x2r . . . xIr−1]T.

We define the NLA constrained rank of X in (8) as the minimal number of NLA constrained rank-1 matrices

arsT

r, where aris of the form (9), that yield X in a linear

combination. We say that the NLA decomposition of X is unique if for every pair (4A, 4S) satisfying the relation (8) we have 4A = AΠ, where Π is a permutation matrix. In DOA applications the goal is to estimate{θr} via A,

observing X. On the other hand, in source separation the goal is to estimate S, observing X. For simplicity, we will limit the exposition to the case where S has full column rank. However, the results presented in this paper can be extended to cases where S does not have full column rank.

In this section we establish a link between NLA pro-cessing and the CMTF. The approach can be interpreted as a MI-ESPRIT method [23] for NLA. More precisely, we explain that by simultaneously exploiting several shift-invariance structures inherent in the given NLA it is

under certain conditions possible to transform the NLA problem into a CMTF problem. The idea of exploiting several shift-invariance structures was first considered in [23] and later also in [13].

In order to clarify the idea we first provide an infor-mative example in subsection III-A. Next, in subsection III-B, we formalize the reasoning and establish the link between 1D NLA and CMTF. Subsection III-C briefly explains that the proposed CMTF approach also pro-vides a multiresolution interpretation of sparse arrays with multiple baselines.

A. Illustration

Consider the NLA observation matrix

X = AST∈ CI×K, a r= ' 1 xr x2r x4r x5r x6r x8r (T , I = 7. (10) The shift-invariance structures within A can be visual-ized as follows:               1 x x2 − x4 x5 x6 − x8 x 1 x x2 x x4 x5 x6 x2 1 x2 x4 x6 x8 x3 x x4 x3 x2 x5 x8 x4 1 x4 x8 x4 x x5 x4 x2 x6 x5 1 x5 x5 x x6 x6 1 x6 x6 x2 x8 x7 x x8 x8 1 x8               , (11)

where the outer column contains the generators from

x up to x8, the outer row denotes the incomplete

Van-dermonde vector ar (in which ’-’ denotes a missing

power of x) and each row inside the matrix denotes the sub-ULA associated with a particular partial shift-invariance. By inspection of (11) it becomes clear that the NLA can be interpreted as a set of superimposed ULAs. From X in (10) we extract the submatrices X(1,1) ∈ C3×K

and X(2,1) ∈ C3×K associated with x, and the submatrix

X(1,2)∈ C5×K associated with x2as follows:

X(1,1) = X([1, 2, 3], :) = A(1,1)ST , (12) X(2,1) = X([4, 5, 6], :) = A(2,1)ST , (13) X(1,2) = X([1, 3, 4, 6, 7], :) = A(1,2)ST , (14)

where the submatrices of A associated with the genera-tors x and x2, denoted by A(1,1)

∈ C3×R, A(2,1)

∈ C3×R and

A(1,2)∈ C5×R, have columns given by

a(1,1)r = ' 1 xrx2r (T , a(2,1)r = ' x4 r x5r x6r (T , (15) a(1,1)r = ' 1 x2r x4r x6r x8r (T . (16)

Note that the vectors (15)–(16) are (column-wise scaled) Vandermonde vectors. Exploiting the shift-invariance, we obtain from (12)–(14) tensors Y(1) ∈ C4×2×K and

(4)

Y(2)∈ C4×2×K of which the matrix representation admits

the coupled decomposition

Y(1) = [S2(X(1,1))T, S2(X(2,1))T]T= (A(1)" B(1))ST, Y(2) = S2(X(1,2)) = (A(2)" B(2))ST, where a(1)r = [a(1,1)Tr | a(2,1)Tr ] = [1 xr | xr4x5r]T, b(1)r = [1 xr]T, a(2)r = a(1,2)r = [1 x 2 r x4r x6r]T, b(2)r = [1 x2r]T.

We now obtain the coupled CPD of {Y(1),Y(2)} with

matrix representation Y = ; Y(1) Y(2) < = ; A(1)" B(1) A(2)" B(2) < ST. (17)

Thanks to Theorem II.2 the coupled CPD of{Y(1),Y(2)}

is generically unique if R≤ 5. Note that if we only use the CPD of Y(1), then DOA identifiability can only be

guaranteed for R≤ 4 (e.g. Theorem II.1).

B. 1D NLA via CMTF

Let us now formalize the reasoning behind the previ-ous example. The first step is to identify ULA structures within A in (8). In the second step we turn the matrix decomposition into a tensor decomposition via spatial smoothing. In the third and final step we combine the generated tensors, leading to a CMTF problem.

1) Multiple generators lead to multiple sub-ULAs:

Con-sider the NLA observation matrix X ∈ CI×K given by

(8) where A ∈ CI×R is a NLA matrix (e.g. of the form

(10)). We call xn

r a generator if there exist Im,n≥ 2 entries

of ar denoted by aσm,n(1),r, aσm,n(2),r, . . . , aσm,n(Im,n),r such that

aσm,n(i),r=aσm,n(1),rx

(i−1)n

r ,∀i ∈ {1, . . . , Im,n}. The first index of

Im,n denotes the mth subvector extracted from ar while

the second index of Im,n denotes the power of xr in the

generator (i.e., xn

r). As an example, in the cases where

n ∈ {1, 2}, the matrix A in (10) has ULA structured submatrices with properties:

I1,1= 3 σ1,1(1) = 1 σ1,1(2) = 2 σ1,1(3) = 3 I2,1= 3 σ2,1(1) = 4 σ2,1(2) = 5 σ2,1(3) = 6

I1,2= 5 σ1,2(1) = 1 σ1,2(2) = 3 σ1,2(3) = 4 σ1,2(4) = 6 σ1,2(5) = 7

Let Ω ⊆ {1, . . . , I} denote the set of powers of x in the generators. Assume that there are Mn

subma-trices of A associated with the generator xn, denoted

by A(1,n) ∈ CI1,n×R, . . . , A(Mn,n) ∈ CIMn ,n×R, with property

a(m,n)p+1,r=a(m,n)p,r xnr. In other words, the submatrices{A(Mn,n)}

are (column-wise scaled) Vandermonde matrices with generators{xn

r}. By way of illustration, consider A in (10)

and the subset Ω ={1, 2}. Then we get M1= 2 with A(1,1)

and A(2,1) given by (15) and M

2= 1 with A(1,2)given by

(16).

For each pair (m, n) we extract the corresponding submatrix X(m,n)∈ CIm,n×K from X in (8) as follows

X(m,n)= X([σm,n(1), . . . , σm,n(Im,n)], :) = A(m,n)ST, (18)

where A(m,n) = A([σm,n(1), . . . , σm,n(Im,n)], :), m

{1, . . . , Mn} and n ∈ Ω.

2) From matrices to tensors: The mapping S2(X(m,n))

producesY(m,n)∈ C(Im,n−1)×2×K with matrix representation

Y(m,n)=S2(X(m,n)) =*A(m,n)" B(n)+ST, (19)

where the rth column of A(m,n)∈ C(Im,n−1)×R is given by

a(m,n)r = a(m,n)r (1 : Im,n− 1, :), (20) and B(n) = ; 1 · · · 1 xn 1 · · · x n R < ∈ C2×R. (21) Define In = 0Mm=1n Im,n− Mn. Stacking Y(1,n), . . . , Y(Mn,n),

yieldsY(n)∈ CIn×2×K for which the matrix representation

Y(n)∈ CIn2×K admits the decomposition

Y(n)='Y(1,n)T . . . Y(Mn,n)T(T=*A(n)" B(n)+ST, (22) where

A(n) = 'A(1,n)T . . . A(Mn,n)T(T∈ CIn×R. (23)

3) From NLA to CMTF: Let us index the elements in

Ω by ω(1), . . . , ω(card (Ω)). For every element n∈ Ω it is assumed that Iω(n) ≥ 2. This condition is necessary for

the definition of G(Ω)given by (25). We now obtain the coupled CPD of {Y(n)} with matrix representation Y ∈

C(0n∈ΩIn)2×K given by Ytot=     Y(ω(1)) .. . Y(ω(card(Ω)))    =     A(ω(1))" B(ω(1)) .. . A(ω(card(Ω)))" B(ω(card(Ω)))    S T. (24)

Proposition III.1 makes use of the following matrix

G(Ω)=     C2*A(ω(1))+" C2 * B(ω(1))+ .. . C2*A(ω(card(Ω)))+" C2 * B(ω(card(Ω)))+    . (25)

Proposition III.1. Consider the NLA factorization of X =

AST∈ CI×Kin (8) where A is a NLA structured matrix with columns ar=21 xnr2 . . . xnrI 3T and dx≤ λ/2 = πcωc. If   

Sin (8) has full column rank, (26a)

G(Ω)in (25) has full column rank, (26b)

GCD(ni, nj) = 1 for some i ! j , (26c)

then the NLA constrained rank of X is R and the NLA decomposition of X is unique.

Proof: Since G(Ω) and S have full column rank, we

know from Theorem II.2 that the coupled rank of{Y(n)}

is R and the CMTF of {X, Y(n)} is unique. This in turn

implies that A is unique up to a column permutation and scaling. In other words, the CMTF of {X, Y(n)} has split

the problem into a set of single source DOA estimation problems. It is now easy to show that we can recover θr from zr = αrar, where αr ∈ C is an arbitrary scaling

factor. We add the details for completeness.

For simplicity, consider first the special case where nj

ni= 1 for some i < j. This implies that

'zir zjr ( = αr 'xni r xnjr ( ⇒

(5)

zjr/zir=xnrj−ni=xrfor some i < j. Since dxπcωc, nj−ni= 1

and 0 ≤ θr < π, we can uniquely recover θr from xr =

e−iωcdx(nj−ni) cos(θr)/c=e−iωcdxcos(θr)/c.

Consider now the general case in which nj− ni ≥ 1,

∀i < j and GCD(ni, nj) = 1 for some i ! j. We have

'z 1r zir ( = αr ' 1 xnir ( ⇒ br=zir/z1r=xnri, (27) 'z1r zjr ( = αr ' 1 xnjr ( ⇒ cr=zjr/z1r=x nj r . (28)

The solutions to (27) are x(br)

k = |br|

1/nie i(2πk+arg(br ))

ni , k

{0, 1, . . . , ni− 1}. Similarly, the solutions to (28) are x(cmr)=

|cr|1/nje

i(2πm+arg(cr))

nj , m∈ {0, 1, . . . , n

j− 1}. Hence, any solution

that satisfies both (27) and (28) must also satisfy the relation (2πk + arg(br))/ni= (2πm + arg(cr))/nj. (29) Note that bnj r = x ninj r = |br|njei arg(br)nj and cnri = x ninj r =

|cr|niei arg(cr)ni, implying that arg(br)nj= arg(cr)ni. Together

with relation (29) this implies that we have knj

ni = m.

Since GCD(ni, nj) = 1 and k < ni we conclude that the

only common solution to (27) and (28) is k = m = 0, cor-responding to arg(br)/ni = arg(cr)/nj = −ωcdxcos(θr)/c.

Since dxπcωc and 0≤ θr< π we can uniquely recover θr

from arg(br) and arg(cr).

We conclude that if conditions (26a)–(26c) are satisfied, then the NLA constrained rank of X is R and the NLA decomposition of X is unique.

Note that condition (26c) is necessary while conditions (26a) and (26b) are only sufficient. Even though we have only discussed the case of unit norm generators, Propo-sition III.1 is also valid for cases where the generators are not necessarily located on the unit circle.

Since each tensorY(n)∈ CIn×2×K, n∈ Ω,only consists of two matrix slices, the coupled CPD of{Y(n)} can be

inter-preted as a coupled Matrix-Pencil (MP) problem. More precisely, let P(n) ∈ C2In×2In denote a row-permutation matrix such that Z(n) := P(n)Y(n) = (B(n)" A(n))ST. Con-sider the partitioning Z(n) = 'Z(1,n)

Z(2,n) ( = ' A(n)ST A(n)D 2(B(n))ST ( with Z(1,n), Z(2,n) ∈ CIn×K. Define ΓΓΓ = [γγγ 1, . . . , γγγR] := (ST)† and ΥΥΥ =D1([λ1, . . . , λR]) := (D1([x1, . . . , xR]))−1. Then we

obtain a coupled MP problem

Z(1,n)ΓΓΓ = Z(2,n)ΓΓΓΥΥΥn⇔ (Z(1,n)− λn

rZ(2,n))γγγr= 0In, (30) where r∈ {1, . . . , R} and n ∈ Ω. In other words, the fac-torization problem (8) amounts to finding the common generalized eigenvectors of a set of MPs{(Z(1,n), Z(2,n))}

n∈Ω

subject to a power constraint on the associated gener-alized eigenvalues ({λn

r}n∈Ω). The power constraint can

make the generalized eigenvalues (λr) unique despite

min(Ω) > 1. Likewise, the coupling between the MPs can make the common generalized eigenvectors (γγγr) in (30)

unique, even in cases where for certain n, Z(1,n)and Z(2,n)

have overlapping kernels (ker*Z(1,n)+∩ ker*Z(2,n)+!∅) so that the eigenvector in (30) is not unique. Note also that the ordinary linear MP case (Ω = {1}) corresponds to classical ESPRIT [15].

Let us illustrate that the aforementioned power con-straint (λn

r) leads to improved DOA identifiability.

Con-sider the NLA for which A has columns of the form:

ar= [1 x3

r x6r x8r]T. (31)

Consider also the sub-ULAs [1 x3

rx6r]Tand [x6rx8r]Tof (31)

with inter-element spacings n = 3 and n = 2, respectively. Since Ω = {2, 3} and GCD(2, 3) = 1, Proposition III.1 guarantees the uniqueness of a generic NLA decompo-sition in which A has columns of the form (31) if R≤ 2. On the other hand, if only the sub-ULA [1 x3

r x6r]T is

used and the element x8

r in (31) is ignored, then unique

DOA recovery is not possible since now Ω = {3} and

min(Ω) > 1. Indeed, the associated MP involves only x3

r

and not xr itself. Similarly, if only the sub-ULA [x6r x8r]T

is used, then unique DOA recovery is not possible since Ω ={2} and min(Ω) > 1.

To illustrate that coupling can reduce the dimension of the eigenspace, consider the NLA for which A has columns of the form:

ar= [1 xrx2r x11r x15r x18r x21r x23r ]T. (32)

By simultaneously exploiting the multiple

shift-invariance structures in (32) that are associated with the generators xr, x2r, x3r and x10r , Proposition III.1 guarantees

the uniqueness of a generic NLA decomposition with columns of the form (32) if R≤ 3. Note that if we only exploit one of the sub-ULAs in (32), say [1 xr x2r]T, then

the more restrictive NLA uniqueness condition R≤ 2 is obtained.

This does not not mean that identifiability properties for all possible NLAs can be derived via the CMTF framework. For instance, nonredundant arrays [26] such as ar = [1 xr x4r x9r x11r ]T do not contain any

shift-invariance structure and hence the CMTF approach does not apply if R > 1.

C. Multiresolution interpretation of CMTF for NLA

Consider the NLA response matrix A with columns of the form (32). We note that the partial sub-ULAs within A have generators with small and large powers, i.e., ar(1)xr = ar(2), ar([1, 7])x2r = ar([3, 8]), ar([1, 7])x2r = ar([3, 8]), ar([5, 6])x3r = ar([6, 7]) and ar([2, 4])x10r = ar([4, 7]). We say that the array is equipped with multiple baselines. This property has been used to develop arrays with multiresolution properties (e.g. [6], [29], [8], [1]). An example taken from [1] is the case of three separated ULAs where the response matrix A has columns of the form ar= [1 xrx2rx14r x15r x21r x22r x23r ]T. We note that arhas

generators xr, x7r, x8r, x13r , x14r and x20r . Roughly speaking,

a small baseline can lead to coarse but unique DOA estimates while a large baseline can lead to accurate but ambiguous DOA estimates. The idea behind array processing involving multiple baselines is to combine the advantages of small and large baselines.

A related example is nested arrays or co-prime arrays composed of superimposed or concatenated ULAs with

(6)

different inter-element spacings [12], [25], [11], [24], [14]. An example could be two superimposed ULAs where the response matrix A has columns of the form ar =

[1 x5

r x7r x10r x14r x15r xr20 x21r x25 x28r x30r ]T. From A we can

extract A(1), A(5)and A(7)with columns of the form a(1)r =

[x14

r x15r x20r xr21]T, a(5)r = [1 x5r x10r x15r x20r x25r x30r ]T and a(7)r = [1 x7r x14r x21r x28r ]T, respectively. It is now clear that

the CMTF approach for NLA can also be used for nested arrays or co-prime arrays without making orthogonality assumptions on S, or by making use of any kind of data imputation/interpolation.

The CMTF approach also supports the use of hybrid arrays composed of sparse arrays with periodic subar-rays [28].

Actually, in the case of sparse array processing, the CMTF approach for NLA can be interpreted as a mul-tiresolution variant of ESPRIT. A distinct advantage of the CMTF approach compared to the existing methods (e.g. [6], [29], [8], [1], [24]) is that it simultaneously exploits the structure of the involved subarrays. As such, none of the involved subarrays is required to yield unique DOAs and also a more relaxed bound on R may be obtained.

As an example, consider the NLA response matrix

A = [A(2)T A(11)T A(14)T]T ∈ C15×R with columns a(2) r = [1 x2 r x4r x6r x8r]T, a (11) r = [x11r x21r xr31 x41r x51r ]T and a (14) r = [x14

r x28r x46r x62r x76r ]T. Methods that process the sub-ULAs

individually can identify at most R≤ 4 DOAs. However, Proposition III.1 implies that this NLA factorization is generically unique if R≤ 6.

IV. New Simultaneous matrix Diagonalization (SD)

method forMI-ESPRIT

In this section we present an SD method for MI-ESPRIT that will be used to compute the NLA factor-ization of X.

a) Step 0: Dimensionality reduction: Assume that the

full column rank conditions (26a) and (26b) in Propo-sition III.1 are satisfied, then A in (8) has full column rank. Indeed, if A does not have full column rank, then there exists a vector y ∈ ker (A) and a vector x ∈ CK

such that X = A(S + xyT)T has rank R− 1. This in turn

contradicts Proposition III.1. Let X = UΣΣΣVH denote the compact SVD of X, where U ∈ CI×R and V ∈ CK×R are

columnwise orthonormal and ΣΣΣ∈ CR×R is diagonal. We

know that range (X) = range (U) and that there exists a nonsingular matrix F∈ CR×Rsuch that

ΣΣ = AFTand VF−1= S. (33)

b) Step 1: From NLA factorization to coupled CPD:

For every m∈ {1, . . . , Mn} and for every n ∈ Ω we extract

from UΣΣΣ in (33) the compressed variant of (18):

U(m,n)= U([σ

m,n(1), . . . , σm,n(Im,n)], :)ΣΣΣ = A(m,n)FT. (34)

From the matrices{U(m,n)} in (34) we obtain the following

compressed variant of (22): Y(n)red=     S2(U(1,n)) .. . S2(U(Mn,n))    = * A(n)" B(n)+FT, n∈ Ω, (35)

where ’red’ stands for reduced. Stacking for all n ∈ Ω yields the following compressed variant of (24):

Ytot,red=     Y(ω(1))red .. . Y(ω(card(Ω)))red    =     A(ω(1))" B(ω(1)) .. . A(ω(card(Ω)))" B(ω(card(Ω)))    F T. (36) Clearly the low rank structured matrix decomposition (36) corresponds to a matrix representation of the cou-pled CPD of the tensorsY(n)red∈ CIn×2×K, n∈ Ω.

c) Step 2: From coupled CPD to SD: Assume that the

full column rank conditions (26a) and (26b) in Propo-sition III.1 are satisfied, then there exist R symmetric matrices{M(r)} built from the matrix (36), admitting the factorizations (see [21], [18] for the precise construction):

M(r)= GΛΛΛ(r)GT, r∈ {1, . . . , R}, (37) where G = F−1 and where ΛΛΛ(r) ∈ CR×R are diagonal

matrices. It is explained in [21], [18] that the complexity of the construction of{M(r)} is only of order 4(0n∈ΩIn)R2.

In the exact case, the SD problem (37) can be solved by means of an eigenvalue decomposition (e.g. [9]), with the columns of G−T as generalized eigenvectors.

d) Step 3: From SD to single-source DOA retrieval:

After G = F−1 has been found from (37), we obtain A and S via (33). The generator xr can now be obtained

from ar. More precisely, define

e(n)r =     a(1,n)r .. . a(Mn,n) r     and f (n) r =     a(1,n)r .. . a(Mn,n) r    , n∈ Ω. (38) If xr∈ C, then due to conditions (26a)–(26c) in

Proposi-tion III.1 we know that the generator can be estimated as the minimizer of the univariate polynomial f (xr) =

0

n∈Ω===e(n)r xnr− f(n)r ===

2

F. However, in DOA estimation xr is

located on the unit circle. In that case xrmust satisfy the

conditions

e(n)r xnr = f(n)r , n∈ Ω subject to xr=ei·φr , 0≤ φr< π.

Proposition IV.1 below states that φrcan be found via

the rooting of a real polynomial.

Proposition IV.1.Consider f (φr) =0n∈Ω===er(n)ei·φr·n− f(n)r ===

2

F,

where e(n)r and f(n)r are given by (38) and Ω⊂ {1, . . . , I}. The

(7)

real polynomial of degree 2Nmax: g(t) =) n∈Ω > Re(n)r @)2n p=0 N)max−n k=0 > 2n p A> Nmax− n k A* − pt(2n−p+1+2k) + (2n− p)t(2n−p−1+2k)+· cos(1 2(2n− p)π) +Im(n)r } 2n ) q=0 N)max−n m=0 > 2n q A> Nmax− n m A* − qt(2n−q+1+2m) + (2n− q)t(2n−q−1+2m)+· sin(1 2(2n− q)π) A , (39)

where β(n)r = e(n)Hr f(n)r . Let ˆt denote the real root of g(t) that

minimizes f (2 tan−1(ˆt)), then

2 tan−1(ˆt) = arg min

0≤φr

f (φr).

Proof: The proof is technical and is given in the

supplementary material of [19].

We now explain that Proposition IV.1 is guaranteed to find the true unit norm generator xr= ei·φr in the exact

case. Recall that conditions (26a) and (26b) in Proposition III.1 guarantee that the columns{ar} of A are unique up

to individual scaling factors. This means that there exists a unit norm generator xr=ei·φrsuch that e(n)r xr= f(n)r ,∀n ∈

Ω. Finally, condition (26c) in Proposition III.1 guarantees the uniqueness of this unit norm generator.

e) Summary of algorithm: The overall procedure is

summarized as Algorithm 1. Note that if conditions (26a)–(26c) in Proposition III.1 are satisfied, then Algo-rithm 1 is guaranteed to find the exact decomposition in the noiseless case.

Algorithm 1SD method for MI-ESPRIT (Prop. III.1).

Input: X = ASTof the form (8).

1.Determine subset Ω⊆ {1, . . . , I}.

2. Compute SVD X = UΣΣΣVH(and find R from ΣΣΣ). 3. Extract U(m,n)in (34) from UΣΣΣ,∀m ∈ {1, . . . , Mn}, ∀n ∈ Ω.

4. Build Y(n)red in (35) for every n∈ Ω.

5. Build matrices{M(r)

} from {Y(n)red}n∈Ω[21], [18].

6. Solve SD problem M(r)= GΛΛΛ(r)GT

, r∈ {1, . . . , R}. 7. Compute A = UΣΣΣGTand S = VG.

8. Obtain{xr} from A (e.g. via Proposition IV.1).

Output:{xr}, A and S.

V. Extension to the multidimensional case A simple extension of the 1D NLA is an array built from superimposed NLAs. Here we adapt some of the results in [18] to the NLA case. Subsection V-A provides an illustrative example. Next, in Subsection V-B we formalize the extension to the multidimensional case.

A. Illustration

Consider an L-shaped array composed of I(1) sensors

along the “x1-axis” and I(2) sensors along the “x2-axis”

such that (8) extends to

X(1)= A(1)ST∈ CI

(1)×K

, X(2)= A(2)ST∈ CI

(2)×K

, (40)

where X(1) denotes the observation matrix associated

with the sensors along the “x1-axis” in which (A(1))r =

' 1 xn2,1 r,1 . . . x nI(1),1 r,1 (T ∈ CI(1)

, where xr,1=e−iωcd1sin(φr) cos(θr)/c,

d1 is the unit spacing along the “x1-axis” and the new

variable φr denotes the elevation angle associated with

the rth source. Similarly, X(2) denotes the observation

matrix associated with the sensors along the “x2-axis”

in which (A(2))r = ' 1 xn2,2 r,2 . . . x nI(2),2 r,2 (T ∈ CI(2) , where

xr,2=e−iωcd2sin(φr) sin(θr)/c and d2 is the unit spacing along

the “x2-axis”.

As in the 1D case, we define the NLA constrained and coupled rank of{X(1), X(2)} in (40) as the minimal

num-ber of NLA constrained and coupled rank-1 matrices {(A(1))rsTr, (A(2))rsTr} that yield {X(1), X(2)} in a linear

combi-nation. We also say that the coupled NLA decomposition of {X(1), X(2)} is unique if for every triplet (4A(1), 4A(2), 4S)

satisfying the relation (40) we have 4A(1) = A(1)Π and

4

A(2)= A(2)Π, where Π is a permutation matrix.

By way of example, let X(1) in (40) be of the form (10)

while X(2) in (40) admits the factorization

X(2)= A(2)ST∈ CI (2)×K , (A(2))r= ' 1 xr,2x4r,2x5r,2x8r,2x12r,2 (T . (41) From X(1) we obtain the coupled CPD of{Y(1)(1),Y(2)(1)} with

matrix representation (17). We will now capitalize on the NLA structure of A(2). Note that the vector (A(2))rin A(2)

contains the generators xr,2, x3r,2and x4r,2:

      1 x − − x4 x5 − − x8 − − − x12 x 1 x x x4 y5 x3 x x4 x3 x5 x8 x4 1 x4 x8 x12 x4 x x5       . (42)

We consider the following submatrices of A(2):

X(1,1)(2) = X(2)([1, 2], :) = A(1,1)(2) ST∈ C2×K, I(2)1,1= 2, (43) X(2,1)(2) = X(2)([3, 4], :) = A(2,1)(2) ST∈ C2×K, I(2)2,1= 2, (44) X(1,3)(2) = X(2)([2, 3], :) = A(1,3)(2) ST∈ C2×K, I(2)1,3= 2, (45) X(2,3)(2) = X(2)([4, 5], :) = A(2,3)(2) ST∈ C2×K, I(2)2,3= 2, (46) X(1,4)(2) = X(2)([1, 3, 5, 6], :) = A(1,4)(2) ST∈ C4×K, I(2)1,4= 4, (47) X(2,4)(2) = X(2)([2, 4], :) = A(2,4)(2) ST∈ C2×K, I(2)2,4= 2, (48) where A(1,1)(2) ='x11,2··· 1··· xR,2 ( , A(2,1)(2) ='x 4 1,2··· x4R,2 x5 1,2··· x5R,2 ( , A(1,3)(2) ='xx1,24 ··· xR,2 1,2··· x 4 R,2 ( , A(2,3)(2) ='x 5 1,2··· x5R,2 x8 1,2··· x 8 R,2 ( , A(1,4)(2) =    1 ··· 1 x4 1,2··· x4R,2 x8 1,2··· x8R,2 x12 1,2··· x 12 R,2   , A(2,4)(2) ='xx1,25 ··· xR,2 1,2··· x5R,2 ( .

From (43)–(48) we obtain the tensors Y(1)(2) ∈ C2×2×K ,

Y(3)(2)∈ C2×2×K andY(4)

(8)

matrix representations Y(1)(2)∈ C4×K, Y(3)

(2)∈ C4×K and Y (4) (2)∈

C8×K admit the following CPDs:

Y(1)(2) = 'S2(X(1,1)(2) )T, S 2(X(2,1)(2) )T (T = (A(1)(2)" B(1)(2))ST, Y(3)(2) = 'S2(X(1,3)(2) )T, S 2(X(2,3)(2) )T (T =*A(3)(2)" B(3)(2)+ST, Y(4)(2) = 'S2(X(1,4)(2) )T, S2(X(2,4)(2) )T (T =*A(4)(2)" B(4)(2)+ST, where A(1)(2)='A (1,1) (2) A(2,1)(2) ( ='x14 ··· 1 1,2··· x4R,2 ( , B(1)(2)='x11,2··· 1··· xR,2(, A(3)(2)='A (1,3) (2) A(2,3)(2) ( ='x15 ··· 1 1,2··· x5R,2 ( , B(3)(2)='x13 ··· 1 1,2··· x3R,2 ( , A(4)(2)='A (1,4) (2) A(2,4)(2) ( =    1 ··· 1 x4 1,2 ··· x4R,2 x8 1,2··· x8R,2 x1,2 ··· xR,2    , B(4)(2)='x14 ··· 1 1,2··· x4R,2 ( . We now obtain the coupled CPD of{Y(2)(1),Y(3)(2),Y(4)(2)} with matrix representation Y(2) =     Y(1)(2) Y(2)(2) Y(3)(2)    =     A(1)(2)" B(1)(2) A(3)(2)" B(3)(2) A(4)(2)" B(4)(2)    ST. (49) Overall, from (17) and (49) we obtain the coupled CPD of{Y(1)(1),Y(1)(2),Y(1)(2),Y(3)(2),Y(4)(2)}. By exploiting the cou-pling between the two NLAs imposed by S, improved uniqueness conditions are obtained. Indeed, from sub-section III-A we know that the coupled CPD structure of only {Y(1)(1),Y(2)(1)} yields the upper bound R ≤ 5. Likewise, from Theorem II.2 the coupled CPD structure of {Y(1)(2),Y(3)(2),Y(4)(2)} yields the upper bound R ≤ 3. On the other hand, by exploiting the overall structure of {Y(1)(1),Y(1)(2),Y(2)(1),Y(3)(2),Y(4)(2)}, Theorem II.2 relaxes the bound to R≤ 6.

B. CMTF framework for multidimensional NLA

Let us now formalize the reasoning behind the previ-ous example. Consider the decomposition

X = AST∈ CI×K, (50)

where the matrix A∈ CI×Ris shift-invariance structured

while S ∈ CK×R is unstructured. More precisely, let

J(p) ∈ CI(p)×R

denote a row-selection matrix such that

A(p) = J(p)A ∈ CI

(p)×R

consists of I(p) rows of A. It is

assumed that each of the following P decompositions

X(p)= J(p)X = A(p)ST∈ CI

(p)×K

, p∈ {1, . . . , P}, (51) corresponds to an NLA factorization of the form (8). This in turn means that A(p) ∈ CI

(p)×R

is an NLA matrix with columns (A(p))r = [xnr,p1,p x

n2,p

r,p . . . x nI(p),p

r,p ]T in which xr,p

C and ni,p ∈ Z. We can now proceed as in Section III. Briefly, let Ωp ⊆ {1, . . . , I(p)} denote the set of powers of

xr,p in the generators associated with pth NLA matrix A(p). Relation (18) now becomes

X(m,n)(p) = X(p)([σm,n,p(1), . . . , σm,n,p(I(p)m,n)], :) = A(m,n)(p) S

T∈ CI(p)m,n×K,

where A(m,n)(p) = A(p)([σm,n,p(1), . . . , σm,n,p(I(p)m,n)], :), m

{1, . . . , Mn,p}, n ∈ Ωp and the new index ’p’ refers to the

associated NLA factorization X(p). Likewise, relation (19)

extends to Y(m,n)(p) = S2(X(m,n)(p) ) = B A(m,n)(p) " B(n)(p) C ST, where B(n)(p) = 'x1n ··· 1 1,p··· xnR,p (

. Let us index the elements in Ωp by

ωp(1), . . . , ωp(card

* Ωp

+

). We now obtain the augmented version of (24): Ytot=     Ytot,1 .. . Ytot,P    , Ytot,p=      Yp(1)) (p) .. . Yp(card(Ωp))) (p)      , Y(n)(p) =     Y(1,n)(p) .. . Y(Mn,p,n) (p)    , in which Yp(n)) (p) = (Ap(n)) (p) " Bp(n)) (p) )S T and A(n) (p) = D A(1,n)T(p) . . . A(Mn,p,n)T (p) ET

. In the same way, the augmented version of (25) is given by

Gtot=

'

G(Ω1,1)T . . . G(ΩP,P)T(T, (52) where the index ’p’ indicates that the matrix G(Ωp,p), which is of the form (25), is built from the factor matrices {Ap(n),p)

(p) , Bp(n),p)

(p) } card(Ωp)

n=1 associated with the pth NLA.

Proposition V.1 below summarizes the extension of Proposition III.1 to the multidimensional case.

Proposition V.1. Consider the multidimensional NLA

fac-torization of X = AST∈ CI×Kin (50). If     

Sin (50) has full column rank, Gtotin (52) has full column rank,

∀p ∈ {1, . . . , P} , ∃ i, j ∈ Ωp : GCD(ni,p, nj,p) = 1 ,

(53)

then the multidimensional NLA constrained rank of X is R and the multidimensional NLA decomposition of X is unique.

Furthermore, if the conditions (53) in Proposition V.1 are satisfied, then Algorithm 1 can easily be extended to the multidimensional case. We summarize the extension as Algorithm 2.

Algorithm 2 SD method for multidimensional

MI-ESPRIT (Proposition V.1).

Input: X = AST of the form (50).

1.Compute SVD X = UΣΣΣVH(and find R from ΣΣΣ).

2. Choose row-selection matrices J(1), . . . , J(P).

3. Extract U(p)=J(p)UΣΣΣ for every p∈ {1, . . . , P}

4. Determine the subsets Ωp⊆ {1, . . . , I(p)}, ∀p ∈ {1, . . . , P}.

5. Extract{U(m,n) (p) }

m∈{1,...,Mn,p}

n∈Ωp of the form (34),∀p ∈ {1, . . . , P}.

6. Build Y(n)red,p of the form (35),∀n ∈ Ωp,∀p ∈ {1, . . . , P}.

7. Build{M(r) } in (37) from {Y(n)red,p} n∈Ωp p∈{1,...,P}[21], [18]. 8. Solve SD problem M(r)= GΛΛΛ(r)GT, r ∈ {1, . . . , R}. 9. Compute A(p)= J(p)ΣΣGT,∀p ∈ {1, . . . , P} and S = VG.

10. Obtain{xr,p} via A(p)(e.g. via Proposition IV.1). Output:{xr,p}, {A(p)} and S.

(9)

VI. Numerical experiments

Consider the factorization X = AST in (8) or (50). The goal is to estimate the angles {θr} or {θr, φr}

from T = X + βN, where N is an unstructured

perturbation matrix and β ∈ R controls the noise

level. The azimuth angles {θr} and elevation angles

r} are randomly chosen on the intervals [0, π] and

[−π/2, π/2], respectively. The real and imaginary en-tries of the perturbation matrices are randomly drawn from a Gaussian distribution with zero mean and unit variance. The following Signal-to-Noise Ratio (SNR) measure will be used: SNR [dB] = 10 log($X$2

F/$βN$2F).

The performance evaluation will be based on the dis-tance between the true angles {θr, φr} and their

es-timates { ˆθr, ˆφr}. The distance is measured according

to: ERMS = minσ∈ΛR F 1 R 0R r=1 * θr− ˆθσ(r) +2 +*φr− ˆφσ(r) +2 , where ΛRdenotes the set of R! possible permutations of

the elements {1, . . . , R} and σ(r) denotes the rth element of the corresponding permuted set.

A. One-dimensional (1D) NLA

Consider the NLA problem with I = 8, K = 50,

R = 3 and NLA response matrix A with columns of

the form (32). We compute the NLA factorization of (32) via Algorithm 1 in which the generators xr, x2r, x3r and

x10

r are used and the associated subvectors of (32) are

[1 xrx2r]T, [1 xr2 | x21r x23r ]T, [x15r x18r x21r ]T and [1 x11r x21r ]T,

respectively. Note that methods that only exploit one shift-invariance structure within the NLA require that

R≤ 2 and hence cannot be used. In order to demonstrate that the multiresolution property of A leads to improved performance, we compare the I = 8 element NLA with an I = 8 (number of elements in the NLA) element ULA and an I = 23 (aperture of the NLA) element ULA. The DOA estimation in the ULA cases will be based on the ESPRIT method [15]. In this experiment we only estimate the azimuth angles and consequently we set φr = ˆφσ(r)

in ERMS. The mean ERMS value with φr= ˆφσ(r)over 500

trials as a function of SNR can be seen in figure 1. We observe that the NLA performs significantly better than the I = 8 element ULA and almost as well as the I = 23 element ULA.

B. L-shaped two-dimensional (2D) NLA

Consider the L-shaped 2D NLA in subsection V-A with I(1)= 7, I(2)= 6, K = 50 and R = 4. Consider also an

(I(1), I(2)) = (7, 6) (number of elements in the L-shaped

2D NLA) element 2D ULA and an (I(1), I(2)) = (9, 13)

(aperture of the L-shaped 2D NLA) element 2D ULA. The DOA estimation in the 2D NLA case and in the 2D ULA cases will be based on Algorithm 2 and the SD algorithm developed in [18], respectively. The mean ERMS

value over 500 trials as a function of SNR can be seen in figure 2. As in the 1D case, 2D NLA performs better than the (I(1), I(2)) = (7, 6) element 2D ULA and almost as well as the (I(1), I(2)) = (9, 13) element 2D ULA.

VII. Conclusion

Sparse array processing presents many important al-gebraic challenges. We first made a link between (mul-tidimensional) NLA and the CMTF. This led to a new uniqueness condition for sparse array processing prob-lems involving shift-invariance structures. To the au-thors’ knowledge, this is the first formal proof that improved identifiability results for a sparse array can be obtained by simultaneously exploiting several shift-invariance structures within the sparse array.

Second, the CMTF approach also provided us with an algebraic SD method that is guaranteed to find the de-composition in the exact case if the presented uniqueness condition is satisfied. In the inexact case, it can be used as a computationally inexpensive initialization procedure for a more costly optimization method. The low cost property of the SD method also makes it interesting for shift-invariant structured large scale and sparse arrays.

Finally, numerical experiments indicated that by ex-ploiting the multiresolution property of sparse arrays, a good compromise between performance, identifiability, complexity and the number of sampling points (e.g. antennas) can be obtained.

10 20 30 40 0 0.05 0.1 0.15 0.2 SNR E RMS

NLA with I=8 elements ULA with I=8 elements ULA with I=23 elements

Fig. 1. Mean ERMSover 500 trials, 1D NLA and ULAs.

10 20 30 40 0 0.05 0.1 0.15 0.2 SNR E RMS

2D NLA with (I(1),I(2))=(7,6) elements 2D ULA with (I(1),I(2))=(7,6) elements 2D ULA with (I(1),I(2))=(9,13) elements

(10)

References

[1] F. Athley, C. Engdahl, and P. Sunnergren, “On radar detection and direction finding using sparse arrays,” IEEE Trans. Aerosp. Electron. Syst., vol. 43, no. 4, pp. 1319–1333, Oct. 2007.

[2] L. De Lathauwer, “A link between the canonical decomposition in multilinear algebra and simultaneous matrix diagonalization,” SIAM J. Matrix Anal. Appl., vol. 28, no. 3, pp. 642–666, 2006. [3] I. Domanov and L. De Lathauwer, “On the uniqueness of the

canonical polyadic decomposition of third-order tensors — Part I: Basic results and uniqueness of one factor matrix,” SIAM J. Matrix Anal. Appl., vol. 34, no. 3, pp. 855–875, 2013.

[4] ——, “On the uniqueness of the canonical polyadic decomposi-tion of third-order tensors — Part II: Overall uniqueness,” SIAM J. Matrix Anal. Appl., vol. 34, no. 3, pp. 876–903, 2013.

[5] C. El Kassis, J. Picheral, and C. Mokbel, “Advantages of nonuni-form arrays using root-MUSIC,” Signal Processing, vol. 90, pp. 689– 695, 2010.

[6] E. Jacobs and E. W. Ralston, “Ambiguity resolution in interferom-etry,” IEEE Trans. Aerosp. Electron. Syst., vol. 17, no. 6, pp. 766–780, Nov. 1981.

[7] T. Jiang and N. D. Sidiropoulos, “Kruskal’s permutation lemma and the identification of CANDECOMP/PARAFAC and bilinear model with constant modulus constraints,” IEEE Trans. Signal Process., vol. 52, no. 9, pp. 2625–2636, Sep. 2004.

[8] A. N. Lemma, A.-J. van der Veen, and E. F. Deprettere, “Multires-olution ESPRIT algorithm,” IEEE Trans. Signal Process., vol. 47, no. 6, pp. 1722–1726, Jun. 1999.

[9] S. Leurgans, R. T. Ross, and R. B. Abel, “A decomposition for three-way arrays,” SIAM J. Matrix Anal. Appl., vol. 14, no. 4, pp. 1064–1083, 1993.

[10] A. T. Moffet, “Minimum-redundancy linear arrays,” IEEE Trans. Antennas Propagat., vol. AP-16, no. 2, pp. 172–175, Mar. 1968. [11] P. Pal and P. P. Vaidyanathan, “Coprime sampling and the MUSIC

algorithm,” in Proc. of the 14th/6th IEEE DSP/SPE Workshop, Jan 4-7, Sedona AZ, USA, 2011.

[12] ——, “Nested arrays: A novel approach to array processing with enhanced degrees of freedom,” IEEE Trans. Signal Processing, vol. 58, no. 8, pp. 4167–4181, Aug. 2010.

[13] P. Parvazi, M. Pesavento, and A. B. Gershman, “Rooting-based harmonic retrieval using multiple shift-invariances: The complete and the incomplete cases,” IEEE Trans. Signal Processing, vol. 60, no. 4, pp. 1556–1570, Apr. 2012.

[14] S. Qin, Y. D. Zhang, and M. G. Amin, “Generalized coprime array configurations for direction-of-arrival estimation,” IEEE Trans. Signal Process., vol. 63, no. 6, pp. 1377–1390, Jun. 2015.

[15] R. Roy and T. Kailath, “Estimation of signal parameters via rota-tional invariance techniques,” IEEE Trans. Acoust., Speech, Signal Processing, vol. 32, no. 7, pp. 984–995, Jul. 1989.

[16] M. Sørensen and L. De Lathauwer, “New uniqueness conditions for the canonical polyadic decomposition of third-order tensors,” Accepted for publication in SIAM J. Matrix Anal. Appl.

[17] ——, “Multidimensional harmonic retrieval via coupled canonical polyadic decompositions,” ESAT-STADIUS, KU Leuven, Belgium, Tech. Rep. 13-240, 2013.

[18] ——, “Coupled tensor decompositions in array processing,” ESAT-STADIUS, KU Leuven, Belgium, Tech. Rep. 13-241, 2014. [19] ——, “Multiple invariance ESPRIT: A coupled matrix-tensor

fac-torization approach,” ESAT-STADIUS, KU Leuven, Belgium, Tech. Rep. 13-242, 2014.

[20] ——, “Coupled canonical polyadic decompositions and (coupled) decompositions in multilinear rank-(Lr,n, Lr,n, 1) terms — Part I: Uniqueness,” SIAM J. Matrix Anal. Appl., vol. 36, no. 2, pp. 496– 522, 2015.

[21] M. Sørensen, I. Domanov, and L. De Lathauwer, “Coupled canon-ical polyadic decompositions and (coupled) decompositions in multilinear rank-(Lr,n, Lr,n, 1) terms — Part II: Algorithms,” SIAM J. Matrix Anal. Appl., vol. 36, no. 3, pp. 1015–1045, 2015. [22] A. Stegeman, J.M.F. ten Berge, and L. De Lathauwer, “Sufficient

conditions for uniqueness in Candecomp/Parafac and Indscal with random component matrices,” Psychometrika, vol. 71, pp. 219–229, 2006.

[23] A. Swindlehurst, B. Ottersten, R. Roy, and T. Kailath, “Multiple invariance ESPRIT,” IEEE Trans. Signal Process., vol. 40, no. 4, pp. 868–881, Apr. 1992.

[24] P. P. Vaidyanathan and P. Pal, “Why does direct-MUSIC on sparse-arrays work ?” in Proc. Asilomar 2013, Nov 3-6, Pacific Grove, CA, USA, 2013.

[25] ——, “Sparse sensing with co-prime samplers and arrays,” IEEE Trans. Signal Process., vol. 59, no. 2, pp. 573–586, Feb. 2011. [26] E. Vertatschitsch and S. Haykin, “Nonredundant arrays,” Proc. of

the IEEE, vol. 74, no. 1, p. 217, Jan. 1986.

[27] A. J. Weiss, A. S. Wilsky, and B. C. Levy, “Nonuniform array processing via the polynomial approach,” IEEE Trans. Aerosp. Electron. Syst., vol. 25, no. 1, pp. 48–55, Jan. 1989.

[28] M. J. Wilson and R. McHugh, “Sparse-periodic hybrid array beamformer,” IET Radar Sonar Navig., vol. 1, no. 2, pp. 116–123, 2007.

[29] K. T. Wong and M. D. Zoltowski, “Direction-finding with sparse rectangular dual-size spatial invariance array,” IEEE Trans. Aerosp. Electron. Syst., vol. 17, no. 4, pp. 1320–1335, Oct. 1998.

Acknowledgment

Research supported by: (1) Research Council KU Leuven: CoE EF/05/006 Optimization in Engineering (OPTEC), C1 project c16/15/059-nD, (2) F.W.O.: project G.0830.14N, G.0881.14N, (3) the Belgian Federal Sci-ence Policy Office: IUAP P7 (DYSCO II, Dynamical systems, control and optimization, 2012–2017), (4) EU: The research leading to these results has received funding from the European Research Council under the European Union’s Seventh Framework Programme (FP7/2007-2013) / ERC Advanced Grant: BIOTENSORS (no. 339804). This paper reflects only the authors’ views and the Union is not liable for any use that may be made of the contained information.

Mikael Sørensenreceived the Master’s degree from Aalborg University, Denmark, and the Ph.D. degree from University of Nice, France, in 2006 and 2010, respectively, both in electrical engineering. Since 2010 he has been a Postdoc-toral Fellow with the KU Leuven, Belgium. His research interests include applied linear alge-bra, tensor decompositions, and tensor-based (array) signal processing.

Lieven De Lathauwer received the Master’s degree in electromechanical engineering and the Ph.D. degree in applied sciences from KU Leuven, Belgium, in 1992 and 1997, respectively. He is currently Professor with KU Leuven, Bel-gium. Dr. De Lathauwer is an Associate Editor of the SIAM Journal on Matrix Analysis and Applications and has served as an Associate Editor for the IEEE Transactions on Signal Pro-cessing. His research concerns the development of tensor tools for engineering applications.

(11)

MIKAEL SØRENSEN∗† AND LIEVEN DE LATHAUWER∗†

Proposition IV.1. Consider f (φr) =

! n∈Ω " " "e (n) r ei·φr·n − f(n)r " " " 2 F, where e (n) r

and f(n)r are given by [1, eq. (38)] and Ω ⊂ {1, . . . , I}. The largest element in Ω is

denoted by Nmax. Consider also the real polynomial of degree 2Nmax:

g(t) =# n∈Ω $ Re%βr(n) &#2n p=0 Nmax−n # k=0 '2n p ('Nmax− n k ( ) − pt(2n−p+1+2k) + (2n − p)t(2n−p−1+2k)*· cos(1 2(2n − p)π) + Im{βr(n)} 2n # q=0 Nmax−n # m=0 '2n q ('Nmax− n m ( ) − qt(2n−q+1+2m) + (2n − q)t(2n−q−1+2m)*· sin(1 2(2n − q)π) + , (S.0.1)

where βr(n)= e(n)Hr f(n)r . Let ˆt denote the real root of g(t) that minimizes f (2 tan−1(ˆt)),

then

2 tan−1t) = arg min 0≤φr<π

f (φr).

Proof. Let e(n)r and f(n)r be given by [1, eq. (38)] and xnr = ei·φr·nwhere 0 ≤ φr<

π. Consider f (φr) = # n∈Ω " " "e (n) r xnr − f(n)r " " " 2 F = # n∈Ω ) α(n) r − βr(n)e−i·φ r·n − β(n)∗ r ei·φ r·n * =# n∈Ω ' α(n)r − 2Re % βr(n) & cos(φr 2 · 2n) − 2Im{β (n) r } sin( φr 2 · 2n) ( =# n∈Ω $ α(n) r − 2 ˆβ(n)r 2n # p=0 g(φr 2, 2n, p) − 2 ˇβ (n) r 2n # q=0 h(φr 2, 2n, q) + , (S.0.2)

Group Science, Engineering and Technology, KU Leuven - Kulak, E. Sabbelaan 53, 8500 Kor-trijk, Belgium, KU Leuven - E.E. Dept. (ESAT) - STADIUS Center for Dynamical Systems, Signal Processing and Data Analytics, and iMinds Medical IT Department, Kasteelpark Arenberg 10, B-3001 Leuven-Heverlee, Belgium. {Mikael.Sorensen, Lieven.DeLathauwer}@kuleuven-kulak.be.

Research supported by: (1) Research Council KU Leuven: CoE EF/05/006 Optimization in Engineering (OPTEC), C1 project c16/15/059-nD, (2) F.W.O.: project G.0830.14N, G.0881.14N, (3) the Belgian Federal Science Policy Office: IUAP P7 (DYSCO II, Dynamical systems, control and optimization, 2012–2017), (4) EU: The research leading to these results has received funding from the European Research Council under the European Union’s Seventh Framework Programme (FP7/2007-2013) / ERC Advanced Grant: BIOTENSORS (no. 339804). This paper reflects only the authors’ views and the Union is not liable for any use that may be made of the contained information.

Referenties

GERELATEERDE DOCUMENTEN

A double coupling model is proposed which allows for subject variation in the Hemodynamic Re- sponse Function (HRF) and misspecifications in the coupling of the two modalities, by

Our receiver is deterministic and relies on a third-order tensor decomposition, called decomposition in rank-(L,L,1) terms, which is a generalization of the well-known Parallel

IV. MI-ESPRIT algorithm for NLA processing In this section we propose a MI-ESPRIT variant of the SD method for coupled CPD that is suited for shift- invariance structured

Methods for joint compression [37], [38] and scalable decomposition of large 1 Decompositions with shared factors can be viewed as constrained ones [21], [22] and, in a way,

De Lathauwer, Multiple Invariance ESPRIT for Nonuniform Linear Arrays: A Coupled Canonical Polyadic Decomposition Approach, ESAT-STADIUS, KU Leuven,

IV. MI-ESPRIT algorithm for NLA processing In this section we propose a MI-ESPRIT variant of the SD method for coupled CPD that is suited for shift- invariance structured

To alleviate this problem, we present a link between MHR and the coupled CPD model, leading to an improved uniqueness condition tailored for MHR and an algebraic method that can

In section IV we demonstrate the usefulness of coupled tensor decompositions in the context of array signal processing problems involving widely separated antenna arrays with at