scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Tucker factorization with missing data with application to low-$$n$$n-rank tensor completion

01 Jul 2015-Multidimensional Systems and Signal Processing (Springer US)-Vol. 26, Iss: 3, pp 677-692
TL;DR: This paper proposes a simple algorithm for Tucker factorization of a tensor with missing data and its application to low-$$n$$n-rank tensor completion and demonstrates in several numerical experiments that the proposed algorithm performs well even when the ranks are significantly overestimated.
Abstract: The problem of tensor completion arises often in signal processing and machine learning. It consists of recovering a tensor from a subset of its entries. The usual structural assumption on a tensor that makes the problem well posed is that the tensor has low rank in every mode. Several tensor completion methods based on minimization of nuclear norm, which is the closest convex approximation of rank, have been proposed recently, with applications mostly in image inpainting problems. It is often stated in these papers that methods based on Tucker factorization perform poorly when the true ranks are unknown. In this paper, we propose a simple algorithm for Tucker factorization of a tensor with missing data and its application to low-$$n$$n-rank tensor completion. The algorithm is similar to previously proposed method for PARAFAC decomposition with missing data. We demonstrate in several numerical experiments that the proposed algorithm performs well even when the ranks are significantly overestimated. Approximate reconstruction can be obtained when the ranks are underestimated. The algorithm outperforms nuclear norm minimization methods when the fraction of known elements of a tensor is low.

Summary (1 min read)

1 Introduction

  • Low-rank matrix completion problem was studied extensively in recent years (Recht et al 2010; Candes and Recht 2009).
  • In Subsection 1.1, the authors review basics of tensor notation and terminology.
  • The emphasis is on synthetic experiments, which are used to demonstrate the efficiency of the proposed method on exactly low-rank problems.
  • There, the authors introduced an extension of nuclear norm to tensors.
  • Therefore, one approach to tensor completion could be to use one of these algorithms, and an approximation of the complete tensor can be obtained from its Tucker factorization.

2 Proposed approach

  • The authors suppose that they have some estimations of true ranks.
  • Any first-order optimization method can be used for minimizing fW in (15) with respect to one of the parameters.
  • For this reason, the authors have used nonlinear conjugate gradient method, as in (Acar et al 2011), from the Poblano toolbox (Dunlavy et al 2010) for MATLAB.
  • The core and the factors are initialized by HOSVD algorithm (De Lathauwer et al 2000; Kolda and Bader 2009), applied to the initial approximation X̂.
  • The experiments show that an accurate reconstruction of tensor X can be obtained when the true ranks are overestimated (for example, r̂i = ri + 10).

3 Experimental results

  • Several experiments were performed on synthetic and realistic data to demonstrate the efficiency of the proposed method.
  • Maximal number of iterations was set to 5000, although the algorithm stabilized (reached the stationary point) even in much less iterations in all experiments.
  • It can be seen that good quality of reconstruction (better than using the nuclear norm minimization) can be obtained when the true ranks are over- or even underestimated.

4 Conclusions

  • The authors have proposed a Tucker factorization-based approach to low-n-rank tensor completion using similar approach as in (Acar et al 2011), where it was used for PARAFAC decomposition with missing data.
  • It was demonstrated that the proposed method can recover the underlying low-n-rank tensor even when the true tensor ranks are unknown.
  • An important assumption was that the true ranks can be overestimated.
  • Of course, there are no theoretical guarantees for the proposed method (since it is based on non-convex optimization), which is its main flaw.
  • The authors would like to thank the project leader Ivica Kopriva.

Did you find this useful? Give us your feedback

Content maybe subject to copyright    Report

Noname manuscript No.
(will be inserted by the editor)
Tucker factorization with missing data with
application to low-n-rank tensor completion
Marko Filipovi´c · Ante Juki´c
Received: date / Accepted: date
Abstract The problem of tens or completi on arises often in signal p rocessing and
machine learning. It consists of recovering a tenso r from a subset of its entries.
The usual structural assumption on a tensor t h at makes the problem well posed
is that the tensor has low rank in every mode. Several tensor completion methods
based on minimization of nuclear norm, which is the closest convex approximation
of rank, have been proposed recently, with applications mostly in image i n paint-
ing problems. It is often stated in these papers that method s based on Tucker
factorization perform poorly when the true ranks are unknown. In this paper,
we propose a simple algorithm for Tucker factoriz ation of a tensor with missing
data and its application to low-n-ra nk tensor completion . The algorithm is simi lar
to previously proposed metho d for PARAFAC decomposition with mis s ing data.
We demonstrate in several numerical experiments that the pr oposed alg orithm
performs well even when the ranks are significantly overesti mated. Approximate
reconstruction can be obtained when the ranks are underestimated. The algorithm
outperforms nuclear norm minimization methods when the frac tion of known ele-
ments of a t ens or is low.
Keywords Tucker factorization · Tensor co mpletion · Low-n-rank tensor ·
Missing data
Mathematics Subject Classification (2000) MSC 68U99
M. Filipovi´c
Rudjer Boˇskovi´c Institute
Bijeniˇcka 54, 10000 Zagreb, Croatia
Tel.: +385-1-4571241
Fax: +385-1-4680104
E-mail: filipov@irb.hr
A. Juki´c
Department of Medical Physics and Acoustics
University of Oldenburg
26111 Oldenburg, Germany
E-mail: ante.jukic@uni-oldenburg.de

2 Marko Filipovi´c, Ante Juki´c
1 Introduction
Low-rank matrix completion problem was studied extensively in recent years
(Recht et al 2010; Candes and Recht 2009). It arises natur ally in many practi-
cal problems when one would like to recover a matrix from a subset of its entries.
On th e other hand, in ma ny applications one is dealing with multi-way data,
which are n aturally represented by tens ors. Tensor s are multi-dimensional arrays,
i.e. higher-order generalizations of matrices. Multi-way data anal ysis was ori gi-
nally developed in t h e fields of psychometrics and chemometrics, but nowadays
it also has applications in signal processing, machine learning and da ta analysis.
Here, we are interested in the problem of recovering a partially observed tensor, or
tensor completion problem. Examples of applicati ons where the problem arises in-
clude image occlusio n /inpainting problems, social network data analysis , network
traffic data ana lysis, bibliometric data a nalysis, spectroscopy, multidimensional
NMR (Nuclear Magnetic Resonance) data analysis, EEG (electroencephalogram)
data analysis and many others. For a more detailed description of appl ications,
interested reader is referred t o (Acar et al 2011) and references therein.
In the matrix case, it is often realistic to assum e that the matrix that we want
to reconstruct from a subset of its entries has a low rank. This assumption en-
ables matrix completion from only a small number of its entries. However, the
rank function is discrete and nonconvex, which makes its optimization hard in
practice. Therefo re, nuclear norm has been used in many papers as its approxi-
mation. Nuclear norm is defined as the sum of singular values of a matrix, and
it is the tightest convex lower bound of th e rank of a matrix on the set of ma-
trices {Y : kY k
2
1} (here, k · k
2
denotes usual matrix 2-norm). When the rank
is replaced by nuclear norm, the resulting problem of nuclear norm minimization
is convex, and, as shown in (Candes and R echt 2009), if the matrix rank is low
enough, the solution of the o riginal (rank minimizati on) problem can be f ound
by minimizing the nuclear norm. In several recent papers on tensor completion,
the definitio n o f nuclear norm was extended to tensors. There, it was stated that
methods based on Tucker factorization perform poorly when the true ranks of
the tensor are unknown. In this paper, we propose a method for Tucker factor-
izati on wit h missing data, with application in tensor completion. We demonstrate
in several numerical experiments that t h e method performs well even when th e
true ranks are significantly overestimated. Namely, it can estimate the exact ranks
from the data. Also, it outperform s nuclear norm minimization methods when the
fracti on of known elements of a tenso r is low.
The rest of the paper is or ganized as follows. In Subsection 1.1, we review
basics of tensor notation and t erminol ogy. Problem setting and previou s work are
described in Subsection 1.3. We describe our approach in Section 2. In Subsections
2.1 and 2.2, details related to optimization method and implementation of the
algo rithm are described. Several numerical experiments are presented in Section
3. The emphasis is on s ynthetic experiments, which are used to demonstrate the
efficiency of th e proposed method on exactly low-rank problems. However, we also
perform some ex periments o n realistic data. Conclusions are presented in Section
4.

Low-n-rank tensor completion 3
1.1 Tensor notation and terminology
We denote scalars by regular lowercase o r uppercase, vector s by bold lowercase,
matrices by bo ld uppercase, and tensors by bold Euler script letters. For more de-
tail s on tensor notation and terminol ogy, the reader is also referred to
(Kolda and Bader 2009).
The order of a tensor is the number of its dimens ions (also called ways or
modes). We denote the vector space of tensors of order N and size I
1
× · · · × I
N
by R
I
1
×···×I
N
. Elements o f tensor X of order N are denoted by x
i
1
...i
N
.
A fiber of a tensor is defined as a vector obtained by fixing all indices but
one. Fibers a re generaliz ations of m atrix columns and rows. Mode-n fibers are
obtained by fixing all indices but n-th. Mode-n matricization (unfolding) of tensor
X, denoted as X
(n)
, is obta ined by arranging all mode-n fibers as columns of a
matrix. Precise order in which fibers are stacked as columns is not important as
long as it is consistent. Folding is the inverse operation of ma tricization/ un folding.
Mode-n prod uc t of tensor X and matrix A is denoted by X ×
n
A. It is defined
as
Y = X ×
n
A Y
(n)
= AX
(n)
.
Mode-n product is commutative (when applied in distinct modes), i.e.
(X ×
n
A) ×
m
B = (X ×
m
B) ×
n
A
for m 6= n. Repeated mode-n produc t can be expressed as
(X ×
n
A) ×
n
B = X ×
n
(BA) .
There are several definitions of tensor rank. In this pap er, we are interested in
n-rank. For N-way tensor X, n-rank is defined as the rank of X
(n)
. If we denote
r
n
= rank X
(n)
, for n = 1, . . . , N, we say that X is a rank-(r
1
, . . . , r
N
) tensor. In
the experi mental section (Section 3) in this paper we denote an estimation of the
n-rank of given tensor X by ˆr
n
.
For completeness, we al s o st ate the usual definition of the rank of a tensor.
We say tha t an N -way tensor X R
I
1
×···×I
N
is rank-1 if it can be written as the
outer product of N vectors, i.e.
X = a
(1)
· · · a
(N)
, (1)
where denotes the vector outer product. Elementwise, (1) is written as x
i
1
...i
N
=
a
(1)
i
1
· · · a
(N)
i
N
, for all 1 i
n
I
N
. Tensor rank of X is d efin ed as minimal number
of rank-1 tensors that generate X in the sum. As opposed to n-rank of a tensor,
tensor rank is hard to compute (H˚astad 1990).
The Hadamard prod uc t of tensors is the componentwise product, i .e. for N -way
tensors X, Y, it is defined as (X Y)
i
1
...ı
N
= x
i
1
...i
N
y
i
1
...i
N
.
The Frobenius norm of tensor X of size I
1
× · · · × I
N
is denoted by kXk
F
and
defined as
kXk
F
=
I
1
X
i
1
=1
· · ·
I
N
X
i
N
=1
x
2
i
1
...i
N
!
1
2
. (2)

4 Marko Filipovi´c, Ante Juki´c
1.2 Tensor factorizations/decompositions
Two of the most often used tensor factorizations/decom positions a re PARAFAC
(parallel factors) decomposition a n d Tucker factorization. PARAFAC decomposi-
tion is also called canonical decomposition (CANDECOMP) or
CANDECOMP/PARA FAC (CP) decomposition. For a given tensor X R
I
1
×···×I
N
,
it is defined as a decomposition of X as a linear combination of minimal number
of rank-1 tensors
X =
R
X
r=1
λ
r
a
(1)
r
· · · a
(N)
r
. (3)
For more details regarding the PARAFAC deco mposition, the reader is referred to
(Kolda and Bader 2009), since here we are interested in Tucker factorization.
Tucker factoriza tion (also ca lled N -mode PCA or higher-order SVD) of a tensor
X can be written as
X = G ×
1
A
1
×
2
· · · ×
N
A
N
, (4)
where G R
J
1
×···×J
N
is the core tensor with J
i
I
i
, for i = 1, . . . , N , an d A
i
,
i = 1, . . . , N a re, usually orthogonal, factor matrices. Factor matrices A
i
are of
size I
i
× r
i
, for i = 1, . . . , N , if X is rank-(r
1
, . . . , r
N
). A tensor that has low ran k
in every mode can be represented with its Tucker fact orization with small core
tensor (whose dimensi ons correspond to r anks in correspo nd ing modes). Mode-n
matricization X
(n)
of X in (4) can be written as
X
(n)
= A
(n)
G
(n)
A
(N)
· · · A
(n+1)
A
(n1)
· · · A
(1)
T
, (5)
where G
(n)
denotes mode-n matricization of G, denotes Kronecker product of
matrices, and M
T
denotes the transpose of matrix M. If the factor matrices A
(i)
are constrained to be orthogonal, then they can be interpreted as the principa l
components in correspondin g modes, whi le the elem ents of the core tensor G show
the level of interaction between different modes. In general, Tucker factorization is
not unique. However, in practica l applications some constraints are often imposed
on the core and the factors to obtain a meaningful fact orization, for example
orthogonality, non-negativity or sparsity. For m ore details, t h e reader i s referred
to (Tucker 1966; Kolda and Bader 2 009; De La thauwer et al 2000).
1.3 Problem definition and previous work
Let us denote by T R
I
1
×···×I
N
a tensor that is low-rank in every mode (low-
n-rank tensor ), and by T
the projection of T onto indexes of observed entries.
Here, is a subset of {1, . . . , I
1
} × {1, . . . , I
2
} × · · · × {1, . . . , I
N
}, consisting of
positions of obs erved tensor entries. The problem of low-n-rank tensor completion
was formulated in (Gandy et al 2011) as
min
XR
I
1
×···×I
N
N
X
n=1
rank
X
(n)
subject to X
= T
. (6)
Some other functi on o f n-ranks of a tensor can also be considered here, for exam-
ple any linear combination of n-ranks. Nuclear norm mini mization approaches to

Low-n-rank tensor completion 5
tensor completion, described in the following, are based on this type of problem
formulation. Namely, the idea is to replace rank
X
(n)
with nuclear norm of X
(n)
.
This leads to the problem formulation
min
XR
I
1
×···×I
N
N
X
n=1
X
(n)
subject to X
= T
.
Here, f or given matrix X R
d
1
×d
2
, kXk
=
P
min(d
1
, d
2
)
i=1
σ
i
(X), where σ
i
(X)
denote singular values of X, denotes nuclear norm (o r trace norm) of X. Corre-
sp onding unconstrained formulation is
min
XR
I
1
×···×I
N
N
X
n=1
X
(n)
+
λ
2
kX
T
k
2
2
, (7)
where λ > 0 is a regularization parameter. This problem formulation was used in
(Gandy et al 2011). In (Tomioka et al. (2011)
1
), similar formulation
min
XR
I
1
×···×I
N
N
X
n=1
γ
n
X
(n)
+
1
2λ
kX
T
k
2
2
, (8)
where λ > 0 and γ
n
0 are parameters, was used. In (Liu et al 2013), formulation
min
XR
I
1
×···×I
N
N
X
n=1
α
n
X
(n)
+
γ
n
2
X
(n)
T
(n)
2
F
subject to X
= T
,
(9)
where again α
n
0 and γ
n
> 0 are para meters, was used. In above equations (7),
(8) and (9), we have used the notation from corresponding pap ers. Therefore, note
that λ in (7) and (8), i.e. γ
n
in (8) and (9), have different interpretations.
The first paper tha t proposed an extension of low-rank matrix completion
concept to tensors seems to be (Liu et al 2009). There, the authors introduced
an extension of nuclear norm to tensors. They focused on n-rank, and defined
the nuclear norm of tensor X as the average of nuclear norms of its unfoldings.
In subsequent paper (Liu et al 2013), they defined the nuclear norm of a tensor
more generally, as a convex combination of nuclear norms of its unfoldings. Similar
approaches were used in (Gandy et al 2011) and (Tomioka et al. (2011)
1
).
In (Liu et al 20 13), three algorithms were proposed. Simple low rank tensor
complet ion (SiLRTC) is a block coordinate descent method that is guaranteed to
find the optimal solution since the objective is convex. To improve its convergence
sp eed, the authors in (Liu et al 2013) proposed another algorithm: fast low rank
tensor completion (FaLRTC). FaLRTC uses a smoothing scheme to convert the
original nonsmooth pro blem into a sm ooth one. Then, acceleration scheme is used
to improve the convergence speed of the algorithm. Finally, the authors also pro-
posed the highly accurat e low rank tensor completion (HaLRTC), which applies
the alternating direction method of multipliers (ADMM) algorithm to the low
rank tensor completion p roblems. It was shown to be slower than FaLRTC, but
can achieve higher accuracy. Similar algorithm was derived in (Gandy et al 2011).
1
R. Tomioka, K. Hayashi and H. Kashima: Estimation of low-rank tensors via convex
optimization. Technical report, http://arxiv.org/abs/1010.0789.

Citations
More filters
Journal ArticleDOI
TL;DR: It is shown how constrained multiblock tensor decomposition methods are able to extract similar or statistically dependent common features that are shared by all blocks, by incorporating the multiway nature of data.
Abstract: With the increasing availability of various sensor technologies, we now have access to large amounts of multi-block (also called multi-set, multi-relational, or multi-view) data that need to be jointly analyzed to explore their latent connections. Various component analysis methods have played an increasingly important role for the analysis of such coupled data. In this paper, we first provide a brief review of existing matrix-based (two-way) component analysis methods for the joint analysis of such data with a focus on biomedical applications. Then, we discuss their important extensions and generalization to multi-block multiway (tensor) data. We show how constrained multi-block tensor decomposition methods are able to extract similar or statistically dependent common features that are shared by all blocks, by incorporating the multiway nature of data. Special emphasis is given to the flexible common and individual feature analysis of multi-block data with the aim to simultaneously extract common and individual latent components with desired properties and types of diversity. Illustrative examples are given to demonstrate their effectiveness for biomedical data analysis.

153 citations


Cites methods from "Tucker factorization with missing d..."

  • ...Several recently proposed tensor completion methods including the high accuracy low rank tensor completion (HaLRTC) method [103], the generalized higher-order orthogonal iteration (gHOOI) method [122], and the one based on weighted Tucker (WTucker) decomposition [123] were also used for comparisons....

    [...]

Journal ArticleDOI
TL;DR: A modern overview of recent advances in tensor completion algorithms from the perspective of big data analytics characterized by diverse variety, large volume, and high velocity is provided.
Abstract: Tensor completion is a problem of filling the missing or unobserved entries of partially observed tensors. Due to the multidimensional character of tensors in describing complex datasets, tensor completion algorithms and their applications have received wide attention and achievement in areas like data mining, computer vision, signal processing, and neuroscience. In this survey, we provide a modern overview of recent advances in tensor completion algorithms from the perspective of big data analytics characterized by diverse variety, large volume, and high velocity. We characterize these advances from the following four perspectives: general tensor completion algorithms, tensor completion with auxiliary information (variety), scalable tensor completion algorithms (volume), and dynamic tensor completion algorithms (velocity). Further, we identify several tensor completion applications on real-world data-driven problems and present some common experimental frameworks popularized in the literature along with several available software repositories. Our goal is to summarize these popular methods and introduce them to researchers and practitioners for promoting future research and applications. We conclude with a discussion of key challenges and promising research directions in this community for future exploration.

145 citations

Journal ArticleDOI
06 Jan 2016
TL;DR: In this paper, a review of existing matrix-based component analysis methods for the joint analysis of multiblock data with a focus on biomedical applications is provided, and the authors discuss their important extensions and generalization to multi-lock multiway (tensor) data.
Abstract: With the increasing availability of various sensor technologies, we now have access to large amounts of multiblock (also called multiset, multirelational, or multiview) data that need to be jointly analyzed to explore their latent connections. Various component analysis methods have played an increasingly important role for the analysis of such coupled data. In this article, we first provide a brief review of existing matrix-based (two-way) component analysis methods for the joint analysis of such data with a focus on biomedical applications. Then, we discuss their important extensions and generalization to multiblock multiway (tensor) data. We show how constrained multiblock tensor decomposition methods are able to extract similar or statistically dependent common features that are shared by all blocks, by incorporating the multiway nature of data. Special emphasis is given to the flexible common and individual feature analysis of multiblock data with the aim to simultaneously extract common and individual latent components with desired properties and types of diversity. Illustrative examples are given to demonstrate their effectiveness for biomedical data analysis.

137 citations

Proceedings Article
13 Feb 2017
TL;DR: It is argued that low-rank constraint, albeit useful, is not effective enough to exploit the local smooth and piecewise priors of visual data, and proposed integrating total variation into low- rank tensor completion (LRTC) to address the drawback.
Abstract: With the advance of acquisition techniques, plentiful higherorder tensor data sets are built up in a great variety of fields such as computer vision, neuroscience, remote sensing and recommender systems. The real-world tensors often contain missing values, which makes tensor completion become a prerequisite to utilize them. Previous studies have shown that imposing a low-rank constraint on tensor completion produces impressive performances. In this paper, we argue that low-rank constraint, albeit useful, is not effective enough to exploit the local smooth and piecewise priors of visual data. We propose integrating total variation into low-rank tensor completion (LRTC) to address the drawback. As LRTC can be formulated by both tensor unfolding and tensor decomposition, we develop correspondingly two methods, namely LRTC-TV-I and LRTC-TVII, and their iterative solvers. Extensive experimental results on color image and medical image inpainting tasks show the effectiveness and superiority of the two methods against state-of-the-art competitors.

107 citations


Cites background or methods from "Tucker factorization with missing d..."

  • ...2014), while the other line performs CP or Tucker decomposition and try to make the decomposition factors be low-rank (Chen, Hsu, and Liao 2014; Liu et al. 2014a; Filipović and Jukić 2015; Zhao, Zhang, and Cichocki 2015; Liu et al. 2014b)....

    [...]

  • ...Similarly, a weighted Tucker model (WTuker), is introduced for LRTC (Filipović and Jukić 2015)....

    [...]

  • ...Similarly, a weighted Tucker model (WTuker), is introduced for LRTC (Filipović and Jukić 2015)....

    [...]

  • ...…Tomioka, Hayashi, and Kashima 2010; Signoretto et al. 2014), while the other line performs CP or Tucker decomposition and try to make the decomposition factors be low-rank (Chen, Hsu, and Liao 2014; Liu et al. 2014a; Filipović and Jukić 2015; Zhao, Zhang, and Cichocki 2015; Liu et al. 2014b)....

    [...]

Journal ArticleDOI
TL;DR: An overview of the emerging field of massive MIMO localization is provided, which can be used to meet the requirements of 5G, by exploiting different spatial signatures of users.

97 citations

References
More filters
Journal ArticleDOI
TL;DR: This survey provides an overview of higher-order tensor decompositions, their applications, and available software.
Abstract: This survey provides an overview of higher-order tensor decompositions, their applications, and available software. A tensor is a multidimensional or $N$-way array. Decompositions of higher-order tensors (i.e., $N$-way arrays with $N \geq 3$) have applications in psycho-metrics, chemometrics, signal processing, numerical linear algebra, computer vision, numerical analysis, data mining, neuroscience, graph analysis, and elsewhere. Two particular tensor decompositions can be considered to be higher-order extensions of the matrix singular value decomposition: CANDECOMP/PARAFAC (CP) decomposes a tensor as a sum of rank-one tensors, and the Tucker decomposition is a higher-order form of principal component analysis. There are many other tensor decompositions, including INDSCAL, PARAFAC2, CANDELINC, DEDICOM, and PARATUCK2 as well as nonnegative variants of all of the above. The N-way Toolbox, Tensor Toolbox, and Multilinear Engine are examples of software packages for working with tensors.

9,227 citations


"Tucker factorization with missing d..." refers background in this paper

  • ...The first paper that proposed an extension of low-rank matrix completion concept to tensors seems to be (Liu et al 2009)....

    [...]

Proceedings ArticleDOI
07 Jul 2001
TL;DR: In this paper, the authors present a database containing ground truth segmentations produced by humans for images of a wide variety of natural scenes, and define an error measure which quantifies the consistency between segmentations of differing granularities.
Abstract: This paper presents a database containing 'ground truth' segmentations produced by humans for images of a wide variety of natural scenes. We define an error measure which quantifies the consistency between segmentations of differing granularities and find that different human segmentations of the same image are highly consistent. Use of this dataset is demonstrated in two applications: (1) evaluating the performance of segmentation algorithms and (2) measuring probability distributions associated with Gestalt grouping factors as well as statistics of image region properties.

6,505 citations

Journal ArticleDOI
TL;DR: It is proved that one can perfectly recover most low-rank matrices from what appears to be an incomplete set of entries, and that objects other than signals and images can be perfectly reconstructed from very limited information.
Abstract: We consider a problem of considerable practical interest: the recovery of a data matrix from a sampling of its entries. Suppose that we observe m entries selected uniformly at random from a matrix M. Can we complete the matrix and recover the entries that we have not seen? We show that one can perfectly recover most low-rank matrices from what appears to be an incomplete set of entries. We prove that if the number m of sampled entries obeys $$m\ge C\,n^{1.2}r\log n$$ for some positive numerical constant C, then with very high probability, most n×n matrices of rank r can be perfectly recovered by solving a simple convex optimization program. This program finds the matrix with minimum nuclear norm that fits the data. The condition above assumes that the rank is not too large. However, if one replaces the 1.2 exponent with 1.25, then the result holds for all values of the rank. Similar results hold for arbitrary rectangular matrices as well. Our results are connected with the recent literature on compressed sensing, and show that objects other than signals and images can be perfectly reconstructed from very limited information.

5,274 citations


"Tucker factorization with missing d..." refers background in this paper

  • ...Low-rank matrix completion problem was studied extensively in recent years (Recht et al 2010; Candes and Recht 2009)....

    [...]

  • ...When the rank is replaced by nuclear norm, the resulting problem of nuclear norm minimization is convex, and, as shown in (Candes and Recht 2009), if the matrix rank is low enough, the solution of the original (rank minimization) problem can be found by minimizing the nuclear norm....

    [...]

Journal ArticleDOI
TL;DR: There is a strong analogy between several properties of the matrix and the higher-order tensor decomposition; uniqueness, link with the matrix eigenvalue decomposition, first-order perturbation effects, etc., are analyzed.
Abstract: We discuss a multilinear generalization of the singular value decomposition. There is a strong analogy between several properties of the matrix and the higher-order tensor decomposition; uniqueness, link with the matrix eigenvalue decomposition, first-order perturbation effects, etc., are analyzed. We investigate how tensor symmetries affect the decomposition and propose a multilinear generalization of the symmetric eigenvalue decomposition for pair-wise symmetric tensors.

4,101 citations


"Tucker factorization with missing d..." refers background or methods in this paper

  • ...For more details, the reader is referred to (Tucker 1966; Kolda and Bader 2009; De Lathauwer et al 2000)....

    [...]

  • ...The core and the factors are initialized by HOSVD algorithm (De Lathauwer et al 2000; Kolda and Bader 2009), applied to the initial approximation X̂....

    [...]

Journal ArticleDOI
TL;DR: The model for three-mode factor analysis is discussed in terms of newer applications of mathematical processes including a type of matrix process termed the Kronecker product and the definition of combination variables.
Abstract: The model for three-mode factor analysis is discussed in terms of newer applications of mathematical processes including a type of matrix process termed the Kronecker product and the definition of combination variables. Three methods of analysis to a type of extension of principal components analysis are discussed. Methods II and III are applicable to analysis of data collected for a large sample of individuals. An extension of the model is described in which allowance is made for unique variance for each combination variable when the data are collected for a large sample of individuals.

3,810 citations


"Tucker factorization with missing d..." refers background in this paper

  • ...For more details, the reader is referred to (Tucker 1966; Kolda and Bader 2009; De Lathauwer et al 2000)....

    [...]

Frequently Asked Questions (15)
Q1. What are the contributions in "Tucker factorization with missing data with application to low-n-rank tensor completion" ?

In this paper, the authors propose a simple algorithm for Tucker factorization of a tensor with missing data and its application to low-n-rank tensor completion. The authors demonstrate in several numerical experiments that the proposed algorithm performs well even when the ranks are significantly overestimated. 

Since the proposed approach is based on unconstrained optimization, possible extensions include introducing some constraints on the factors in the model, for example orthogonality or non-negativity. 

Nonlinear conjugate gradient method, implemented in the Poblano toolbox (Dunlavy et al 2010), was used for optimization because of its speed. 

Examples of applications where the problem arises include image occlusion/inpainting problems, social network data analysis, network traffic data analysis, bibliometric data analysis, spectroscopy, multidimensional NMR (Nuclear Magnetic Resonance) data analysis, EEG (electroencephalogram) data analysis and many others. 

In the first setting, the size of the tensor was 20× 20× 20× 20× 20, all n-mode ranks were set to 2, and the fraction of known entries of the tensor was 0.2. 

The problem with the approaches that use nuclear norm is their computational complexity, since in every iteration the singular value decomposition (SVD) needs to be computed. 

When the rank is replaced by nuclear norm, the resulting problem of nuclear norm minimization is convex, and, as shown in (Candes and Recht 2009), if the matrix rank is low enough, the solution of the original (rank minimization) problem can be found by minimizing the nuclear norm. 

The objective function used in (Tomasi and Bro 2005; Acar et al 2011) was of the form (for 3-way tensors)fW (A,B,C) =I ∑i=1J ∑j=1K ∑k=1{wijk(xijk −R ∑r=1airbjrckr)}2, (11)where W is a tensor of the same size as X defined aswijk =1 , if xijk is known 0 , if xijk is missing (12)This approach differs from the one taken in (Andersson and Bro 1998). 

The authors say that an N -way tensor X ∈ RI1×···×IN is rank-1 if it can be written as the outer product of N vectors, i.e.X = a(1) ◦ · · · ◦ a(N), (1)where ◦ denotes the vector outer product. 

It can be seen that the proposed method can reconstruct the underlying low-n-rank tensor even for small number of observed entries (for 20 percent or more), smaller than the nuclear norm minimization approach, despite the fact that the ranks were significantly overestimated. 

as demonstrated in numerical experiments in Section 3, the proposed algorithm can estimate the exact n-ranks of a tensor as long as initial approximations of n-ranks are over-estimations of exact ranks. 

Relative error was calculated as ∥∥ ∥ X̂−X∥ ∥ ∥ΩC , F‖X‖ΩC , F (18)where X̂ denotes the output of the algorithm and ‖ · ‖ΩC , F denotes the error calculated only on the set ΩC . 

As another contribution, the authors show that the proposed method performs better than nuclear norm minimization methods when the fraction of known tensor elements is low. 

Since here the authors concentrate mostly on synthetic experiments, where data are generated randomly, a natural question is how confident the reported results are2 MATLAB codesince the authors used at most 100 repetitions (i.e. different random realizations) for given problem setting. 

For this number of iterations, the algorithm took about 22 minutes (the algorithms from (Tomioka et al. (2011)1) took about 42 minutes for 5000 iterations).