2,047 research outputs found
Parallel Algorithms for Constrained Tensor Factorization via the Alternating Direction Method of Multipliers
Tensor factorization has proven useful in a wide range of applications, from
sensor array processing to communications, speech and audio signal processing,
and machine learning. With few recent exceptions, all tensor factorization
algorithms were originally developed for centralized, in-memory computation on
a single machine; and the few that break away from this mold do not easily
incorporate practically important constraints, such as nonnegativity. A new
constrained tensor factorization framework is proposed in this paper, building
upon the Alternating Direction method of Multipliers (ADMoM). It is shown that
this simplifies computations, bypassing the need to solve constrained
optimization problems in each iteration; and it naturally leads to distributed
algorithms suitable for parallel implementation on regular high-performance
computing (e.g., mesh) architectures. This opens the door for many emerging big
data-enabled applications. The methodology is exemplified using nonnegativity
as a baseline constraint, but the proposed framework can more-or-less readily
incorporate many other types of constraints. Numerical experiments are very
encouraging, indicating that the ADMoM-based nonnegative tensor factorization
(NTF) has high potential as an alternative to state-of-the-art approaches.Comment: Submitted to the IEEE Transactions on Signal Processin
Kullback-Leibler Principal Component for Tensors is not NP-hard
We study the problem of nonnegative rank-one approximation of a nonnegative
tensor, and show that the globally optimal solution that minimizes the
generalized Kullback-Leibler divergence can be efficiently obtained, i.e., it
is not NP-hard. This result works for arbitrary nonnegative tensors with an
arbitrary number of modes (including two, i.e., matrices). We derive a
closed-form expression for the KL principal component, which is easy to compute
and has an intuitive probabilistic interpretation. For generalized KL
approximation with higher ranks, the problem is for the first time shown to be
equivalent to multinomial latent variable modeling, and an iterative algorithm
is derived that resembles the expectation-maximization algorithm. On the Iris
dataset, we showcase how the derived results help us learn the model in an
\emph{unsupervised} manner, and obtain strikingly close performance to that
from supervised methods.Comment: Asilomar 201
- …
