Algorithms for an Efficient Tensor Biclustering

@inproceedings{Faneva2019AlgorithmsFA,
  title={Algorithms for an Efficient Tensor Biclustering},
  author={Andriantsiory Dina Faneva and Mustapha Lebbah and Hanene Azzag and Ga{\"e}l Beck},
  booktitle={PAKDD},
  year={2019}
}
Consider a data set collected by (individuals-features) pairs in different times. It can be represented as a tensor of three dimensions (Individuals, features and times). The tensor biclustering problem computes a subset of individuals and a subset of features whose signal trajectories over time lie in a low-dimensional subspace, modeling similarity among the signal trajectories while allowing different scalings across different individuals or different features. This approach are based on… 
Multi-Slice Clustering for 3-order Tensor Data
TLDR
This work proposes a new method, namely the multi-slice clustering (MSC) for a 3-order tensor data set, which defines a similarity measure between matrix slices up to a threshold (precision) parameter, and from that, identifies a cluster.

References

SHOWING 1-8 OF 8 REFERENCES
Tensor Biclustering
TLDR
This work proposes an efficient spectral algorithm to solve the tensor biclustering problem and analyzes its achievability bound in an asymptotic regime and shows the efficiency of the proposed method in several synthetic and real datasets.
A statistical model for tensor PCA
TLDR
It turns out that the Principal Component Analysis problem for large tensors of arbitrary order $k$ under a single-spike (or rank-one plus noise) model is possible as soon as the signal-to-noise ratio $\beta$ becomes larger than $C\sqrt{k\log k}$ (and in particular $\beta can remain bounded as the problem dimensions increase).
Tensor Decompositions and Applications
This survey provides an overview of higher-order tensor decompositions, their applications, and available software. A tensor is a multidimensional or $N$-way array. Decompositions of higher-order
Fast spectral algorithms from sum-of-squares proofs: tensor decomposition and planted sparse vectors
TLDR
This work gives an algorithm with running time nearly linear in the input size that approximately recovers a planted sparse vector with up to constant relative sparsity in a random subspace of ℝn of dimension up to Ω(√n).
On the Limitation of Spectral Methods: From the Gaussian Hidden Clique Problem to Rank One Perturbations of Gaussian Tensors
TLDR
A lower bound on the critical signal-to-noise ratio below which a rank-one signal cannot be detected is established on the basis of a more general result on rank- one perturbations of the Gaussian tensors.
Tensor principal component analysis via sum-of-square proofs
TLDR
It is shown that degree-$4 sum-of-squares relaxations break down for $\tau \leq O(n^{3/4}/\log(n)^{1/4})$, which demonstrates that improving the current guarantees would require new techniques or might even be intractable.
Statistical-Computational Tradeoffs in Planted Problems and Submatrix Localization with a Growing Number of Clusters and Submatrices
TLDR
The results establish the minimax recovery limits, which are tight up to universal constants and hold even with a growing number of clusters/submatrices, and provide order-wise stronger performance guarantees for polynomial-time algorithms than previously known.
Computational and Statistical Boundaries for Submatrix Localization in a Large Noisy Matrix
TLDR
An interesting phenomenon is shown that $\sf SNR_c$ is always significantly larger than $sigma/\sigma$, which implies an essential gap between statistical optimality and computational efficiency for submatrix localization.