• Corpus ID: 235826011

Local Correlation Clustering with Asymmetric Classification Errors

@article{Jafarov2021LocalCC,
  title={Local Correlation Clustering with Asymmetric Classification Errors},
  author={Jafar Jafarov and Sanchit Kalhan and Konstantin Makarychev and Yury Makarychev},
  journal={ArXiv},
  year={2021},
  volume={abs/2108.05697}
}
In the Correlation Clustering problem, we are given a complete weighted graph G with its edges labeled as “similar” and “dissimilar” by a noisy binary classifier. For a clustering C of graph G, a similar edge is in disagreement with C, if its endpoints belong to distinct clusters; and a dissimilar edge is in disagreement with C if its endpoints belong to the same cluster. The disagreements vector, dis, is a vector indexed by the vertices of G such that the v-th coordinate disv equals the weight… 

Figures from this paper

Robust Correlation Clustering with Asymmetric Noise

TLDR
It is demonstrated that l2-norm-diag recovers nodes with sufficiently strong cluster membership in graph instances generated by the NFM, thereby making progress towards establishing the provable robustness of the proposed algorithm.

THE UNIVERSITY OF CHICAGO FOUR ALGORITHMS FOR CORRELATION CLUSTERING: A SURVEY A DISSERTATION SUBMITTED TO THE FACULTY OF THE DIVISION OF THE PHYSICAL SCIENCES IN CANDIDACY FOR THE DEGREE OF MASTERS

  • Computer Science
  • 2020
TLDR
This exposition focuses on the case when G is complete and unweighted, and explores four approximation algorithms for the Correlation Clustering problem under this assumption.

Correlation Clustering with Sherali-Adams

TLDR
This paper shows that there exists a (1 . 994+ ε )-approximation algorithm based on O (1 /ε 2 ) rounds of the Sherali-Adams hierarchy and reaches an approximation ratio of 2 + ε for Correlation Clustering.

Sublinear Time and Space Algorithms for Correlation Clustering via Sparse-Dense Decompositions

TLDR
A new approach for solving (minimum disagreement) correlation clustering that results in sublinear algorithms with highly efficient time and space complexity for this problem is presented, with a novel connection to sparse-dense graph decompositions that are used extensively in the graph coloring literature.

Correlation Clustering via Strong Triadic Closure Labeling: Fast Approximation Algorithms and Practical Lower Bounds

TLDR
This work presents faster approximation algorithms that avoid linear programming relaxations, for two well-studied special cases: cluster editing and cluster deletion, by draw-ing new connections to edge labeling problems related to the principle of strong triadic closure.

References

SHOWING 1-10 OF 20 REFERENCES

Correlation Clustering with Asymmetric Classification Errors

TLDR
This work gives a (3 + 2 loge(1/α) approximation algorithm for the correlation clustering problem and shows an asymptotically matching Linear Programming integrality gap of Ω(log 1/α).

Correlation clustering with local objectives

TLDR
This paper presents the first known algorithm for minimizing the \ell_q norm of the disagreements vector on arbitrary graphs and provides an improved algorithm for minimize the\ell-q norm (q >= 1) of the disagreement vector on complete graphs.

Clustering with qualitative information

TLDR
This work considers the problem of clustering a collection of elements based on pairwise judgments of similarity and dissimilarity, and gives a factor 4 approximation for minimization on complete graphs, and a factor O(log n) approximation for general graphs.

Min-Max Correlation Clustering via MultiCut

TLDR
The goal is to obtain a partitioning of the vertices that minimizes disagreements – weight of negative edges trapped inside a cluster plus positive edges between different clusters.

Local Guarantees in Graph Cuts and Clustering

TLDR
This work presents an \(O(\sqrt{n})\)-approximation for the problem of minimizing the maximum total weight of disagreement edges incident on any node, providing the first known approximation for the above family of min-max graph cut problems.

Correlation clustering: from theory to practice

TLDR
This tutorial is to show how correlation clustering can be a powerful addition to the toolkit of the data mining researcher and practitioner, and to encourage discussions and further research in the area.

Correlation Clustering and Biclustering With Locally Bounded Errors

TLDR
This work provides a rounding algorithm which converts “fractional clusterings” into discrete clusterings while causing only a constant-factor blowup in the number of errors at each vertex.

Deterministic pivoting algorithms for constrained ranking and clustering problems

TLDR
D deterministic algorithms for constrained weighted feedback arc set in tournaments, constrained correlation clustering, and constrained hierarchical clustering related to finding good ultrametrics and a combinatorial algorithm that improves on the best known factor given by deterministic combinatorsial algorithms for the unconstrained case are given.

Aggregating inconsistent information: Ranking and clustering

TLDR
This work almost settles a long-standing conjecture of Bang-Jensen and Thomassen and shows that unless NP⊆BPP, there is no polynomial time algorithm for the problem of minimum feedback arc set in tournaments.