Understanding the Cluster Linear Program for Correlation Clustering

被引:0
作者
Cao, Nairen [1 ]
Cohen-Addad, Vincent [2 ]
Lee, Euiwoong [3 ]
Li, Shi [4 ]
Newman, Alantha [5 ]
Vogl, Lukas [6 ]
机构
[1] Boston Coll, Brighton, MA 02135 USA
[2] Google Res, Grenoble, France
[3] Univ Michigan, Ann Arbor, MI 48109 USA
[4] Nanjing Univ, Dept Comp Sci & Technol, Nanjing, Peoples R China
[5] Univ Grenoble Alpes, CNRS, Grenoble, France
[6] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
来源
PROCEEDINGS OF THE 56TH ANNUAL ACM SYMPOSIUM ON THEORY OF COMPUTING, STOC 2024 | 2024年
关键词
Clustering; approximation algorithms; exponential size linear programming; semi-definite programming;
D O I
10.1145/3618260.3649749
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In the classic Correlation Clustering problem introduced by Bansal, Blum, and Chawla (FOCS 2002), the input is a complete graph where edges are labeled either + or -, and the goal is to find a partition of the vertices that minimizes the sum of the +edges across parts plus the sum of the -edges within parts. In recent years, Chawla, Makarychev, Schramm and Yaroslavtsev (STOC 2015) gave a 2.06-approximation by providing a near-optimal rounding of the standard LP, and Cohen-Addad, Lee, Li, and Newman (FOCS 2022, 2023) finally bypassed the integrality gap of 2 for this LP giving a 1.73-approximation for the problem. While introducing new ideas for Correlation Clustering, their algorithm is more complicated than typical approximation algorithms in the following two aspects: (1) It is based on two different relaxations with separate rounding algorithms connected by the round-or-cut procedure. (2) Each of the rounding algorithms has to separately handle seemingly inevitable correlated rounding errors, coming from correlated rounding of Sherali-Adams and other strong LP relaxations. In order to create a simple and unified framework for Correlation Clustering similar to those for typical approximate optimization tasks, we propose the cluster LP as a strong linear program that might tightly capture the approximability of Correlation Clustering. It unifies all the previous relaxations for the problem. It is exponential-sized, but we show that it can be (1 + epsilon)-approximately solved in polynomial time for any epsilon > 0, providing the framework for designing rounding algorithms without worrying about correlated rounding errors; these errors are handled uniformly in solving the relaxation.
引用
收藏
页码:1605 / 1616
页数:12
相关论文
共 42 条
  • [1] Agrawal Rakesh, 2009, INT C WEB SEARCH WEB, P172
  • [2] Ahmadian Sara, 2023, PMLR, P9499
  • [3] Ahn KJ, 2015, PR MACH LEARN RES, V37, P2237
  • [4] Aggregating Inconsistent Information: Ranking and Clustering
    Ailon, Nir
    Charikar, Moses
    Newman, Alantha
    [J]. JOURNAL OF THE ACM, 2008, 55 (05)
  • [5] Large-Scale Deduplication with Constraints using Dedupalog
    Arasu, Arvind
    Re, Christopher
    Suciu, Dan
    [J]. ICDE: 2009 IEEE 25TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING, VOLS 1-3, 2009, : 952 - 963
  • [6] Assadi Sepehr, 2022, P 13 C INN THEOR COM, V215
  • [7] Bansal N., 2006, STOC'06. Proceedings of the 38th Annual ACM Symposium on Theory of Computing, P31, DOI 10.1145/1132516.1132522
  • [8] Correlation clustering
    Bansal, N
    Blum, A
    Chawla, S
    [J]. MACHINE LEARNING, 2004, 56 (1-3) : 89 - 113
  • [9] Behnezhad S, 2023, Disc Algorithms, P819
  • [10] Almost 3-Approximate Correlation Clustering in Constant Rounds
    Behnezhad, Soheil
    Charikar, Moses
    Ma, Weiyun
    Tan, Li-Yang
    [J]. 2022 IEEE 63RD ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS), 2022, : 720 - 731