Date | Title | Speaker | Links |
---|---|---|---|
9/6 | Sparsification through Codes | Aaron (Louie) Putterman | Abstract |
9/20 | Sparsification for Maximum Matching: EDCS and Robust Communication Complexity | Amir Azarmehr | Abstract |
The seminal work of Karger and Benczur and Karger from the
90s showed that graphs can be "sparsified" to a number of edges nearly
linear in the number of vertices, so that every cut is preserved to
within multiplicative factors close to 1. In the years since, graph
(and hypergraph) sparsification has emerged as a central tool in the
algorithmic literature, with numerous applications to streaming, fully
dynamic algorithms, and more. The utility of this idea frames the
natural question of which other combinatorial objects are amenable to
sparsification. In this talk, I will provide a partial answer through
the recently introduced notion of "code sparsification" which provides
a unified framework to construct graph / hypergraph sparsifiers, while
also yielding new sparsifications for CSPs and Cayley graphs.
Specifically, the given structure to be preserved is a linear error
correcting code; the goal is to select a re-weighted subset of the
coordinates of the code such that the Hamming weight of every codeword
is preserved within multiplicative factors close to 1. We show that
this class of problems gives the right language to abstract the
techniques of Karger and Benczur and Karger --- and indeed all codes
can be sparsified to length nearly linear in the number of message
bits. This generalization already resolves some basic questions in CSP
sparsification. A further generalization to additive codes over finite
abelian groups gives even more powerful results and in particular
completely classifies the class of symmetric Boolean CSPs that allow
nearly linear sized sparsification. Finally, by developing a linear
algebraic analog of spanning trees, all of these sparsifiers become
efficiently constructible.
Based on joint works with Sanjeev Khanna (Penn) and Madhu Sudan (Harvard).
With massive graphs becoming increasingly prevalent, there has been growing interest in sparsifiers which preserve certain properties of graphs while having a small number of edges. The Edge Degree Constrained Subgraph (EDCS), introduced by Bernstein and Stein (2015, 2016), is such a sparsifier that approximately preserves the size of the maximum matching. EDCS is extremely robust as it remains a good sparsifier under various operations. Consequently, it has found applications in a myriad of different settings including the dynamic, streaming, and sublinear time settings, as well as in the communication model, stochastic matching, and massively parallel computations (MPC).
EDCS is often used to obtain 2/3-approximate matchings. However, we showed in a recent work that it is powerful enough to obtain a much better 5/6-approximation in the robust communication model, significantly improving over prior bounds. The talk will present this result while also covering the basic properties of EDCS.
Based on joint works with Behnezhad (2023), and Behnezhad and Roghani (2024).