Concept Graph & Resume using Claude 3.5 Sonnet | Chat GPT4o | Llama 3:
Resume:
1.- Robustness can mean different things in different contexts. This talk explores robustness in algorithm design for high-dimensional problems.
2.- In 1D robust parameter estimation, the median and median absolute deviation are more robust than the mean and variance.
3.- In high dimensions, a tradeoff emerges between provable robustness guarantees and computational tractability of estimation algorithms.
4.- A general recipe for robust high-dimensional estimation: 1) Find good parameter distance 2) Detect corruptions 3) Create win-win algorithm
5.- For robust mean estimation with covariance=I, use empirical mean but detect corruptions in 2nd moment. Filter and repeat.
6.- For robust covariance estimation, look at 4th moment tensor structure to detect corruptions in empirical covariance.
7.- Can get dimension-independent error of ε*polylog(1/ε) for robustly estimating Gaussian mean and covariance in polynomial time.
8.- Statistical query lower bounds show ε*√log(1/ε) error is needed for computationally efficient robust Gaussian estimation.
9.- With >50% corruptions, can still do robust Gaussian mean estimation by outputting a small list containing a good estimate.
10.- Combining sparsity with robustness assumptions improves sample complexity to logarithmic in dimension for sparse robust mean estimation.
11.- Robust Gaussian estimation algorithms, despite distributional assumptions, can help with exploratory data analysis on real datasets.
12.- The stochastic block model generates graphs with planted community structure based on within-community and across-community edge probabilities.
13.- Many algorithms solve partial recovery in the stochastic block model when a-b^2 > 2(a+b). This was conjectured optimal.
14.- Belief propagation and statistical physics techniques predict the a-b^2 > 2(a+b) threshold in the stochastic block model.
15.- The a-b^2 > 2(a+b) threshold was proven tight for the stochastic block model. Convex relaxations don't match it.
16.- Semi-random models allow "helpful" changes to stochastic block model, e.g. strengthening within-community edges and weakening between-community edges.
17.- Simple algorithms like common neighbor counting fail in semi-random models due to the allowed edge probability changes.
18.- Algorithms that match the a-b^2 > 2(a+b) threshold fail in semi-random models. The information-theoretic threshold increases.
19.- Semi-definite programs still solve partial recovery in semi-random models, sacrificing sharpness but gaining robustness compared to belief propagation.
20.- The broadcast tree model captures local structure of the stochastic block model. It has a sharp recovery threshold.
21.- Kesten-Stigum bound: In the broadcast tree model, root color is recoverable iff a-b^2 > 2(a+b). Optimal algorithm is majority.
22.- In a semi-random broadcast tree model allowing "helpful" changes, the Kesten-Stigum bound and majority reconstruction break down.
23.- Recursive majority still works in the semi-random broadcast tree model, showing a robustness vs sharpness tradeoff.
24.- Studying semi-random models provides insight into which algorithms are overtuned to exactly the distributional assumptions of average-case models.
25.- Semi-random models are a way to "stress test" average-case algorithmic guarantees by adding realistic deviations to the model.
26.- Matrix completion is the problem of recovering a low-rank matrix from sparse observations. Two main approaches: convex relaxation and alternating minimization.
27.- Convex relaxation for matrix completion is more robust to an adversary revealing additional entries. Alternating minimization breaks down.
28.- No single notion of robustness is universally appropriate. Studying different robustness models provides insight into algorithms.
29.- An open question is to get matrix completion algorithms with the computational benefits of alternating minimization and robustness of convex relaxation.
30.- Investigating appropriate notions of robustness for problems in algorithms and machine learning is a promising research direction.
Knowledge Vault built byDavid Vivancos 2024