Randomly sparsified Richardson iteration is really fast

Speaker: 

Robert Webber

Institution: 

UCSD

Time: 

Wednesday, November 6, 2024 - 3:00pm to 4:00pm

Location: 

510R Rowland Hall

Recently, a class of algorithms combining classical fixed point iterations with repeated random sparsification of approximate solution vectors has been successfully applied to eigenproblems with matrices as large as 10^{108} x 10^{108}. So far, a complete mathematical explanation for their success has proven elusive. The family of random sparsification methods has not yet been extended to the important case of linear system solves. This talk proposes a new algorithm based on repeated random sparsification that is capable of solving linear systems in extremely high dimensions and provides a complete mathematical analysis of the new algorithm. The analysis establishes a faster-than-Monte Carlo convergence rate and justifies use of the scheme even when the solution vector is too large to store.

Fairness and Foundations in Machine Learning

Speaker: 

Deanna Needell

Institution: 

UCLA

Time: 

Wednesday, November 13, 2024 - 3:00pm to 4:00pm

Location: 

Rowland Hall 510 R

In this talk, we will address areas of recent work centered around the themes of fairness and foundations in machine learning as well as highlight the challenges in this area. We will discuss recent results involving linear algebraic tools for learning, such as methods in non-negative matrix factorization that include tailored approaches for fairness. We will showcase our derived theoretical guarantees as well as practical applications of those approaches.  Then, we will discuss new foundational results that theoretically justify phenomena like benign overfitting in neural networks.  Throughout the talk, we will include example applications from collaborations with community partners, using machine learning to help organizations with fairness and justice goals. 

Introduction to Robust Statistics III

Speaker: 

Pedro Abdalla

Institution: 

UCI

Time: 

Wednesday, October 16, 2024 - 3:00pm to 4:00pm

Host: 

Location: 

RH 510R

Robust Statistics is a classical topic that dates back to the seminal work of Huber in the 1980s. In essence, the main goal of the field is to account for the effect of outliers when performing estimation tasks, such as mean estimation. A recent line of research, inspired by the seminal work of Catoni, has revisited some classical problems in robust statistics from a non-asymptotic perspective. The goal of this short seminar series is to introduce the key ideas related to robust estimation and discuss various notions of robustness, including heavy-tailed distributions and adversarial contamination. The primary example will be the mean estimation problem, and if time permits, I will also cover covariance estimation

Introduction to Robust Statistics II

Speaker: 

Pedro Abdalla

Institution: 

UCI

Time: 

Wednesday, October 9, 2024 - 3:00pm to 4:00pm

Host: 

Location: 

RH 510R

Robust Statistics is a classical topic that dates back to the seminal work of Huber in the 1980s. In essence, the main goal of the field is to account for the effect of outliers when performing estimation tasks, such as mean estimation. A recent line of research, inspired by the seminal work of Catoni, has revisited some classical problems in robust statistics from a non-asymptotic perspective. The goal of this short seminar series is to introduce the key ideas related to robust estimation and discuss various notions of robustness, including heavy-tailed distributions and adversarial contamination. The primary example will be the mean estimation problem, and if time permits, I will also cover covariance estimation

Introduction to Robust Statistics I

Speaker: 

Pedro Abdalla

Institution: 

UCI

Time: 

Wednesday, October 2, 2024 - 3:00pm to 4:00pm

Host: 

Location: 

RH 510R

Robust Statistics is a classical topic that dates back to the seminal work of Huber in the 1980s. In essence, the main goal of the field is to account for the effect of outliers when performing estimation tasks, such as mean estimation. A recent line of research, inspired by the seminal work of Catoni, has revisited some classical problems in robust statistics from a non-asymptotic perspective. The goal of this short seminar series is to introduce the key ideas related to robust estimation and discuss various notions of robustness, including heavy-tailed distributions and adversarial contamination. The primary example will be the mean estimation problem, and if time permits, I will also cover covariance estimation

A new proof of Friedman's second eigenvalue theorem with strong implications

Speaker: 

Jorge Garza Vargas

Institution: 

Caltech

Time: 

Wednesday, June 5, 2024 - 2:00pm

Location: 

510R Rowland Hall

In 2004 J. Friedman wrote a ~100 page paper proving a conjecture of Alon which stated that random d-regular graphs are nearly optimal expanders. Since then, Friedman's result has been refined and generalized in several directions, perhaps most notably by Bordenave and Collins who in 2019 established strong convergence of independent permutation matrices (a massive generalization of Friedman's theorem), a result that led to groundbreaking results in spectral theory and geometry.  

In this talk I will present joint work with C. Chen, J. Tropp and R. van Handel, where we introduce a new proof technique that allows one to convert qualitative results in random matrix theory into quantitative ones. This technique yields a fundamentally new approach to the study of strong convergence which is more flexible and significantly simpler than the existing techniques. Concretely, we're able to obtain  (1) a remarkably short of Friedman's theorem (2) a quantitative version of the result of Bordenave and Collins (3) a proof of strong convergence for arbitrary stable representations of the symmetric group, which constitutes a substantial generalization of the result of Bordenave and Collins. 

Online Ordered Ramsey Numbers

Speaker: 

Dylan King

Institution: 

Caltech

Time: 

Wednesday, May 29, 2024 - 2:00pm to 3:00pm

Host: 

Location: 

510R Rowland Hall

The Ramsey number of graphs G1 and G2, the smallest N so that any red/blue coloring of the N-clique contains either a red G1 or a blue G2, is one of the most studied notions in combinatorics. We study a related process called the online ordered Ramsey game, played between two players, Builder and Painter. Builder has two graphs, G1 and G2, each of which has a specific ordering on its vertices. Builder starts with an edgeless graph on an ordered vertex set (the integers) and attempts to build either an ordered red copy of G1 or an ordered blue copy of G2 by adding one edge at a time. When Builder adds an edge, Painter is required to decide, at the time of creation, whether an edge is red or blue. Ramsey’s Theorem tells us that Builder can eventually win; their objective is to do so using the minimum number of turns, and Painter’s objective is to delay them as long as possible. The *online ordered Ramsey number* of G1 and G2 is the number of turns taken when both players play optimally.

Online ordered Ramsey numbers were introduced by Perez-Gimenez, P. Pralat, and West in 2021. In this talk we will discuss their relation to other types of Ramsey numbers and present some results on the case when at least one of G1,G2 is sparse.

(Joint work with Emily Heath, Grace McCourt, Hannah Sheats, and Justin Wisby)

Random permutations using GEPP

Speaker: 

John Peca-Medlin

Institution: 

University of Arizona

Time: 

Wednesday, May 8, 2024 - 2:00pm to 3:00pm

Location: 

510R Rowland Hall

Gaussian elimination with partial pivoting (GEPP) remains the most used dense linear solver. For a nxn matrix A, GEPP results in the factorization PA = LU where L and U are lower and upper triangular matrices and P is a permutation matrix. If A is a random matrix, then the associated permutation from the P factor is random. When is this a uniform permutation? How many disjoint cycles are in its cycle decomposition (which equivalently answers how many GEPP pivot movements are needed on A)? What is the longest increasing subsequence of this permutation? We will provide some statistical answers to these questions for select random matrix ensembles and transformations. For particular butterfly permutations, we will present full distributional descriptions for these particular statistics. Moreover, we introduce a random butterfly matrix ensemble that induces the Haar measure on the full 2-Sylow subgroup of the symmetric group on a set of size 2ⁿ.

MCMC, variational inference, and reverse diffusion Monte Carlo

Speaker: 

Yian Ma

Institution: 

UCSD

Time: 

Wednesday, May 1, 2024 - 2:00pm to 3:00pm

Location: 

510R Rowland Hall

I will introduce some recent progress towards understanding the scalability of Markov chain Monte Carlo (MCMC) methods and their comparative advantage with respect to variational inference. I will fact-check the folklore that "variational inference is fast but biased, MCMC is unbiased but slow". I will then discuss a combination of the two via reverse diffusion, which holds promise of solving some of the multi-modal problems. This talk will be motivated by the need for Bayesian computation in reinforcement learning problems as well as the differential privacy requirements that we face.

 

Online differential privacy

Speaker: 

Yiyun He

Institution: 

UCI

Time: 

Wednesday, April 10, 2024 - 2:00pm to 3:00pm

Location: 

510R Rowland Hall

We present a polynomial-time algorithm for online differentially private synthetic data generation. For a data stream within the hypercube [0,1]^d and an infinite time horizon, we develop an online algorithm that generates a differentially private synthetic dataset at each time t. This algorithm achieves a near-optimal accuracy bound in the 1-Wasserstein distance.

Pages

Subscribe to RSS - Combinatorics and Probability