MCMC, variational inference, and reverse diffusion Monte Carlo

Speaker: 

Yian Ma

Institution: 

UCSD

Time: 

Wednesday, May 1, 2024 - 2:00pm to 3:00pm

Location: 

510R Rowland Hall

I will introduce some recent progress towards understanding the scalability of Markov chain Monte Carlo (MCMC) methods and their comparative advantage with respect to variational inference. I will fact-check the folklore that "variational inference is fast but biased, MCMC is unbiased but slow". I will then discuss a combination of the two via reverse diffusion, which holds promise of solving some of the multi-modal problems. This talk will be motivated by the need for Bayesian computation in reinforcement learning problems as well as the differential privacy requirements that we face.

 

Online differential privacy

Speaker: 

Yiyun He

Institution: 

UCI

Time: 

Wednesday, April 10, 2024 - 2:00pm to 3:00pm

Location: 

510R Rowland Hall

We present a polynomial-time algorithm for online differentially private synthetic data generation. For a data stream within the hypercube [0,1]^d and an infinite time horizon, we develop an online algorithm that generates a differentially private synthetic dataset at each time t. This algorithm achieves a near-optimal accuracy bound in the 1-Wasserstein distance.

Extreme Eigenvalues of a Random Laplacian Matrix

Speaker: 

Kyle Luh

Institution: 

CU Boulder

Time: 

Wednesday, April 17, 2024 - 2:00pm

Host: 

Location: 

510R Rowland Hall

The extreme eigenvalues of a random matrix have been important objects of study since the inception of random matrix theory and also have a variety of applications.  The Laplacian matrix is the workhorse of spectral graph theory and is the key player in many practical algorithms for graph clustering, network control theory and combinatorial optimization.  In this talk, we discuss the fluctuations of the extreme eigenvalues of a random Laplacian matrix with gaussian entries.  The proof relies on a broad set of techniques from random matrix theory and free probability.  We will also describe some recent progress on a broader class of random Laplacian matrices.

This is joint work with Andrew Campbell and Sean O'Rourke.

Concentration Inequalities and Moment Bounds for Self-Adjoint Operators with Heavy tails

Speaker: 

Stanislav Minsker

Institution: 

USC

Time: 

Wednesday, May 15, 2024 - 2:00pm

Location: 

510R Rowland Hall

We present Fuk-Nagaev - type inequality for the sums of independent self-adjoint operators. This bound could be viewed as an extension of the well known “Matrix Bernstein” inequality to the case of operators with heavy-tailed norms. As a corollary, we deduce Rosenthal moment inequality that improves upon the previously known versions even in the scalar case. Finally, we will discuss applications of these bounds to the covariance estimation problem. 

Measurable tilings

Speaker: 

Jan Grebik

Institution: 

UCLA

Time: 

Wednesday, April 3, 2024 - 2:00pm to 3:00pm

Host: 

Location: 

510R Rowland Hall

Let $(X,\mu)$ be a standard probability space and $G\curvearrowright (X,\mu)$ be a measure-preserving action of a group $G$ on $X$. The general problem that we consider is to understand the structure of measurable tilings $F\oplus A=X$ of $X$ by a measurable tile $A\subseteq X$ shifted by a finite set $F\subseteq G$, thus the shifts $f\cdot A$, $f\in F$ partition $X$ up to null sets. The motivation comes from the theory of (paradoxical) equidecompositions and tilings in $\mathbb{R}^n$. After a summary of recent results that concern the spheres and tori, I will focus on the intersection of these cases, that is, the case of the circle. Using the structure theorem of Greenfeld and Tao for tilings of $\mathbb{Z}^d$, we show that measurable tilings of the circle can be reduced to tilings of finite cyclic groups.

This is a joint work with Conley and Pikhurko, and Greenfeld, Rozhon and Tao.

A quantum algorithm for learning a hidden graph of bounded degree

Speaker: 

Liam Hardiman

Institution: 

UCI

Time: 

Wednesday, February 14, 2024 - 2:00pm

Location: 

510R Rowland Hall

We are presented with a graph, $G$, on $n$ vertices and $m$ edges whose edge set is unknown. Our goal is to learn the edges of $G$ with as few queries to an oracle as possible. When we submit a set $S$ of vertices to the oracle, it tells us whether or not $S$ induces at least one edge in $G$. This so-called OR-query model has been well studied, with Angluin and Chen giving an upper bound on the number of queries needed of $O(m \log n)$ for a general graph $G$ with $m$ edges.
   
When we allow ourselves to make *quantum* queries (we may query subsets in superposition), then we can achieve speedups over the best possible classical algorithms. In the case where $G$ has maximum degree $d$ and is $O(1)$-colorable, Montanaro and Shao presented an algorithm that learns the edges of $G$ in at most $\tilde{O}( d^2 m^{3/4} )$ quantum queries. This gives an upper bound of $\tilde{O}( m^{3/4} )$ quantum queries when $G$ is a matching or a Hamiltonian cycle, which is significantly larger than the lower bound of $\Omega( \sqrt{m} )$ queries given by Ambainis and Montanaro.

We improve on the work of Montanaro and Shao in the case where $G$ has bounded degree. In particular, we present a randomized algorithm that, with high probability, learns cycles and matchings in $\tilde{O}( \sqrt{m} )$ quantum queries, matching the theoretical lower bound up to logarithmic factors.

Doubly Noisy Linear Systems and the Kaczmarz Algorithm

Speaker: 

Anna Ma

Institution: 

UCI

Time: 

Wednesday, April 24, 2024 - 2:00pm

Location: 

510R Rowland Hall

Large-scale linear systems, Ax=b, frequently arise in data science and scientific computing at massive scales, thus demanding effective iterative methods to solve them. Often, these systems are noisy due to operational errors or faulty data-collection processes. In the past decade, the randomized Kaczmarz algorithm (RK) was studied extensively as an efficient iterative solver for such systems. However, the convergence study of RK in the noisy regime is limited and considers measurement noise in the right-hand side vector, b. Unfortunately, that is not always the case, and the coefficient matrix A can also be noisy. In this talk, we motivate and discuss doubly noise linear systems and the performance of the Kaczmarz algorithm applied to such systems. 

Approximately Hadamard matrices and random frames

Speaker: 

Mark Rudelson

Institution: 

University of Michigan

Time: 

Wednesday, February 28, 2024 - 2:00pm

Host: 

Location: 

510R Rowland Hall

We will discuss a problem concerning random frames which arises in signal processing. A frame is an overcomplete set of vectors in the n-dimensional linear space which allows a robust decomposition of any vector in this space as a linear combination of these vectors. Random frames are used in signal processing as a means of encoding since the loss of a fraction of coordinates does not prevent the recovery. We will discuss a question when a random frame contains a copy of a nice (almost orthogonal) basis.

Despite the probabilistic nature of this problem it reduces to a completely deterministic question of existence of approximately Hadamard matrices.  An n by n matrix with plus-minus 1 entries is called Hadamard if it acts on the space as a scaled isometry. Such matrices exist in some, but not in all dimensions. Nevertheless, we will construct plus-minus 1 matrices of every size which act as approximate scaled isometries. This construction will bring us back to probability as we will have to combine number-theoretic and probabilistic methods.

Joint work with Xiaoyu Dong.

Variation of no-three-in-line problem

Speaker: 

Ji Zeng

Institution: 

UCSD

Time: 

Wednesday, March 6, 2024 - 2:00pm to 3:00pm

Host: 

Location: 

510R Rowland Hall

The famous no-three-in-line problem by Dudeney more than a century ago asks whether one can select $2n$ points from the grid $[n]^2$ such that no three are collinear. We present two results related to this problem. First, we give a non-trivial upper bound for the maximum size of a set in $[n]^4$ such that no four are coplanar. Second, we characterize the behavior of the maximum size of a subset such that no three are collinear in a random set of $\mathbb{F}_q^2$, that is, the plane over the finite field of order $q$. We discuss their proofs and related open problems.

A Theory of Neuronal Synaptic Balance

Speaker: 

Pierre Baldi

Institution: 

UCI

Time: 

Wednesday, February 7, 2024 - 2:00pm

Host: 

Location: 

510R Rowland Hall

AI today can pass the Turing test and is in the process of transforming science, technology, society, humans, and beyond.
Surprisingly, modern AI is built out of two very simple and old ideas, rebranded as deep learning: neural networks and
gradient descent learning. When a typical feed-forward neural network is trained by gradient descent, with an L2 regularizer
to avoid overly large synaptic weights, a strange phenomenon occurs: at the optimum, each neuron becomes "balanced"
in the sense that the L2 norm of its incoming synaptic weights becomes equal to the L2 norm of its outgoing synaptic weights. We develop a theory that explains this phenomenon and exposes its generality. Balance emerges with a variety of activation functions, a variety of regularizers including all Lp regularizers, and a variety of networks including recurrent networks. A simple local balancing algorithm can be applied to any neuron and at any time, instead of just at the optimum. Most remarkably, stochastic iterated application of the local balancing algorithm always converges to a unique, globally balanced, state.

Pages

Subscribe to RSS - Combinatorics and Probability