Structure of tight (k,0)-stable graphs

Speaker: 

Dingding Dong

Institution: 

Caltech

Time: 

Monday, May 4, 2026 - 2:00pm to 3:00pm

Host: 

Location: 

340P Rowland Hall

We say that a graph G is (k,l)-stable if removing k vertices from it reduces its independence number by at most l. We say that G is tight (k,l)-stable if it is (k,l)-stable and its independence number equals ⌊(n−k+1)/2⌋+l, the maximum possible, where n is the vertex number of G. Answering a question of Dong and Wu, we show that every tight (2,0)-stable graph with odd vertex number must be an odd cycle. Moreover, we show that for all k≥3, every tight (k,0)-stable graph has at most k+6 vertices. This is joint work with Sammy Luo.

Statistics in Card Shuffling

Speaker: 

Alexander Clay

Institution: 

USC

Time: 

Monday, May 11, 2026 - 2:00pm to 3:00pm

Location: 

340P Rowland Hall

How many times does one need to shuffle a deck of cards to randomize them? This question has been throughly studied for many types of card shuffling, with connections to Markov chains, representation theory, and group theory. In this talk, we will discuss a weaker version of this topic. What if we only need certain “features” of the deck to be close to uniformly random? This is typically considered through the lens of permutation statistics. We will mention some corresponding results on riffle shuffles by many authors. Then we will give an overview of some of our recent work in this area on random-to-top (and related) shuffling.

How large are sums of large sets?

Speaker: 

Antoine Song

Institution: 

Caltech

Time: 

Monday, April 6, 2026 - 2:00pm to 3:00pm

Location: 

340P Rowland Hall

I will discuss a circle of questions about the geometry of sums of large sets in R^n endowed with the Gaussian measure, related to the convexity problem of Talagrand: if a subset A of R^n has Gaussian measure at least 2/3 say, for which integer q does A+...+A (q times) contain a large convex body? I will explain why it is closely connected to the problem of understanding sums of non-independent Gaussian random vectors. I will discuss some sharp results and partial answers.

A Statistical Framework for Multidimensional Scaling From Noisy Data

Speaker: 

Siddharth Vishwanath

Institution: 

UC San Diego

Time: 

Monday, March 9, 2026 - 2:00pm to 3:00pm

Location: 

340P Rowland Hall

Multidimensional scaling (MDS) has a long history in statistics and underpins a broad class of unsupervised learning and spectral/nonlinear dimension reduction techniques. The objective of MDS is to extract meaningful information from relational data (e.g., distances between sensors, correlations between brain regions, or disagreement scores between individuals) by embedding these relationships into a Euclidean space. In practice, the observed relational information is often subject to measurement errors and/or corrupted by noise. However, the resulting embeddings are typically interpreted as exploratory visualizations without accounting for these variations. This talk presents recent work developing a principled statistical framework for MDS. We show that the classical MDS algorithm achieves minimax-optimal performance across a wide range of noise models and loss functions. Building on this, we develop a framework for constructing valid confidence sets for the embedded points obtained via MDS, enabling formal uncertainty quantification for geometric structure inferred from noisy relational data. These results provide a theoretical foundation for interpreting MDS embeddings, and extend naturally to a wide range of unsupervised learning techniques in modern data science.

Data-driven Breakpoint detection with the QML estimator

Speaker: 

Georg Menz

Institution: 

UCLA

Time: 

Monday, February 23, 2026 - 2:00pm to 3:00pm

Location: 

340P Rowland Hall

 We study quasi-maximum likelihood (QML) breakpoint estimation for
 covariance regime switches in multivariate time series. We move beyond
 the classical framework and show that regime switches can be detected
 as soon as the signal-to-noise ratio is high enough. We identify a
 quantitative global recovery threshold that compares signal separation
 between regimes to signal fluctuations within regimes, and show its
 sharpness via an explicit counterexample.  We also discuss further
 developments in breakpoint detection in high dimensions.

 Joint work with Hubeyb Gurdogan (UCLA)

Singular values and singular vectors of sparse random rectangular matrices at criticality

Speaker: 

Haixiao Wang

Institution: 

University of Wisconsin Madison

Time: 

Monday, April 20, 2026 - 2:00pm to 3:00pm

Location: 

340P Rowland Hall

In modern machine learning applications, data matrices are always assumed to admit the signal-plus-noise structure. Typically, we assume that the spectra of signal and noise matrices are well-separated and that noise subspaces only produce a marginal influence. While these assumptions are readily verified for dense matrices via classical random matrix theory, real-world data is often sparse, posing significant theoretical challenges. We investigate the spectral properties of a matrix $X \in \mathbb{R}^{n \times m}$ with i.i.d. $\text{Bernoulli}(p)$ entries. Previous literature proved that when $np \gg \log(n)$, the singular values of $X$ almost surely remain within the compact support of the Marčenko-Pastur (MP) distribution. However, we identify a critical sparsity regime $p = b \log(n) / \sqrt{mn}$ where this classical result fails. We provide a quantitative characterization of the emergence of outlier singular values. For explicit thresholds $b_*$ and $b^*$ as functions of the aspect ratio $\gamma = n/m \ge 1$, we prove a three-phase transition: (1) for $b > b_*$, no outliers exist; (2) for $b^* < b < b_*$, outliers emerge only beyond the right edge of the MP law; and (3) for $b < b^*$, outliers appear on both sides of the bulk, all with high probability. The locations of those outliers are precisely determined by the largest and smallest degree vertices of the underlying random graph. Besides, behavior of singular vectors corresponding to bulk and edge singular values can be characterized precisely. Our approach follows the framework established by Alt, Ducatez, and Knowles (2021), which can be extended to sparse matrices with general bounded entries.

Recovery limits for geometric planted matchings beyond Gaussian assumptions

Speaker: 

Lucas R. Schwengber

Institution: 

UC Berkeley

Time: 

Monday, March 2, 2026 - 2:00pm to 3:00pm

Location: 

340P Rowland Hall

We consider the problem of recovering an unknown planted matching between a set of $n$ randomly placed points in \mathbb{R}^d and random perturbations of these points. Some recent works have established results for the error rates for this problem under Gaussian assumptions for both initial positions and noise at different scaling regimes of sample size and dimension. I will discuss some recent progress on establishing results which hold under general distributional assumption for both the the initial positions and noise. More precisely, I will show a general recipe to establish lower bounds via showing the existence of large matchings in random geometric graphs, which leads to simplified and generalized proofs of previous results. Time allowing I will also make some remarks regarding sufficient conditions for perfect recovery in high-dimensions for models where the noise is not isotropic. This is joint work with Roberto Oliveira.

Pages

Subscribe to RSS - Combinatorics and Probability