Jon A. Wellner Lecture Series
The Jon A. Wellner Lecture was established by Jon and Vera Wellner to provide educational experiences outside the classroom for students and faculty. It brings distinguished statisticians to University of Idaho to share advances in mathematics and statistics while honoring the legacy of Professor Jon A. Wellner.
Upcoming lecture
2026 Jon A. Wellner Lecture
Carl Bergstrom, Ph.D., Professor of Biology, University of Washington
2:30 p.m. on April 13
About the speaker
Carl T. Bergstrom, Ph.D., is a Professor of Biology at University of Washington and a member of the External Faculty at the Santa Fe Institute.
The unifying theme running throughout his work is the concept of information. Within biology, he studies how communication evolves and how the process of evolution encodes information in genomes. In the philosophy and sociology of science, he examines how norms and institutions influence scholars’ research strategies and, in turn, our scientific understanding of the world. Within informatics, he investigates how citations and other traces of scholarly activity can be used to better navigate the overwhelming volume of scholarly literature.
Lately, Dr. Bergstrom has become concerned with the spread of disinformation on social networks and is exploring what can be done to address it.
About Jon A. Wellner
Professor Wellner earned his B.S. in Mathematics at University of Idaho and his Ph.D. in Statistics at University of Washington. He served on the faculty of University of Rochester from 1975 to 1983, and in 1983 joined the faculty of University of Washington, where he is Professor of Statistics and Biostatistics. His research interests include uses of large sample theory in statistics.
Professor Wellner is a fellow of the Institute of Mathematical Statistics, a fellow of the American Statistical Association and a fellow of the American Association for the Advancement of Science. In 2010, he was named a Knight of the Order of the Netherlands Lion in recognition of his exceptional research in probability and statistics, much of which has been done in collaboration with Dutch colleagues. He also received a John Simon Guggenheim Foundation Fellowship (1987-1988) and an Alexander von Humboldt Foundation Fellowship (1980-1981). He was inducted into the College of Science Academy of Distinguished Alumni in 2014.
Recent lectures
The Jon A. Wellner Lecture Series has hosted many distinguished statisticians and researchers over the years. Explore past lectures and learn more about their contributions below.
October, 2024
Annie Qu, Ph.D. | Data Integration for Heterogeneous Data
Abstract
In this presentation, I will showcase advanced statistical machine learning techniques and tools designed for the seamless integration of information from multi-source datasets. These datasets may originate from various sources, encompass distinct studies with different variables, and exhibit unique dependent structures. One of the greatest challenges in investigating research findings is the systematic heterogeneity across individuals, which could significantly undermine the power of existing machine learning methods to identify the underlying true signals. This talk will investigate the advantages and drawbacks of current methods such as multi-task learning, optimal transport, missing data imputations, matrix completions and transfer learning. Additionally, we will introduce a new latent representation method aimed at mapping heterogeneous observed data to a latent space, facilitating the extraction of shared information and knowledge, and disentanglement of source-specific information and knowledge. The key idea of the proposal is to project heterogeneous raw observations to the representation retriever library, and the novelty of our method is that we can retrieve partial representations from the library for a target study. The main advantages of the proposed method are that it can increase statistical power through borrowing common representation retrievers from multiple sources of data. This approach ultimately allows one to extract information from heterogeneous data sources and transfer generalizable knowledge beyond observed data and enhance the accuracy of prediction and statistical inference.
About the speaker
Annie Qu is Chancellor's Professor, Department of Statistics, University of California, Irvine. She received her Ph.D. in Statistics from the Pennsylvania State University in 1998. Qu's research focuses on solving fundamental issues regarding structured and unstructured large-scale data and developing cutting-edge statistical methods and theory in machine learning and algorithms for personalized medicine, text mining, recommender systems, medical imaging data and network data analysis for complex heterogeneous data. The newly developed methods can extract essential and relevant information from large volumes of intensively collected data, such as mobile health data. Her research impacts many fields, including biomedical studies, genomic research, public health research, social and political sciences. Before joining UC Irvine, Dr. Qu was a Data Science Founder Professor of Statistics and the Director of the Illinois Statistics Office at the University of Illinois at Urbana-Champaign. She was awarded the Brad and Karen Smith Professorial Scholar by the College of LAS at UIUC and was a recipient of the NSF Career award from 2004 to 2009. She is a Fellow of the Institute of Mathematical Statistics (IMS), the American Statistical Association and the American Association for the Advancement of Science. She is also a recipient of IMS Medallion Award and Lecturer in 2024. She serves as Journal of the American Statistical Association Theory and Methods Co-Editor from 2023 to 2025 and as IMS Program Secretary from 2021 to 2027.
January, 2024
Daniela Witten, Ph.D. | Data thinning and its applications
Abstract
We propose data thinning, a new approach for splitting an observation from a known distributional family with unknown parameter(s) into two or more independent parts that sum to yield the original observation, and that follow the same distribution as the original observation, up to a (known) scaling of a parameter. This proposal is very general, and can be applied to a broad class of distributions within he natural exponential family, including the Gaussian, Poisson, negative binomial, Gamma and binomial distributions, among others. Furthermore, we generalize data thinning to enable splitting an observation into two or more parts that can be combined to yield the original observation using an operation other than addition; this enables the application of data thinning far beyond the natural exponential family. Data thinning has a number of applications to model selection, evaluation and inference. For instance, cross-validation via data thinning provides an attractive alternative to the “usual” approach of cross-validation via sample splitting, especially in unsupervised settings in which the latter is not applicable. We will present an application of data thinning to single-cell RNA-sequencing data, in a setting where sample splitting is not applicable. This is joint work with Anna Neufeld (Fred Hutch), Ameer Dharamshi (University of Washington), Lucy Gao (University of British Columbia) and Jacob Bien (University of Southern California).
About the speaker
Daniela Witten is a professor of Statistics and Biostatistics at University of Washington, and the Dorothy Gilford Endowed Chair in Mathematical Statistics. She develops statistical machine learning methods for high-dimensional data, with a focus on unsupervised learning.
She has received a number of awards for her research in statistical machine learning: most notably the Spiegelman Award from the American Public Health Association for a (bio)statistician under age 40, and the Presidents' Award from the Committee of Presidents of Statistical Societies for a statistician under age 41.
Daniela is a co-author of the textbook "Introduction to Statistical Learning", and beginning in 2023 will serve as Joint Editor of Journal of the Royal Statistical Society, Series B.
September, 2022
Susan A. Murphy, Ph.D. | Inference for Longitudinal Data After Adaptive Sampling
Abstract
Stochastic epidemic models describe how infectious diseases spread through a population of interest. These models are constructed by first assigning individuals to compartments (e.g., susceptible, infectious and recovered) and then defining a stochastic process that governs the evolution of sizes of these compartments through time. I will review multiple lines of attack of a challenging and not fully solved problem of fitting these models to noisy infectious disease surveillance data. These solutions involve a range of mathematical techniques: particle filter Markov chain Monte Carlo algorithms, approximations of stochastic differential equations and Poisson random measure-based Bayesian data augmentation. Importantly, many of these computational strategies open the door for integration of multiple infectious disease surveillance data streams, including less conventional ones (e.g., pathogen wastewater monitoring and genomic surveillance). Such data integration is critical for making key parameters of stochastic epidemic models identifiable. I will illustrate the state-of-the-art statistical inference for stochastic epidemic models using Influenza, Ebola and SARS-CoV-2 surveillance data and will conclude with open problems and challenges that remain to be addressed.
About the speaker
Minin’s research interests revolve around developing statistically rigorous solutions to problems that arise in biological sciences. These solutions often involve formulating stochastic models that can describe complex dynamics of biological systems and devising computationally efficient algorithms to fit these models to data. Minin is currently most active in infectious disease epidemiology, working on Bayesian estimation of disease transmission model parameters. His other research interests include phylogenetics, population genetics, computational immunology and systems biology. Minin received a B.S. in Mathematics from Odesa National University, an M.S. in Mathematics from the University of Idaho and a Ph.D. in Biomathematics from the University of California, Los Angeles.
March, 2022
Vladimir N. Minin | Fitting stochastic epidemic models to noisy surveillance data: are we there yet?
September, 2019
Richard J. Samworth, Ph.D. | Nonparametric Inference Under Shape Constraints: Past, Present and Future
Abstract
Traditionally, we think of statistical methods as being divided into parametric approaches, which can be restrictive, but where estimation is typically straightforward (e.g. using maximum likelihood), and nonparametric methods, which are more flexible but often require careful choices of tuning parameters. The area of nonparametric inference under shape constraints sits somewhere in the middle, seeking in some ways the best of both worlds. I will give an introduction to this currently very active area, providing some history, recent developments and a future outlook.
About the speaker
Professor Richard J. Samworth's main research interests are in nonparametric and high-dimensional statistics. Particular topics include shape-constrained estimation problems; data perturbation methods (e.g. subsampling, bootstrap sampling, random projections, knockoffs); nonparametric classification; (conditional) independence testing; estimation of entropy and other functionals; changepoint detection and estimation; missing data; variable selection; and applications, including genetics, archaeology and oceanography.
September, 2018
Jon A. Wellner, Ph.D. | New Multiplier Inequalities and Applications
Abstract
Multiplier inequalities have proved to be one of the key tools of modern empirical process theory, with applications to central limit theorems, bootstrap theory and weighted likelihood methods in statistics. In this talk, I will review some classical multiplier inequalities, present a new multiplier inequality, and discuss several statistical applications. The applications include new results concerning convergence rates of least squares estimators (LSE) in regression models with possibly heavy-tailed errors. Particular cases involving sparse linear regression and shape restrictions will be mentioned.] [Professor Wellner earned his B.S. in Mathematics at the University of Idaho and his Ph.D. in Statistics at the University of Washington. He served on the faculty of the University of Rochester from 1975 to 1983, and in 1983 joined the faculty of the University of Washington, where he is Professor of Statistics and Biostatistics. His research interests include uses of large sample theory in statistics.
Professor Wellner is a fellow of the Institute of Mathematical Statistics, a fellow of the American Statistical Association and a fellow of the American Association for the Advancement of Science. In 2010, he was named a Knight of the Order of the Netherlands Lion in recognition of his exceptional research in probability and statistics, much of which has been done in collaboration with Dutch colleagues. He also received a John Simon Guggenheim Foundation Fellowship and an Alexander von Humboldt Foundation Fellowship. He was inducted into the College of Science Academy of Distinguished Alumni in 2014.