Date & Time:
March 25, 2026 11:30 am – 12:30 pm
Location:
DSI 105, 5460 S University, Chicago, IL,
03/25/2026 11:30 AM 03/25/2026 12:30 PM America/Chicago Qing Qu (UMich): Harnessing Low-Dimensionality for Generalizable and Scientific Generative AI DSI 105, 5460 S University, Chicago, IL,

Abstract: The empirical success of modern generative AI, from diffusion models to Large Language Models (LLMs), often outpaces our classical understanding of how machine learning models generalize from finite, out-of-distribution (OOD) data. This talk introduces a unified mathematical framework identifying intrinsic low-dimensional structures as the primary driver of generalization and a critical lever for advancing scientific AI. First, we deconstruct the generalization mechanism of diffusion models, revealing a training transition from memorization to generalization that effectively breaks the curse of dimensionality. By using a mixture of low-rank Gaussian models, we demonstrate that sample complexity scales linearly with the intrinsic dimension rather than exponentially with the ambient dimension, through establishing a formal equivalence with the canonical subspace clustering problem. Moreover, by exploring nonlinearity in two-layer denoising autoencoders, we uncover how weight structures differ between memorization and generalization. This distinction allows us to provide a unified understanding of how models learn representations and how they generate new data. Second, we characterize the OOD generalization of in-context learning (ICL) in transformers. For linear regression tasks in which vectors lie in low-dimensional subspaces, we show that OOD capabilities emerge from interpolating across training task subspaces. We derive precise conditions under which linear attention models interpolate across distribution shifts, highlighting task diversity as a prerequisite for ICL efficacy. Finally, we translate these theoretical insights into practical guidelines for controlled generation, ensuring model safety and privacy, and solving high-dimensional inverse problems in science and engineering.

Speakers

headshot

Qing Qu

Assistant Professor in EECS, the University of Michigan

Qing Qu is an Assistant Professor in EECS at the University of Michigan. He works at the intersection of the foundations of machine learning, numerical optimization, and signal/image processing, with a current focus on the theory of deep generative models and representation learning. Prior to joining Michigan in 2021, he was a Moore–Sloan Data Science Fellow at the Center for Data Science, New York University (2018–2020). He received his Ph.D. in Electrical Engineering from Columbia University in October 2018 and his B.Eng. in Electrical and Computer Engineering from Tsinghua University in July 2011. His work has been recognized with multiple honors, including the Best Student Paper Award at SPARS 2015, a Microsoft PhD Fellowship in Machine Learning (2016), the Best Paper Award at the NeurIPS Diffusion Models Workshop (2023), NSF CAREER Award (2022), Amazon Research Award (AWS AI, 2023), UM CHS Junior Faculty Award (2025), Google Research Scholar Award (2025), and the 1938E Award in Michigan Engineering. He has led and delivered multiple tutorials at ICASSP, CPAL, CVPR, ICCV, and ICML. He was one of the founding organizers and Program Chair for the new Conference on Parsimony & Learning (CPAL), regularly serves as an Area Chair for NeurIPS, ICML, and ICLR, senior area chair for ICASSP’26, and is an Action Editor for TMLR.

Related News & Events

headshot
UChicago CS News

University of Chicago PhD Student Riki Otaki Receives MongoDB PhD Fellowship Award

Feb 26, 2026
Robert Grossman presenting
UChicago CS News

M3 Workshop Advances Federated AI for Biomedical Research

Feb 23, 2026
headshot
UChicago CS News

Aloni Cohen Named Sloan Research Fellow for Work Bridging Law and Computer Science

Feb 17, 2026
TEI conference announcement
UChicago CS News

This Spring at UChicago: TEI’26 Unites Technology, Art, and Design on Campus

Feb 03, 2026
neutron star
UChicago CS News

RADAR: A new era of collaborative cosmic exploration

Jan 28, 2026
privacy settings example
UChicago CS News

Designed to Deceive: Why Knowledge Isn’t Enough to Beat Dark Patterns

Jan 27, 2026
headshot
UChicago CS News

Bridging Physics and CS: A Conversation with our latest IBM PhD Fellow, Soumik Ghosh

Jan 23, 2026
Tanya presenting research
UChicago CS News

Ranya Sharma Receives CRA Outstanding Undergraduate Researcher Award

Jan 22, 2026
Tensormesh CEO Junchen Jiang
Video

Building Tensormesh: A Conversation with the CEO (Junchen Jiang)

Jan 08, 2026
cityscape
UChicago CS News

UChicago Researchers Help Launch First International Conference on AI Scientists in Beijing

Jan 08, 2026
test of time headshots
UChicago CS News

Five Paths to Lasting Influence: Celebrating Five UChicago CS Test of Time Award Recipients

Dec 02, 2025
technology architecture
UChicago CS News

Researchers Built Their Own ISP to Fix the Internet– A Decade Later, It’s Still Running

Nov 20, 2025
arrow-down-largearrow-left-largearrow-right-large-greyarrow-right-large-yellowarrow-right-largearrow-right-smallbutton-arrowclosedocumentfacebookfacet-arrow-down-whitefacet-arrow-downPage 1CheckedCheckedicon-apple-t5backgroundLayer 1icon-google-t5icon-office365-t5icon-outlook-t5backgroundLayer 1icon-outlookcom-t5backgroundLayer 1icon-yahoo-t5backgroundLayer 1internal-yellowinternalintranetlinkedinlinkoutpauseplaypresentationsearch-bluesearchshareslider-arrow-nextslider-arrow-prevtwittervideoyoutube