Upcoming 2026 Seminars
Title: The Genesis of Time Series Foundation Models: From Generative Pre-training to Physically-Consistent Inference
Date/Time/Location: Wednesday, April 8th at 4:00 p.m. in Barnard 347
Speaker: Defu Cao
Abstract: While foundation models have transformed fields like Natural Language Processing and Computer Vision, their application to time series analysis has been limited by unique challenges inherent to temporal data, such as complex structures, pervasive data imperfections, and the need for task generality. This talk explores the emerging frontier of foundation models designed to overcome these obstacles, tracing a path through recent breakthroughs that are redefining time series analysis. We begin by examining TEMPO, a prompt-based generative pre-trained transformer that addresses the inability of standard attention mechanisms to disentangle the non-orthogonal trend, seasonal, and residual components of time series data. Building on these insights, we introduce TimeDiT, a more general-purpose diffusion transformer architecture that employs a unified masking strategy to robustly handle missing values and perform diverse tasks like forecasting, imputation, and anomaly detection from a single framework. Derived from TimeDiT, we introduce PINFDIT, a plug-and-play, physics-informed inference process to inject domain knowledge and ensure physical consistency without retraining. The primary contribution of this research will be a novel foundation model for time series that is physically consistent, advancing the field closer to the vision of a universal "world model" for temporal data. The discussion will cover the opportunities and challenges in scaling these models, their potential impact on real-world applications in finance, healthcare, and climate science, and future research directions, including multimodal integration and enhanced interpretability.
Brief Bio: Defu Cao is a Ph.D. candidate in the Department of Computer Science at the University of Southern California and a visiting scholar at Caltech. His research focuses on building practical time series foundation models and LLM-guided decision-making systems for real-world temporal data. He works on large-scale pretraining methods, model orchestration, and interpretable inference, with applications in domains such as finance, infrastructure monitoring, and large-scale forecasting systems. His work has been published in top machine learning conferences including NeurIPS, ICML, and ICLR, and has achieved state-of-the-art performance on large-scale time series benchmarks, including a top 1 result on the GIFT-Eval leaderboard. He is a recipient of the USC Best Research Assistant Award.
PAST 2026 SEMINARS
Title: Towards Collaborative Intelligence: Learning from Decentralized Data at Scale
Date/Time/Location: Monday, February 9th at 4:10 p.m. in Barnard 108
Speaker: Yujia Wang
Abstract: As modern data increasingly comes from decentralized sources, e.g., phones, smart devices, and medical systems, learning must occur without centralizing sensitive data. Federated learning (FL) enables learning from decentralized data sources but faces significant challenges in real-world deployments, including data heterogeneity, system variability, and communication bottlenecks. In this talk, I will present the algorithmic and optimization foundations of collaborative intelligence, focusing on building efficient and scalable learning from decentralized data. My work addresses FL’s challenges both individually and in a more systematic, integrated way, depending on what the problem demands. I will first diagnose how stale updates and data heterogeneity jointly destabilize asynchronous FL and introduce a cached calibration mechanism that probably removes the harmful delay-heterogeneity interaction. I will then introduce a modularized and parallel block-coordinate framework for federated fine-tuning of large language models. Together, these results establish optimization-driven principles that enable efficient and scalable federated learning. The talk concludes with a vision for the next generation of collaborative AI, where models learn efficiently while respecting privacy, system constraints, and social trustworthiness
Bio: Yujia Wang is a Ph.D. candidate in the College of Information Sciences and Technology at The Pennsylvania State University, advised by Dr. Jinghui Chen. Her research spans the theories and applications of collaborative intelligence and privacy-preserving machine learning. Her work has been published in top venues such as ICML, NeurIPS, ICLR, AISTATS, ACL and TMLR. She has delivered technical talks at the SIAM-NNP Section Conference and IBM Research, and presented her work at the SDM Doctoral Forum. She actively serves as a reviewer for leading AI conferences and journals. Beyond academia, she gained industry experience as a Research Intern at IBM Research.
Title: Visualization Design and Artificial Intelligence for Scientific Inquiry
Date/Time/Location: Monday, February 23rd at 4:10 p.m. in Barnard 108
Speaker: Devin Lange
Abstract: Scientific inquiry seeks to understand how the world works, and data is a fundamental tool for representing underlying phenomena. In this talk, I discuss how well-designed visualizations, together with the integration of artificial intelligence, can help domain scientists interact with and reason about their data. Cancer research serves as a motivating application domain, as it involves complex, large-scale, and multimodal datasets. These data sets range from images of individual cancer cells to datasets assembled through collaborations across multiple institutions. Visualization plays a critical role in helping scientists navigate this complexity and extract meaningful insight from their data.
Bio: Devin Lange is a postdoctoral fellow in biomedical informatics at Harvard Medical School in the HIDIVE Lab, where he works with Nils Gehlenborg. He earned a PhD in computer science from the University of Utah under the supervision of Alexander Lex. His research investigates how visualization and artificial intelligence can support the understanding, discovery, and quality control of scientific data, with a particular focus on biological and biomedical data.His work has been recognized with multiple awards at IEEE VIS, including a Best Paper Award for Aardvark, a visualization system for integrated analysis of imaging, time-series, and cellular divisions. His research spans biological visualization, data forensics, and natural-language interfaces for interactive visual analysis.
Title: QCORE Tour
Date/Time/Location: Monday, March 30th at 4:00 p.m. at EngineWorks (2425 Technology Blvd)
Abstract: Learn about quantum initiatives at MSU and opportunities to engage with QCORE
Arrival: The Engine works building is located at 2425 Technology Blvd, Bozeman, MT 59718. Sometimes the address doesn't get you right to the entrance, but you park in the main parking lot, then enter the lobby under the "EngineWorks" sign. From there we have an IPAD that EngineWorks requires all members of your party to sign in on for security tracking purposes. It will prompt you to enter the name of the person you are meeting with, in this case, Jayne Morrow
Seminars from 2025.
