Events
SkAI Institute Events
SkAI Institute events are held at the SkAI Hub in the John Hancock Center, Suite 3500, unless otherwise indicated, and are in Central (Standard) Time. If you have any questions about the events, please contact Emma Alexander (ealexander at northwestern.edu) and Alex Ćiprijanović (aleksand at fnal.gov) and, for logistics, please contact SkAI Operations Coordinator, Gema Tinoco (gema.tinoco at northwestern.edu). For questions about a Journal Club meeting, please contact Xinfeng Xu (xinfeng.xu at northwestern.edu).
All SkAI events are presented in a hybrid (in-person and Zoom) format. Whether you plan to attend in person or virtually, please RSVP using the registration forms below, which will help us to get an accurate head count for catering and reporting purposes, and to make sure a visitor pass is requested for everyone who needs one. Please note that participants agree to follow the SkAI Institute Code of Conduct.
First time visiting the SkAI Hub? We have a brief orientation to help you know what to expect.
Location: John Hancock Center, Suite 3500 (unless otherwise indicated)
Zoom: Link
Registration Forms:
Registration Form for February 4 SkAI Works-In-Progress Wednesday
Registration Form for February 18 SkAI Colloquium
Upcoming Events:
Click on the speaker’s photo or name to learn more about them. Click on the title to reveal the abstract.
All SkAI events are presented in a hybrid (in-person and Zoom) format. Whether you plan to attend in person or virtually, please RSVP using the registration forms below, which will help us to get an accurate head count for catering and reporting purposes, and to make sure a visitor pass is requested for everyone who needs one. Please note that participants agree to follow the SkAI Institute Code of Conduct.
First time visiting the SkAI Hub? We have a brief orientation to help you know what to expect.
Location: John Hancock Center, Suite 3500 (unless otherwise indicated)
Zoom: Link
Registration Forms:
Registration Form for February 4 SkAI Works-In-Progress Wednesday
Registration Form for February 18 SkAI Colloquium
Upcoming Events:
Click on the speaker’s photo or name to learn more about them. Click on the title to reveal the abstract.
| Date/Location | Speaker | Affiliation | Title | Links | ||
| Upcoming | ||||||
| February 4, 2026 Works-In-Progress Wednesday 11:00–11:55 a.m. Hancock, 3500 |
![]()
|
Vicky Kalogera Aggelos Katsaggelos |
Northwestern University |
TBA.
| — | |
| February 11, 2026 Works-In-Progress Wednesday 11:00–11:55 a.m. Hancock, 3500 |
![]()
|
Billy Moses Joaquin Vieira |
University of Illinois Urbana-Champaign |
TBA.
| — | |
| February 18, 2026 Colloquium 11:00–11:55 a.m. Hancock, 3500 |
|
Berthy Feng | Massachusetts Institute of Technology |
Computational imaging aims to visualize scientific phenomena beyond the reach of conventional optics by incorporating assumptions, or priors, about the object being imaged. In the age of AI, the priors available to us are more sophisticated than ever. However, it is yet unclear how to use these priors rigorously, especially when using imaging in the scientific process. The main challenge of computational imaging for science is discerning when to use what prior and how. The art of imaging lies in deciding the correct balance of assumptions to obtain trustworthy and informative images. Once the prior has been decided, the question is how to incorporate it rigorously.
In this talk, I will present my previous work on building principled routes for incorporating data-driven and physics-based priors. On the data-driven side, I will show results of re-imagining the famous M87 black hole from real data with score-based priors. On the physics-based side, I will show how we have tackled extremely underdetermined imaging problems by enforcing physics constraints, including the problem of single-viewpoint dynamic tomography of emission near a black hole. In the intersection of AI and physics, I will present neural approximate mirror maps, a way to enforce physics constraints on generative models. I will then discuss future directions for further taming priors so that we can rigorously create, interpret, and extract insights from scientific images. | — |
|
| March 11, 2026 Works-in-Progress Wednesday 10:00–11:00 a.m. Hancock, 3500 |
![]()
|
Adam Miller Han Liu |
Northwestern University |
TBA.
| — | |
| 11:00–11:55 a.m |
![]()
|
Ermin Wei Jeff McMahon |
Northwestern University The University of Chicago |
TBA.
| — | |
| March 18, 2026 Colloquium 11:00–11:55 a.m. Hancock, 3500 |
|
Yuanyuan Shi | University of California, San Diego |
TBA.
| — |
|
| March 25, 2026 Works-In-Progress Wednesday 11:00–11:55 a.m. Hancock, 3500 |
![]() ![]()
|
Alex Drlica-Wagner Aravindan Vijayaraghavan Paul Chichura |
The University of Chicago Northwestern University SkAI Institute |
TBA.
| — | |
| April 1, 2026 Works-In-Progress Wednesday 11:00–11:55 a.m. Hancock, 3500 |
![]()
|
Rina Barber Daniel Holz |
The University of Chicago |
TBA.
| — | |
| April 8, 2026 Works-in-Progress Wednesday 11:00–11:55 a.m. Hancock, 3500 |
![]()
|
Rebecca Willett Chihway Chang |
The University of Chicago |
TBA.
| — | |
| April 15, 2026 Colloquium 11:00–11:55 a.m. Hancock, 3500 |
|
Keith Bechtol | University of Wisconsin–Madison |
TBA.
| — |
|
| April 22, 2026 Works-in-Progress Wednesday 10:00–11:00 a.m. Hancock, 3500 |
![]()
|
Aleksandra Ćiprijanović Sandeep Madireddy |
Fermilab Argonne National Laboratory |
TBA.
| — | |
| 11:00–11:55 a.m |
![]()
|
Tjitske Starkenburg Emma Alexander |
Northwestern University |
TBA.
| — | |
| May 6, 2026 Works-In-Progress Wednesday 11:00–11:55 a.m. Hancock, 3500 |
![]()
|
Noelle Samia Gautham Narayan |
Northwestern University University of Illinois Urbana-Champaign |
TBA.
| — | |
| May 13, 2026 Colloquium 11:00–11:55 a.m. Hancock, 3500 |
|
Tom Downes | Microsoft AI |
TBA.
| — | |
| Past | ||||||
| January 14, 2025 Colloquium 11:00–11:55 a.m. Hancock, 3500 |
|
David Hogg | New York University |
Large language models (LLMs) are beginning to obtain the ability to design, execute, write up, and referee scientific projects on the data-science side of astrophysics. What implications does this have for our profession? I will try to elucidate a set of facts or points of agreement about what astrophysics is, or should be. I will also deliver a set of putative benefits that astrophysics perhaps brings to us, and to science, and to universities, and to the world. My goal is to deliver some kind of take on why astronomy and astrophysics are worth doing, and how LLMs might fit into a practice with that justification
| — |
|
| December 17, 2025 Works-in-Progress Wednesday 10:00–11:00 a.m. Hancock, 3500 |
![]()
|
Aleksandra Ćiprijanović Sandeep Madireddy |
Fermilab Argonne National Laboratory |
TBA.
| — | |
| 11:00–11:55 a.m. |
![]() ![]()
|
Alex Drlica-Wagner Aravindan Vijayaraghavan Paul Chichura |
The University of Chicago Northwestern University SkAI Institute |
TBA.
| — | |
| December 10, 2025 SkAI AI Ethics Tutorial 10:00–11:00 a.m. Hancock, 3500 |
|
Lu Cheng | University of Illinois Chicago |
Generative foundation models (GenFMs), including large language and multimodal models, are transforming information retrieval and knowledge management. However, their rapid adoption raises urgent concerns about social responsibility, trustworthiness, and governance. This mini-tutorial offers an overview of recent advances in responsible GenFMs, covering foundational concepts and multidimensional risk taxonomies (including safety, privacy, robustness, truthfulness, fairness, and machine ethics).
| — |
|
| December 3, 2025 Journal Club 10:00–11:00 a.m. Hancock, 3500 |
|
Yiping Lu | Northwestern University |
Scaling scientific machine learning (SciML) requires overcoming bottlenecks at both training and inference. On the training side, we study the statistical convergence rate and limits of deep learning for solving elliptic PDEs from random samples. While our theory predicts optimal polynomial convergence for PINNs, optimization becomes prohibitively ill-conditioned as networks widen. By adapting descent strategies to the optimization geometry, we obtain scale-invariant training dynamics that translate polynomial convergence into concrete compute and yield compute-optimal configurations. On the inference side, I will introduce Simulation-Calibrated SciML (SCaSML), a physics-informed post-processing framework that improves surrogate models without retraining or fine-tuning. By enforcing physical laws, SCaSML delivers trustworthy corrections (via Feynman-Kac simulation) with approximate confidence intervals, achieves faster and near-optimal convergence rates, and supports online updates for digital twins. Together, these results integrate theory and numerics to enable predictable, reliable scaling of SciML in both training and inference.
| — |
|
| 10:00–11:00 a.m. |
![]()
|
Lindsay House Benjamin Remy |
SkAI Institute |
We are merging a large participatory science effort with machine learning to enhance the Hobby-Eberly Telescope Dark Energy Experiment (HETDEX). Our overall goal is to remove false positives, allowing us to use lower signal-to-noise data and sources with low goodness-of-fit. With six million classifications through Dark Energy Explorers, we can confidently determine if a source is not real at over 94% confidence level when classified by at least 10 individuals; this confidence level increases for higher signal-to-noise sources. To date, we have only been able to apply this direct analysis to 190,000 sources. The full sample of HETDEX will contain around 2–3 million sources, including nearby galaxies ([O II] emitters), distant galaxies (Lyα emitters or LAEs), false positives, and contamination from instrument issues. We can accommodate this tenfold increase by using machine learning with visually vetted samples from Dark Energy Explorers. We have already increased by over tenfold the number of sources that have been visually vetted from our previous pilot study where we only had 14,000 visually vetted LAE candidates. This paper expands on the previous work by increasing the visually vetted sample from 14,000 to 190,000. In addition, using our currently visually vetted sample, we generate a real or false positive classification for the full candidate sample of 1.2 million LAEs. We currently have approximately 17,000 volunteers from 159 countries around the world. Thus, we are applying participatory or citizen scientist analysis to our full HETDEX data set, creating a free educational opportunity that requires no prior technical knowledge.
Context. Weak lensing mass-mapping is a useful tool for accessing the full distribution of dark matter on the sky, but because of intrinsic galaxy ellipticies, finite fields, and missing data, the recovery of dark matter maps constitutes a challenging, ill-posed inverse problem.
Aims. We introduce a novel methodology that enables the efficient sampling of the high-dimensional Bayesian posterior of the weak lensing mass-mapping problem, relying on simulations to define a fully non-Gaussian prior. We aim to demonstrate the accuracy of the method to simulated fields, and then proceed to apply it to the mass reconstruction of the HST/ACS COSMOS field. Methods. The proposed methodology combines elements of Bayesian statistics, analytic theory, and a recent class of deep generative models based on neural score matching. This approach allows us to make full use of analytic cosmological theory to constrain the 2pt statistics of the solution, to understand any differences between this analytic prior and full simulations from cosmological simulations, and to obtain samples from the full Bayesian posterior of the problem for robust uncertainty quantification. Results. We demonstrate the method in the κTNG simulations and find that the posterior mean significantly outperforms previous methods (Kaiser–Squires, Wiener filter, Sparsity priors) both for the root-mean-square error and in terms of the Pearson correlation. We further illustrate the interpretability of the recovered posterior by establishing a close correlation between posterior convergence values and the S/N of the clusters artificially introduced into a field. Finally, we apply the method to the reconstruction of the HST/ACS COSMOS field, which yields the highest-quality convergence map of this field to date. Conclusions. We find the proposed approach to be superior to previous algorithms, scalable, providing uncertainties, and using a fully non-Gaussian prior. | Article Article |
|
| November 19, 2025 Works-in-Progress Wednesday 10:00–11:00 a.m. Hancock, 3500 |
![]() ![]()
|
Rina Barber Daniel Holz Ruiting Liang |
The University of Chicago |
TBA.
| — | |
| 10:00–11:00 a.m. |
![]() ![]()
|
Roy Zhao Shi-Hui Zang Benjamin Remy |
The University of Chicago University of Wisconsin–Madison SkAI Institute Additional speakers: Yuuki Omori, Shrihan Agarwal, Georgios Valogiannis, Rebecca Willett, Chihway Chang |
TBA.
| — | |
| November 12, 2025 Seminar and Colloquium 11:00–11:55 a.m. Hancock, 3500 |
|
Yuan-Sen Ting | The Ohio State University |
The expansive, interdisciplinary nature of astronomy, combined with its open-access culture, makes it an ideal testing ground for exploring how Large Language Models (LLMs) can accelerate scientific discovery. Recent developments in LLM reasoning capabilities have shown substantial progress—our work demonstrates that AI agents can now achieve gold medal performance on International Olympiad on Astronomy and Astrophysics (IOAA) problems, indicating their growing analytical abilities. In this talk, I will present our recent advances in applying LLMs as agents to real-world astronomical challenges. Through self-play reinforcement learning, we demonstrate how LLM agents can conduct end-to-end research tasks in galaxy spectral fitting, encompassing data analysis, strategy refinement, and outlier detection—approaching capabilities similar to human intuition and domain knowledge. However, limitations remain. While autonomous research agents like Mephisto could theoretically help analyze all observed sources, the cost of closed-source solutions remains prohibitive for large-scale applications involving billions of objects. Additionally, the Moravec paradox manifests clearly in astronomy: tasks requiring abstract reasoning may be easier for AI than seemingly simple perceptual tasks. Current models still struggle with chart reading, multimodal data interpretation, and other fundamental astronomical workflows. To address the cost limitation, we developed lightweight, open-source specialized models (AstroSage and AstroLLaMA)—trained on arXiv literature—and evaluated them against carefully curated astronomical benchmarks. Our research shows that these specialized LLMs can outperform larger general-purpose models on astronomy Q&A tasks when appropriately pretrained and fine-tuned, demonstrating a path forward for building more capable and accessible astronomy-specific models. Looking ahead, the path forward involves integrating more function-calling tools and building a comprehensive ecosystem—not just better models. The astronomical community's collaborative infrastructure will be important for scaling up automated inference and expanding the role of AI in astronomical research.
| ![]() |
|
| 10:00–10:50 a.m. |
![]()
|
Cliff Johnson Michael Zevin |
Northwestern University Adler Planetarium |
Over the past decade, the Zooniverse has grown into the most successful and widely used online platform for people-powered research. While one of the Zooniverse’s key strengths is its large community of nearly 3 million engaged volunteers who contribute their effort to performing research tasks, the platform has increasingly committed resources toward building and supporting machine-learning infrastructure. While human effort remains essential to the success of the platform’s projects, the combination of human and machine intelligence is becoming increasingly necessary to keep up with ever-increasing data volumes and complexity. In this talk, we review Zooniverse’s current strategies and tools for machine-learning integration, and look ahead at a number of exciting possibilities for new and enhanced AI/ML platform features. We wish to share ideas and solicit feedback from AI and Astro SkAI researchers regarding potential new directions, which include: anomaly detection and curation; latent space exploration; active learning, and rapid model refinement.
| ![]() |
|
| November 5, 2025 Journal Club 10:00–11:00 a.m. Hancock, 3500 |
![]()
|
Paul Chichura Christian Jespersen |
SkAI Institute Princeton University |
We present improvements to the pointing accuracy of the South Pole Telescope (SPT) using machine learning. The ability of the SPT to point accurately at the sky is limited by its structural imperfections, which are impacted by the extreme weather at the South Pole. Pointing accuracy is particularly important during SPT participation in observing campaigns with the Event Horizon Telescope (EHT), which requires stricter accuracy than typical observations with the SPT. We compile a training dataset of historical observations of astronomical sources made with the SPT-3G and EHT receivers on the SPT. We train two XGBoost models to learn a mapping from current weather conditions to two telescope drive control arguments—one which corrects for errors in azimuth and the other for errors in elevation. Our trained models achieve root mean squared errors on withheld test data of 2.14″ in cross-elevation and 3.57″ in elevation, well below our goal of 5″ along each axis. We deploy our models on the telescope control system and perform further in situ test observations during the EHT observing campaign in 2024 April. Our models result in significantly improved pointing accuracy: for sources within the range of input variables where the models are best trained, average combined pointing error improved 33%, from 15.9″ to 10.6″. These improvements, while significant, fall shy of our ultimate goal, but they serve as a proof of concept for the development of future models. Planned upgrades to the EHT receiver on the SPT will necessitate even stricter pointing accuracy, which will be achievable with our methods.
Galaxies are often modelled as composites of separable components with distinct spectral signatures, implying that different wavelength ranges are only weakly correlated. They are not. We present a data-driven model which exploits subtle correlations between physical processes to accurately predict infrared (IR) WISE photometry from a neural summary of optical SDSS spectra. The model achieves accuracies of ꭓ2N ≈ 1 for all photometric bands in WISE, as well as good colors. We are also able to tightly constrain typically IR-derived properties, e.g. the bolometric luminosities of AGN and dust parameters such as qPAH. We find that current SED-fitting methods are incapable of making comparable predictions, and that model misspecification often leads to correlated biases in star-formation rates and AGN luminosities. To help improve SED models, we determine what features of the optical spectrum are responsible for our improved predictions, and identify several lines (CaII, SrII, FeI, [OII] and Hα), which point to the complex chronology of star formation and chemical enrichment being incorrectly modelled.
| Article Article |
|
| October 29, 2025 SkAI Industry Panel for Early-Career Scientists 10:00–11:30 a.m. Hancock, 3500 |
![]() ![]() ![]() ![]()
|
Patrick Aleo Frank Fineis James Guillochon Kara Ponder Dimitrios Tanoglidis |
Honeywell Federal Manufacturing & Technologies, LLC Clearcover Esri Charles Schwab Walgreens |
TBA.
| — | |
| October 22, 2025 Works-In-Progress Wednesday 11:00–11:55 a.m. Hancock, 3500 |
![]()
|
Ermin Wei Jeff McMahon |
Northwestern University The University of Chicago |
TBA.
| — | |
| 10:00–11:00 a.m. |
![]()
|
Tjitske Starkenburg Emma Alexander |
Northwestern University |
TBA.
| — | |
| October 15, 2025 SkAI Onboarding Tutorial 10:00–11:00 a.m. Hancock, 3500 |
|
Jennifer Li | University of Illinois Urbana-Champaign |
This onboarding tutorial will include an overview on UIUC’s Delta/DeltaAI computing resources, step-by-step instructions to request computing time on Delta/DeltaAI, and a getting-started guide on using Delta/DeltaAI. Jennifer will also give a quick overview of the National Center for Supercomputing Applications’ (NCSA) extensive online course on high-performance computing (HPC). This interactive tutorial will also outline how SkAI members can apply for compute resources.
| — |
|
| October 8, 2025 Works-In-Progress Wednesday 11:00–11:55 a.m. Hancock, 3500 |
![]()
|
Adam Miller Han Liu |
Northwestern University |
TBA.
| — | |
| 10:00–11:00 a.m. |
![]()
|
Noelle Samia Gautham Narayan |
Northwestern University University of Illinois Urbana-Champaign |
TBA.
| — | |
| October 1, 2025 Journal Club and Colloquium 11:10–11:55 a.m. Hancock, 3500 |
|
Christopher Stubbs | Harvard University |
TBA.
| — | |
10:00–11:00 a.m. |
![]()
|
Sydney Erickson Yongseok Jo |
Stanford University SkAI Institute |
Strongly lensed quasars can be used to constrain cosmological parameters through time-delay cosmography. Models of the lens masses are a necessary component of this analysis. To enable time-delay cosmography from a sample of O(103)
lenses, which will soon become available from surveys like the Rubin Observatory’s Legacy Survey of Space and Time and the Euclid Wide Survey, we require fast and standardizable modeling techniques. To address this need, we apply neural posterior estimation (NPE) for modeling galaxy-scale strongly lensed quasars from the Strong Lensing Insights into the Dark Energy Survey (STRIDES) sample. NPE brings two advantages: speed and the ability to implicitly marginalize over nuisance parameters. We extend this method by employing sequential NPE to increase precision of mass model posteriors. We then fold individual lens models into a hierarchical Bayesian inference to recover the population distribution of lens mass parameters, accounting for out-of-distribution shift. After verifying our method using simulated analogs of the STRIDES lens sample, we apply our method to 14 Hubble Space Telescope single-filter observations. We find the population mean of the power-law elliptical mass distribution slope, γlens, to be Μγlens = 2.13 +/– 0.06. Our result represents the first population-level constraint for these systems. This population-level inference from fully automated modeling is an important stepping stone toward cosmological inference with large samples of strongly lensed quasars.
The rapid advancement of large-scale cosmological simulations has opened new avenues for cosmological and astrophysical research. However, the increasing diversity among cosmological simulation models presents a challenge to the robustness. In this work, we develop the Model-Insensitive ESTimator (MIEST), a machine that can robustly estimate the cosmological parameters, Ωm and σ8, from neural hydrogen maps of simulation models in the CAMELS project—TNG, SIMBA, ASTRID, and EAGLE. An estimator is considered robust if it possesses a consistent predictive power across all simulations, including those used during the training phase. We train our machine using multiple simulation models and ensure that it only extracts common features between the models while disregarding the model-specific features. This allows us to develop a novel model that is capable of accurately estimating parameters across a range of simulation models, without being biased towards any particular model. Upon the investigation of the latent space—a set of summary statistics, we find that the implementation of robustness leads to the blending of latent variables across different models, demonstrating the removal of model-specific features. In comparison to a standard machine lacking robustness, the average performance of MIEST on the unseen simulations during the training phase has been improved by ~17% for Ωm and ~38% for σ8. By using a machine learning approach that can extract robust, yet physical features, we hope to improve our understanding of galaxy formation and evolution in a (subgrid) model-insensitive manner, and ultimately, gain insight into the underlying physical processes responsible for robustness. This is a Learning the Universe publication.
| Article Article |
|
| September 17, 2025 Journal Club 10:00–11:00 a.m. Hancock, 3500 |
![]()
|
Tri Nguyen Jimena González |
Northwestern University SkAI Institute |
TBA.
We conduct a search for strong gravitational lenses in the Dark Energy Survey (DES) Year 6 imaging data. We implement a pre-trained Vision Transformer (ViT) for our machine learning (ML) architecture and adopt Interactive Machine Learning to construct a training sample with multiple classes to address common types of false positives. Our ML model reduces 236 million DES cutout images to 22,564 targets of interest, including around 85% of previously reported galaxy-galaxy lens candidates discovered in DES. These targets were visually inspected by citizen scientists, who ruled out approximately 90% as false positives. Of the remaining 2,618 candidates, 149 were expert-classified as ‘definite’ lenses and 516 as ‘probable’ lenses, with 147 of these candidates being newly identified. Additionally, we trained a second ViT to find double-source plane lens systems, finding at least one double-source system. Our main ViT excels at identifying galaxy-galaxy lenses, consistently assigning high scores to candidates with high confidence. The top 800 ViT-scored images include around 100 of our ‘definite’ lens candidates. This selection is an order of magnitude higher in purity than previous convolutional neural network-based lens searches and demonstrates the feasibility of applying our methodology for discovering large samples of lenses in future surveys.
| — Article |
|
| September 10, 2025 Works-in-Progress Wednesday 11:00–11:55 a.m. Hancock, 3500 |
![]()
|
Billy Moses Joaquin Vieira |
University of Illinois Urbana-Champaign |
TBA.
| — | |
| 10:00–11:00 a.m. |
![]()
|
Vicky Kalogera Aggelos Katsaggelos |
Northwestern University |
TBA.
| — | |
| August 29, 2025 Journal Club and Featured Conversation 11:00–11:45 a.m. Hancock, 3500 |
|
Andreas Berlind | NSF Astronomy Division |
TBA.
| — | |
| 10:00–11:00 a.m. |
|
Elizabeth Teng | Northwestern University |
TBA.
| — | |
| May 28, 2025 Works-in-Progress Wednesday, SkAI Hub Orientation, and Journal Club 11:00–11:55 a.m. Hancock, 4010 |
![]()
|
Aldana Grichener Dalya Baron |
The University of Arizona Stanford University |
One of the main challenges in modeling massive stars to the onset of core collapse is the computational bottleneck of nucleosynthesis during advanced burning stages. The number of isotopes formed requires solving a large set of fully-coupled stiff ordinary differential equations (ODEs), making the simulations computationally intensive and prone to numerical instability. To overcome this barrier, we design a nuclear neural network (NNN) framework with multiple hidden layers to emulate nucleosynthesis calculations and conduct a proof-of-concept to evaluate its performance. The NNN takes the temperature, density and composition of a burning region as input and predicts the resulting isotopic abundances along with the energy generation and loss rates. We generate training sets for initial conditions corresponding to oxygen core depletion and beyond using large nuclear reaction networks, and compare the predictions of the NNNs to results from a commonly used small net. We find that the NNNs improve the accuracy of the electron fraction by 280–660% and the nuclear energy generation by 250–750%, consistently outperforming the small network across all timesteps. They also achieve significantly better predictions of neutrino losses on relatively short timescales, with improvements ranging from 100–106%. While further work is needed to enhance their accuracy and applicability to different stellar conditions, integrating NNN trained models into stellar evolution codes is promising for facilitating large-scale generation of core-collapse supernova (CCSN) progenitors with higher physical fidelity.
The structure and chemistry of the dusty interstellar medium (ISM) are shaped by complex processes that depend on the local radiation field, gas composition, and dust grain properties. Of particular importance are polycyclic aromatic hydrocarbons (PAHs), which emit strong vibrational bands in the mid-infrared, and play a key role in the ISM energy balance. We recently identified global correlations between PAH band and optical line ratios across three nearby galaxies, suggesting a connection between PAH heating and gas ionization throughout the ISM. In this work, we perform a census of the PAH heating–gas ionization connection using ∼700,000 independent pixels that probe scales of 40–150 pc in 19 nearby star-forming galaxies from the PHANGS survey. We find a universal relation between PAH(11.3 μm/7.7 μm) and ([S II]/Hα) with a slope of ∼0.2 and a scatter of ∼0.025 dex. The only exception is a group of anomalous pixels that show unusually high (11.3 μm/7.7 μm) PAH ratios in regions with old stellar populations and high starlight-to-dust emission ratios. Their mid-infrared spectra resemble those of elliptical galaxies. Active galactic nucleus hosts show modestly steeper slopes, with a ∼10% increase in PAH(11.3 μm/7.7 μm) in the diffuse gas on kiloparsec scales. This universal relation implies an emerging simplicity in the complex ISM, with a sequence that is driven by a single varying property: the spectral shape of the interstellar radiation field. This suggests that other properties, such as gas-phase abundances, gas ionization parameter, and grain charge distribution, are relatively uniform in all but specific cases.
|
Article Article |
|
| 10:00–11:00 a.m. |
![]() SkAI Operations Team |
Julian Cuevas-Zepeda |
The University of Chicago Northwestern University |
TBA.
TBA.
|
— — |
|
| May 21, 2025 Works-in-Progress Wednesday and Colloquium 10:00–11:55 a.m. Hancock, 35th Floor |
![]()
|
Weijian Li Souvik Chakraborty |
Northwestern University Indian Institute of Technology (IIT) Delhi |
TBA.
Operator learning is an emerging area in scientific machine learning which aims to learn mappings between infinite dimensional function spaces. In the first half of the talk, I will delve into the foundations of Wavelet Neural Operator (WNO), a recently developed operator learning algorithm. I will discuss its working principles and its potential applications in complex engineering problems including fracture propagation in materials, tumor detection using USG data and elastography, and climate modelling.
The second half of the talk will focus on what lies beyond neural operators. I will introduce a new scientific machine learning architecture that is loosely motivated from cognitive science. This architecture is a first of its kind foundation model and offers two key advantages: (i) it can simultaneously learn solution operators for multiple parametric PDEs, and (ii) rapid generalization to new parametric PDEs with minimal fine-tuning. We observe that the proposed architecture is robust against catastrophic forgetting and facilitate knowledge transfer across dissimilar tasks. Across a diverse array of mechanics problems, consistent performance enhancements are observed with this architecture compared to task-specific baseline operator learning frameworks. |
— — |
|
| May 14, 2025 Journal Club 11:00–11:55 a.m. Hancock, 4010 |
![]()
|
Jason Sun Philipp Rajah Moura Srivastava |
Northwestern University |
The first measurements of the 21 cm brightness temperature power spectrum from the epoch of reionization will very likely be achieved in the near future by radio interferometric array experiments such as the Hydrogen Epoch of Reionization Array (HERA) and the Square Kilometre Array (SKA). Standard MCMC analyses use an explicit likelihood approximation to infer the reionization parameters from the 21 cm power spectrum. In this paper, we present a new Bayesian inference of the reionization parameters where the likelihood is implicitly defined through forward simulations using density estimation likelihood-free inference (DELFI). Realistic effects, including thermal noise and foreground avoidance, are also applied to the mock observations from the HERA and SKA. We demonstrate that this method recovers accurate posterior distributions for the reionization parameters, and it outperforms the standard MCMC analysis in terms of the location and size of credible parameter regions. With the minute-level processing time once the network is trained, this technique is a promising approach for the scientific interpretation of future 21 cm power spectrum observation data. Our code 21cmDELFI-PS is publicly available at this link (https://github.com/Xiaosheng-Zhao/21cmDELFI).
Tomographic three-dimensional 21 cm images from the epoch of reionization contain a wealth of information about the reionization of the intergalactic medium by astrophysical sources. Conventional power spectrum analysis cannot exploit the full information in the 21 cm data because the 21 cm signal is highly non-Gaussian due to reionization patchiness. We perform a Bayesian inference of the reionization parameters where the likelihood is implicitly defined through forward simulations using density estimation likelihood-free inference (DELFI). We adopt a trained 3D convolutional neural network (CNN) to compress the 3D image data into informative summaries (DELFI-3D CNN). We show that this method recovers accurate posterior distributions for the reionization parameters. Our approach outperforms earlier analysis based on two-dimensional 21 cm images. In contrast, a Monte Carlo Markov Chain analysis of the 3D light-cone-based 21 cm power spectrum alone and using a standard explicit likelihood approximation results in less accurate credible parameter regions than inferred by the DELFI-3D CNN, both in terms of the location and shape of the contours. Our proof-of-concept study implies that the DELFI-3D CNN can effectively exploit more information in the 3D 21 cm images than a 2D CNN or power spectrum analysis. This technique can be readily extended to include realistic effects and is therefore a promising approach for the scientific interpretation of future 21 cm observation data.
Modeling of large populations of binary stellar systems is an integral part of many areas of astrophysics, from radio pulsars and supernovae to X-ray binaries, gamma-ray bursts, and gravitational-wave mergers. Binary population synthesis codes that employ self-consistently the most advanced physics treatment available for stellar interiors and their evolution and are at the same time computationally tractable have started to emerge only recently. One element that is still missing from these codes is the ability to generate the complete time evolution of binaries with arbitrary initial conditions using precomputed three-dimensional grids of binary sequences. Here, we present a highly interpretable method, from binary evolution track interpolation. Our method implements simulation generation from irregularly sampled time series. Our results indicate that this method is appropriate for applications within binary population synthesis and computational astrophysics with time-dependent simulations in general. Furthermore, we point out and offer solutions to the difficulty surrounding evaluating the performance of signals exhibiting extreme morphologies akin to discontinuities.
| Article Article Article |
|
| May 7, 2025 Works-in-Progress Wednesday 11:20–11:55 a.m. Hancock, 35th Floor |
|
Ugur Demir | Northwestern University |
TBA.
| — |
|
| 10:40–11:20 a.m. |
|
Roxie (Ruoxi) Jiang | The University of Chicago |
TBA.
| — |
|
| 10:00–10:40 a.m. |
|
Tri Nguyen | Northwestern University |
TBA.
| — |
|
| April 30, 2025 Journal Club 11:00–11:55 a.m. Hancock, 4010 |
![]()
|
Elizabeth Teng Kyle Rocha |
Northwestern University |
Knowledge about the internal physical structure of stars is crucial to understanding their evolution. The novel binary population synthesis code POSYDON includes a module for interpolating the stellar and binary properties of any system at the end of binary MESA evolution based on a pre-computed set of models. In this work, we present a new emulation method for predicting stellar profiles, i.e., the internal stellar structure along the radial axis, using machine learning techniques. We use principal component analysis for dimensionality reduction and fully-connected feed-forward neural networks for making predictions. We find accuracy to be comparable to that of nearest neighbor approximation, with a strong advantage in terms of memory and storage efficiency. By providing a versatile framework for modeling stellar internal structure, the emulation method presented here will enable faster simulations of higher physical fidelity, offering a foundation for a wide range of large-scale population studies of stellar and binary evolution.
Dusty stellar point sources are a significant stage in stellar evolution and contribute to the metal enrichment of galaxies. These objects can be classified using photometric and spectroscopic observations with color-magnitude diagrams (CMD) and infrared excesses in spectral energy distributions (SED). We employed supervised machine learning spectral classification to categorize dusty stellar sources, including young stellar objects (YSOs) and evolved stars (oxygen- and carbon-rich asymptotic giant branch stars, AGBs), red supergiants (RSGs), and post-AGB (PAGB) stars in the Large and Small Magellanic Clouds, based on spectroscopic labeled data from the Surveying the Agents of Galaxy Evolution (SAGE) project, which used 12 multiwavelength filters and 618 stellar objects. Despite missing values and uncertainties in the SAGE spectral datasets, we achieved accurate classifications. To address small and imbalanced spectral catalogs, we used the Synthetic Minority Oversampling Technique (SMOTE) to generate synthetic data points. Among models applied before and after data augmentation, the Probabilistic Random Forest (PRF), a tuned Random Forest (RF), achieved the highest total accuracy, reaching 89% based on recall in categorizing dusty stellar sources. Using SMOTE does not improve the best model's accuracy for the CAGB, PAGB, and RSG classes; it remains 100%, 100%, and 88%, respectively, but shows variations for OAGB and YSO classes. We also collected photometric labeled data similar to the training dataset, classifying them using the top four PRF models with over 87% accuracy. Multiwavelength data from several studies were classified using a consensus model integrating four top models to present common labels as final predictions.
| Article Article |
April 23, 2025 Works-in-Progress Wednesday 11:20–11:55 a.m. Hancock, 4010 |
|
Sneh Pandya | Northeastern University |
TBA.
| — |
| 10:40–11:20 a.m. |
|
Jennifer Li | University of Illinois Urbana-Champaign |
TBA.
| — |
|
| 10:00–10:40 a.m. |
|
Anarya Ray | Northwestern University |
TBA.
| — |
|
| April 16, 2025 Journal Club 11:00–11:55 a.m. Hancock, 4010 |
![]()
|
Anowar J. Shajib Tri Nguyen |
The University of Chicago Northwestern University |
Strong gravitational lensing is a powerful tool for probing the internal structure and evolution of galaxies, the nature of dark matter, and the expansion history of the Universe, among many other scientific applications. For almost all of these science cases, modeling the lensing mass distribution is essential. For that, forward modeling of imaging data to the pixel level is the standard method used for galaxy-scale lenses. However, the traditional workflow of forward lens modeling necessitates a significant amount of human investigator time, requiring iterative tweaking and tuning of the model settings through trial and error. An automated lens modeling pipeline can substantially reduce the need for human investigator time. In this paper, we present \\textsc{dolphin}, an automated lens modeling pipeline that combines artificial intelligence with the traditional forward modeling framework to enable full automation of the modeling workflow. \\textsc{dolphin} uses a neural network model to perform visual recognition of the strong lens components, then autonomously sets up a lens model with appropriate complexity, and fits the model with the modeling engine, lenstronomy. Thanks to the versatility of lenstronomy, dolphin can autonomously model both galaxy-galaxy and galaxy-quasar strong lenses.
The phase space of stellar streams is proposed to detect dark substructure in the Milky Way through the perturbations created by passing subhalos - and thus is a powerful test of the cold dark matter paradigm and its alternatives. Using graph convolutional neural network (GCNN) data compression and simulation-based inference (SBI) on a simulated GD-1-like stream, we improve the constraint on the mass of a [108, 107, 106] M⊙ perturbing subhalo by factors of [11, 7, 3] with respect to the current state-of-the-art density power spectrum analysis. We find that the GCNN produces posteriors that are more accurate (better calibrated) than the power spectrum. We simulate the positions and velocities of stars in a GD-1-like stream and perturb the stream with subhalos of varying mass and velocity. Leveraging the feature encoding of the GCNN to compress the input phase space data, we then use SBI to estimate the joint posterior of the subhalo mass and velocity. We investigate how our results scale with the size of the GCNN, the coordinate system of the input and the effect of incomplete observations. Our results suggest that a survey with 10× fewer stars (300 stars) with complete 6-D phase space data performs about as well as a deeper survey (3000 stars) with only 3-D data (photometry, spectroscopy). The stronger constraining power and more accurate posterior estimation motivate further development of GCNNs in combining future photometric, spectroscopic and astrometric stream observations.
| Article Article |
|
| April 9, 2025 Works-in-Progress Wednesday 11:00–11:55 a.m. Hancock, 4010 |
![]()
|
Aleksandra Ćiprijanović Sandeep Madireddy |
Fermilab Argonne National Laboratory |
TBA.
| — | |
| 10:00–11:00 a.m. |
![]()
|
Rebecca Willett Chihway Chang |
The University of Chicago |
TBA.
| — | |
| April 4, 2025 Colloquium 10:30–11:55 a.m. Hancock, 4010 |
|
Alyssa Goodman | Harvard University |
I have been lucky enough in my career so far to watch, and I hope help, computational technology change what we can learn about our Universe. Today, in 2025, I somewhat unexpectedly find myself involved in a broad range of AI-based and AI-enhanced efforts designed to speed learning and discovery in astrophysics and in science. In this talk, I will offer glimpses into a handful of ongoing AI-enhanced efforts, each of which is very different from the others, yet which work together in a researcher/educator's life to speed progress. Work to be highlighted includes: (1) automated data-set linking in the “glue” and LIVE-Environments visualization environments; (2) The “Reading Time Machine,” which uses AI to “read” graphics and images and ingest their content into the ADS Literature archive, as “data”; (3) approaches to 3D selection in volumetric data, using both AI and augmented reality (AR); (4) a quest to understand why LLMs are so good at describing infographics, but so terrible at creating them; (5) capabilities of AI for writing code for visualization, in both research and education. The plan of the talk will be to present an overview of each of these efforts in order to inspire broader discussion of whichever topics evolve as most interesting to the assembled audience.
| ![]() |
|
| April 02, 2025 Journal Club 11:00–11:55 a.m. Hancock, 4010 |
![]()
|
George Winstone Cliff Johnson |
Northwestern University |
Gravitational waves, detected a century after they were first theorized, are spacetime distortions caused by some of the most cataclysmic events in the universe, including black hole mergers and supernovae. The successful detection of these waves has been made possible by ingenious detectors designed by human experts. Beyond these successful designs, the vast space of experimental configurations remains largely unexplored, offering an exciting territory potentially rich in innovative and unconventional detection strategies. Here, we demonstrate the application of artificial intelligence (AI) to systematically explore this enormous space, revealing novel topologies for gravitational wave (GW) detectors that outperform current next-generation designs under realistic experimental constraints. Our results span a broad range of astrophysical targets, such as black hole and neutron star mergers, supernovae, and primordial GW sources. Moreover, we are able to conceptualize the initially unorthodox discovered designs, emphasizing the potential of using AI algorithms not only in discovering but also in understanding these novel topologies. We've assembled more than 50 superior solutions in a publicly available Gravitational Wave Detector Zoo which could lead to many new surprising techniques. At a bigger picture, our approach is not limited to gravitational wave detectors and can be extended to AI-driven design of experiments across diverse domains of fundamental physics.
We present a catalogue of 497 galaxy-galaxy strong lenses in the Euclid Quick Release 1 data (63 deg2). In the initial 0.45\\% of Euclid's surveys, we double the total number of known lens candidates with space-based imaging. Our catalogue includes 250 grade A candidates, the vast majority of which (243) were previously unpublished. Euclid's resolution reveals rare lens configurations of scientific value including double-source-plane lenses, edge-on lenses, complete Einstein rings, and quadruply-imaged lenses. We resolve lenses with small Einstein radii ($\\theta_{\\rm E} < \\ang{;;1}$) in large numbers for the first time. These lenses are found through an initial sweep by deep learning models, followed by Space Warps citizen scientist inspection, expert vetting, and system-by-system modelling. Our search approach scales straightforwardly to Euclid Data Release 1 and, without changes, would yield approximately 7000 high-confidence (grade A or B) lens candidates by late 2026. Further extrapolating to the complete Euclid Wide Survey implies a likely yield of over 100000 high-confidence candidates, transforming strong lensing science.
| Article Article |
|
| March 26, 2025 Works-in-Progress Wednesday 11:20–11:55 a.m. Hancock, 4010 |
![]()
|
Alex Drlica-Wagner Aravindan Vijayaraghavan |
The University of Chicago Northwestern University |
TBA.
| — | |
| 10:00–10:50 a.m. |
![]()
|
Rina Barber Daniel Holz |
The University of Chicago |
TBA.
| — | |
| March 19, 2025 Journal Club 11:00–11:55 a.m. Hancock, 4010 |
![]()
|
Tianao Li Xinfeng Xu |
Northwestern University |
Diffusion models recently proved to be remarkable priors for Bayesian inverse problems. However, training these models typically requires access to large amounts of clean data, which could prove difficult in some settings. In this work, we present a novel method based on the expectation-maximization algorithm for training diffusion models from incomplete and noisy observations only. Unlike previous works, our method leads to proper diffusion models, which is crucial for downstream tasks. As part of our method, we propose and motivate an improved posterior sampling scheme for unconditional diffusion models. We present empirical evidence supporting the effectiveness of our method.
Mergers of binary neutron stars emit signals in both the gravitational-wave (GW) and electromagnetic spectra. Famously, the 2017 multi-messenger observation of GW170817 led to scientific discoveries across cosmology, nuclear physics and gravity. Central to these results were the sky localization and distance obtained from the GW data, which, in the case of GW170817, helped to identify the associated electromagnetic transient, AT 2017gfo, 11 h after the GW signal. Fast analysis of GW data is critical for directing time-sensitive electromagnetic observations. However, owing to challenges arising from the length and complexity of signals, it is often necessary to make approximations that sacrifice accuracy. Here we present a machine-learning framework that performs complete binary neutron star inference in just 1 s without making any such approximations. Our approach enhances multi-messenger observations by providing: (1) accurate localization even before the merger; (2) improved localization precision by around 30% compared to approximate low-latency methods; and (3) detailed information on luminosity distance, inclination and masses, which can be used to prioritize expensive telescope time. Additionally, the flexibility and reduced cost of our method open new opportunities for equation-of-state studies. Finally, we demonstrate that our method scales to long signals, up to an hour in length, thus serving as a blueprint for data analysis for next-generation ground- and space-based detectors.
| Article Article |
|
| March 12, 2025 Colloquium 10:30–11:30 a.m. Hancock, 4010 |
|
Bill Gropp | University of Illinois Urbana-Champaign |
The end of Dennard or frequency scaling in computer processors nearly twenty years ago has caused a transformation in computing. Innovations in computer architecture have enabled continued improvements in performance, but at the cost of increasing software complexity. GPUs have been key in providing performance for many applications and have enabled the revolution in machine learning and AI. This talk will provide some background on the transformations in computing over the last two decades, describe NCSA and its approach to this revolution in computing, and close with a description of NCSA's efforts in AI.
| — | |
| March 3, 2025 Colloquium 11:00–11:55 a.m. Hancock, 4010 |
|
Renée Hložek | University of Toronto |
In the sky: the Legacy Survey of Space and Time (LSST) on the Vera C. Rubin Observatory will generate a data deluge: millions of astronomical transients and variable sources will need to be classified from their light curves. I'll discuss the efforts within the Dark Energy Science Collaboration (DESC) to get ready for transient classification through efforts like public Photometric LSST Astronomical Time-series Classification Challenge (PLAsTiCC) and the Extended LSST Astronomical Time-series Classification Challenge (ELAsTiCC) was an expert challenge to LSST broker teams themselves to classify alert streams. I'll place this work in the context of pushing from detections to cosmology. Looking to the brain: I'll present AstroBEATS, a pipeline derived from astronomical image analysis techniques and designed for high-resolution images of the brain. I'll describe how AstroBEATS can be used to study the synaptic firing in the brain and to search for signs of neurodegeneration and describe the processes that generated this interdisciplinary research.
| ![]() |
|
| February 12, 2025 Works-in-Progress Wednesday 11:20–11:55 a.m. Hancock, 4010 |
![]()
|
Jeff McMahon Ermin Wei |
The University of Chicago Northwestern University |
TBA.
| — | |
| 10:45–11:20 a.m. |
![]()
|
Vicky Kalogera Aggelos Katsaggelos |
Northwestern University |
TBA.
| — | |
| 10:00–10:45 a.m. |
![]()
|
Emma Alexander Tjitske Starkenburg |
Northwestern University |
TBA.
| — | |
SkAI Institute
875 N. Michigan Ave., Suite 3500
Chicago, IL 60611
skai-institute@u.northwestern.edu
(Get to the SkAI Hub via the 172 E. Chestnut St. entrance)








































































