Events

SkAI Institute Events

SkAI Institute events are held at the SkAI Hub in the John Hancock Center, Suite 3500, unless otherwise indicated, and are in Central (Standard) Time. If you have any questions about the events, please contact Emma Alexander (ealexander at northwestern.edu) and Alex Ćiprijanović (aleksand at fnal.gov) and, for logistics, please contact SkAI Operations Coordinator, Gema Tinoco (gema.tinoco at northwestern.edu). For questions about a Journal Club meeting, please contact Xinfeng Xu (xinfeng.xu at northwestern.edu).

All SkAI events are presented in a hybrid (in-person and Zoom) format. Whether you plan to attend in person or virtually, please RSVP using the registration forms below, which will help us to get an accurate head count for catering and reporting purposes, and to make sure a visitor pass is requested for everyone who needs one. Please note that participants agree to follow the SkAI Institute Code of Conduct.

First time visiting the SkAI Hub? We have a brief orientation to help you know what to expect.


Location: John Hancock Center, Suite 3500 (unless otherwise indicated)
Zoom: Link

Registration Forms:
Registration Form for February 4 SkAI Works-In-Progress Wednesday
Registration Form for February 18 SkAI Colloquium

Upcoming Events:
Click on the speaker’s photo or name to learn more about them. Click on the title to reveal the abstract.

Date/Location Speaker Affiliation Title Links
Upcoming
February 4, 2026
Works-In-Progress Wednesday
11:00–11:55 a.m.
Hancock, 3500

Vicky Kalogera


Aggelos Katsaggelos
Northwestern University
TBA.
    —
February 11, 2026
Works-In-Progress Wednesday
11:00–11:55 a.m.
Hancock, 3500

Billy Moses


Joaquin Vieira
University of Illinois Urbana-Champaign
TBA.
    —
February 18, 2026
Colloquium
11:00–11:55 a.m.
Hancock, 3500
Berthy Feng Massachusetts Institute of Technology
Computational imaging aims to visualize scientific phenomena beyond the reach of conventional optics by incorporating assumptions, or priors, about the object being imaged. In the age of AI, the priors available to us are more sophisticated than ever. However, it is yet unclear how to use these priors rigorously, especially when using imaging in the scientific process. The main challenge of computational imaging for science is discerning when to use what prior and how. The art of imaging lies in deciding the correct balance of assumptions to obtain trustworthy and informative images. Once the prior has been decided, the question is how to incorporate it rigorously.
In this talk, I will present my previous work on building principled routes for incorporating data-driven and physics-based priors. On the data-driven side, I will show results of re-imagining the famous M87 black hole from real data with score-based priors. On the physics-based side, I will show how we have tackled extremely underdetermined imaging problems by enforcing physics constraints, including the problem of single-viewpoint dynamic tomography of emission near a black hole. In the intersection of AI and physics, I will present neural approximate mirror maps, a way to enforce physics constraints on generative models. I will then discuss future directions for further taming priors so that we can rigorously create, interpret, and extract insights from scientific images.
    —
March 11, 2026
Works-in-Progress Wednesday
10:00–11:00 a.m.
Hancock, 3500

Adam Miller


Han Liu
Northwestern University
TBA.
    —
11:00–11:55 a.m
Ermin Wei


Jeff McMahon
Northwestern University


The University of Chicago
TBA.
    —
March 18, 2026
Colloquium
11:00–11:55 a.m.
Hancock, 3500
Yuanyuan Shi University of California, San Diego
TBA.
    —
March 25, 2026
Works-In-Progress Wednesday
11:00–11:55 a.m.
Hancock, 3500


Alex Drlica-Wagner


Aravindan Vijayaraghavan


Paul Chichura
The University of Chicago



Northwestern University



SkAI Institute
TBA.
    —
April 1, 2026
Works-In-Progress Wednesday
11:00–11:55 a.m.
Hancock, 3500

Rina Barber


Daniel Holz
The University of Chicago
TBA.
    —
April 8, 2026
Works-in-Progress Wednesday
11:00–11:55 a.m.
Hancock, 3500

Rebecca Willett


Chihway Chang
The University of Chicago
TBA.
    —
April 15, 2026
Colloquium
11:00–11:55 a.m.
Hancock, 3500
Keith Bechtol University of Wisconsin–Madison
TBA.
    —
April 22, 2026
Works-in-Progress Wednesday
10:00–11:00 a.m.
Hancock, 3500

Aleksandra Ćiprijanović


Sandeep Madireddy
Fermilab



Argonne National Laboratory
TBA.
    —
11:00–11:55 a.m Tjitske Starkenburg


Emma Alexander
Northwestern University
TBA.
    —
May 6, 2026
Works-In-Progress Wednesday
11:00–11:55 a.m.
Hancock, 3500

Noelle Samia


Gautham Narayan
Northwestern University


University of Illinois Urbana-Champaign
TBA.
    —
May 13, 2026
Colloquium
11:00–11:55 a.m.
Hancock, 3500
Tom Downes Microsoft AI
TBA.
    —
Past
January 14, 2025
Colloquium
11:00–11:55 a.m.
Hancock, 3500
David Hogg New York University
Large language models (LLMs) are beginning to obtain the ability to design, execute, write up, and referee scientific projects on the data-science side of astrophysics. What implications does this have for our profession? I will try to elucidate a set of facts or points of agreement about what astrophysics is, or should be. I will also deliver a set of putative benefits that astrophysics perhaps brings to us, and to science, and to universities, and to the world. My goal is to deliver some kind of take on why astronomy and astrophysics are worth doing, and how LLMs might fit into a practice with that justification
    —
December 17, 2025
Works-in-Progress Wednesday
10:00–11:00 a.m.
Hancock, 3500

Aleksandra Ćiprijanović


Sandeep Madireddy
Fermilab



Argonne National Laboratory
TBA.
    —
11:00–11:55 a.m.

Alex Drlica-Wagner


Aravindan Vijayaraghavan


Paul Chichura
The University of Chicago



Northwestern University



SkAI Institute
TBA.
    —
December 10, 2025
SkAI AI Ethics Tutorial
10:00–11:00 a.m.
Hancock, 3500
Lu Cheng University of Illinois Chicago
Generative foundation models (GenFMs), including large language and multimodal models, are transforming information retrieval and knowledge management. However, their rapid adoption raises urgent concerns about social responsibility, trustworthiness, and governance. This mini-tutorial offers an overview of recent advances in responsible GenFMs, covering foundational concepts and multidimensional risk taxonomies (including safety, privacy, robustness, truthfulness, fairness, and machine ethics).
    —
December 3, 2025
Journal Club
10:00–11:00 a.m.
Hancock, 3500
Yiping Lu Northwestern University
Scaling scientific machine learning (SciML) requires overcoming bottlenecks at both training and inference. On the training side, we study the statistical convergence rate and limits of deep learning for solving elliptic PDEs from random samples. While our theory predicts optimal polynomial convergence for PINNs, optimization becomes prohibitively ill-conditioned as networks widen. By adapting descent strategies to the optimization geometry, we obtain scale-invariant training dynamics that translate polynomial convergence into concrete compute and yield compute-optimal configurations. On the inference side, I will introduce Simulation-Calibrated SciML (SCaSML), a physics-informed post-processing framework that improves surrogate models without retraining or fine-tuning. By enforcing physical laws, SCaSML delivers trustworthy corrections (via Feynman-Kac simulation) with approximate confidence intervals, achieves faster and near-optimal convergence rates, and supports online updates for digital twins. Together, these results integrate theory and numerics to enable predictable, reliable scaling of SciML in both training and inference.
    —
10:00–11:00 a.m.
Lindsay House



Benjamin Remy
SkAI Institute
We are merging a large participatory science effort with machine learning to enhance the Hobby-Eberly Telescope Dark Energy Experiment (HETDEX). Our overall goal is to remove false positives, allowing us to use lower signal-to-noise data and sources with low goodness-of-fit. With six million classifications through Dark Energy Explorers, we can confidently determine if a source is not real at over 94% confidence level when classified by at least 10 individuals; this confidence level increases for higher signal-to-noise sources. To date, we have only been able to apply this direct analysis to 190,000 sources. The full sample of HETDEX will contain around 2–3 million sources, including nearby galaxies ([O II] emitters), distant galaxies (Lyα emitters or LAEs), false positives, and contamination from instrument issues. We can accommodate this tenfold increase by using machine learning with visually vetted samples from Dark Energy Explorers. We have already increased by over tenfold the number of sources that have been visually vetted from our previous pilot study where we only had 14,000 visually vetted LAE candidates. This paper expands on the previous work by increasing the visually vetted sample from 14,000 to 190,000. In addition, using our currently visually vetted sample, we generate a real or false positive classification for the full candidate sample of 1.2 million LAEs. We currently have approximately 17,000 volunteers from 159 countries around the world. Thus, we are applying participatory or citizen scientist analysis to our full HETDEX data set, creating a free educational opportunity that requires no prior technical knowledge.



Context. Weak lensing mass-mapping is a useful tool for accessing the full distribution of dark matter on the sky, but because of intrinsic galaxy ellipticies, finite fields, and missing data, the recovery of dark matter maps constitutes a challenging, ill-posed inverse problem.
Aims. We introduce a novel methodology that enables the efficient sampling of the high-dimensional Bayesian posterior of the weak lensing mass-mapping problem, relying on simulations to define a fully non-Gaussian prior. We aim to demonstrate the accuracy of the method to simulated fields, and then proceed to apply it to the mass reconstruction of the HST/ACS COSMOS field.
Methods. The proposed methodology combines elements of Bayesian statistics, analytic theory, and a recent class of deep generative models based on neural score matching. This approach allows us to make full use of analytic cosmological theory to constrain the 2pt statistics of the solution, to understand any differences between this analytic prior and full simulations from cosmological simulations, and to obtain samples from the full Bayesian posterior of the problem for robust uncertainty quantification.
Results. We demonstrate the method in the κTNG simulations and find that the posterior mean significantly outperforms previous methods (Kaiser–Squires, Wiener filter, Sparsity priors) both for the root-mean-square error and in terms of the Pearson correlation. We further illustrate the interpretability of the recovered posterior by establishing a close correlation between posterior convergence values and the S/N of the clusters artificially introduced into a field. Finally, we apply the method to the reconstruction of the HST/ACS COSMOS field, which yields the highest-quality convergence map of this field to date.
Conclusions. We find the proposed approach to be superior to previous algorithms, scalable, providing uncertainties, and using a fully non-Gaussian prior.
Article



Article
November 19, 2025
Works-in-Progress Wednesday
10:00–11:00 a.m.
Hancock, 3500


Rina Barber



Daniel Holz



Ruiting Liang
The University of Chicago
TBA.
    —
10:00–11:00 a.m.

Roy Zhao



Shi-Hui Zang



Benjamin Remy





The University of Chicago



University of Wisconsin–Madison



SkAI Institute

Additional speakers:
Yuuki Omori, Shrihan Agarwal, Georgios Valogiannis,
Rebecca Willett, Chihway Chang
TBA.
    —
November 12, 2025
Seminar and Colloquium
11:00–11:55 a.m.
Hancock, 3500
Yuan-Sen Ting The Ohio State University
The expansive, interdisciplinary nature of astronomy, combined with its open-access culture, makes it an ideal testing ground for exploring how Large Language Models (LLMs) can accelerate scientific discovery. Recent developments in LLM reasoning capabilities have shown substantial progress—our work demonstrates that AI agents can now achieve gold medal performance on International Olympiad on Astronomy and Astrophysics (IOAA) problems, indicating their growing analytical abilities. In this talk, I will present our recent advances in applying LLMs as agents to real-world astronomical challenges. Through self-play reinforcement learning, we demonstrate how LLM agents can conduct end-to-end research tasks in galaxy spectral fitting, encompassing data analysis, strategy refinement, and outlier detection—approaching capabilities similar to human intuition and domain knowledge. However, limitations remain. While autonomous research agents like Mephisto could theoretically help analyze all observed sources, the cost of closed-source solutions remains prohibitive for large-scale applications involving billions of objects. Additionally, the Moravec paradox manifests clearly in astronomy: tasks requiring abstract reasoning may be easier for AI than seemingly simple perceptual tasks. Current models still struggle with chart reading, multimodal data interpretation, and other fundamental astronomical workflows. To address the cost limitation, we developed lightweight, open-source specialized models (AstroSage and AstroLLaMA)—trained on arXiv literature—and evaluated them against carefully curated astronomical benchmarks. Our research shows that these specialized LLMs can outperform larger general-purpose models on astronomy Q&A tasks when appropriately pretrained and fine-tuned, demonstrating a path forward for building more capable and accessible astronomy-specific models. Looking ahead, the path forward involves integrating more function-calling tools and building a comprehensive ecosystem—not just better models. The astronomical community's collaborative infrastructure will be important for scaling up automated inference and expanding the role of AI in astronomical research.
10:00–10:50 a.m.
Cliff Johnson



Michael Zevin
Northwestern University



Adler Planetarium
Over the past decade, the Zooniverse has grown into the most successful and widely used online platform for people-powered research. While one of the Zooniverse’s key strengths is its large community of nearly 3 million engaged volunteers who contribute their effort to performing research tasks, the platform has increasingly committed resources toward building and supporting machine-learning infrastructure. While human effort remains essential to the success of the platform’s projects, the combination of human and machine intelligence is becoming increasingly necessary to keep up with ever-increasing data volumes and complexity. In this talk, we review Zooniverse’s current strategies and tools for machine-learning integration, and look ahead at a number of exciting possibilities for new and enhanced AI/ML platform features. We wish to share ideas and solicit feedback from AI and Astro SkAI researchers regarding potential new directions, which include: anomaly detection and curation; latent space exploration; active learning, and rapid model refinement.
November 5, 2025
Journal Club
10:00–11:00 a.m.
Hancock, 3500

Paul Chichura



Christian Jespersen
SkAI Institute




Princeton University
We present improvements to the pointing accuracy of the South Pole Telescope (SPT) using machine learning. The ability of the SPT to point accurately at the sky is limited by its structural imperfections, which are impacted by the extreme weather at the South Pole. Pointing accuracy is particularly important during SPT participation in observing campaigns with the Event Horizon Telescope (EHT), which requires stricter accuracy than typical observations with the SPT. We compile a training dataset of historical observations of astronomical sources made with the SPT-3G and EHT receivers on the SPT. We train two XGBoost models to learn a mapping from current weather conditions to two telescope drive control arguments—one which corrects for errors in azimuth and the other for errors in elevation. Our trained models achieve root mean squared errors on withheld test data of 2.14″ in cross-elevation and 3.57″ in elevation, well below our goal of 5″ along each axis. We deploy our models on the telescope control system and perform further in situ test observations during the EHT observing campaign in 2024 April. Our models result in significantly improved pointing accuracy: for sources within the range of input variables where the models are best trained, average combined pointing error improved 33%, from 15.9″ to 10.6″. These improvements, while significant, fall shy of our ultimate goal, but they serve as a proof of concept for the development of future models. Planned upgrades to the EHT receiver on the SPT will necessitate even stricter pointing accuracy, which will be achievable with our methods.



Galaxies are often modelled as composites of separable components with distinct spectral signatures, implying that different wavelength ranges are only weakly correlated. They are not. We present a data-driven model which exploits subtle correlations between physical processes to accurately predict infrared (IR) WISE photometry from a neural summary of optical SDSS spectra. The model achieves accuracies of 2N ≈ 1 for all photometric bands in WISE, as well as good colors. We are also able to tightly constrain typically IR-derived properties, e.g. the bolometric luminosities of AGN and dust parameters such as qPAH. We find that current SED-fitting methods are incapable of making comparable predictions, and that model misspecification often leads to correlated biases in star-formation rates and AGN luminosities. To help improve SED models, we determine what features of the optical spectrum are responsible for our improved predictions, and identify several lines (CaII, SrII, FeI, [OII] and Hα), which point to the complex chronology of star formation and chemical enrichment being incorrectly modelled.
Article



Article
October 29, 2025
SkAI Industry Panel for Early-Career Scientists
10:00–11:30 a.m.
Hancock, 3500




Patrick Aleo



Frank Fineis



James Guillochon



Kara Ponder



Dimitrios Tanoglidis
Honeywell Federal Manufacturing & Technologies, LLC



Clearcover



Esri



Charles Schwab



Walgreens
TBA.
    —
October 22, 2025
Works-In-Progress Wednesday
11:00–11:55 a.m.
Hancock, 3500

Ermin Wei


Jeff McMahon
Northwestern University


The University of Chicago
TBA.
    —
10:00–11:00 a.m. Tjitske Starkenburg



Emma Alexander
Northwestern University
TBA.
    —
October 15, 2025
SkAI Onboarding Tutorial
10:00–11:00 a.m.
Hancock, 3500
Jennifer Li University of Illinois Urbana-Champaign
This onboarding tutorial will include an overview on UIUC’s Delta/DeltaAI computing resources, step-by-step instructions to request computing time on Delta/DeltaAI, and a getting-started guide on using Delta/DeltaAI. Jennifer will also give a quick overview of the National Center for Supercomputing Applications’ (NCSA) extensive online course on high-performance computing (HPC). This interactive tutorial will also outline how SkAI members can apply for compute resources.
    —
October 8, 2025
Works-In-Progress Wednesday
11:00–11:55 a.m.
Hancock, 3500

Adam Miller


Han Liu
Northwestern University
TBA.
    —
10:00–11:00 a.m.
Noelle Samia


Gautham Narayan
Northwestern University


University of Illinois Urbana-Champaign
TBA.
    —
October 1, 2025
Journal Club and Colloquium
11:10–11:55 a.m.
Hancock, 3500
Christopher Stubbs Harvard University
TBA.
    —

10:00–11:00 a.m.

Sydney Erickson



Yongseok Jo
Stanford University



SkAI Institute
Strongly lensed quasars can be used to constrain cosmological parameters through time-delay cosmography. Models of the lens masses are a necessary component of this analysis. To enable time-delay cosmography from a sample of O(103) lenses, which will soon become available from surveys like the Rubin Observatory’s Legacy Survey of Space and Time and the Euclid Wide Survey, we require fast and standardizable modeling techniques. To address this need, we apply neural posterior estimation (NPE) for modeling galaxy-scale strongly lensed quasars from the Strong Lensing Insights into the Dark Energy Survey (STRIDES) sample. NPE brings two advantages: speed and the ability to implicitly marginalize over nuisance parameters. We extend this method by employing sequential NPE to increase precision of mass model posteriors. We then fold individual lens models into a hierarchical Bayesian inference to recover the population distribution of lens mass parameters, accounting for out-of-distribution shift. After verifying our method using simulated analogs of the STRIDES lens sample, we apply our method to 14 Hubble Space Telescope single-filter observations. We find the population mean of the power-law elliptical mass distribution slope, γlens, to be Μγlens = 2.13 +/– 0.06. Our result represents the first population-level constraint for these systems. This population-level inference from fully automated modeling is an important stepping stone toward cosmological inference with large samples of strongly lensed quasars.




The rapid advancement of large-scale cosmological simulations has opened new avenues for cosmological and astrophysical research. However, the increasing diversity among cosmological simulation models presents a challenge to the robustness. In this work, we develop the Model-Insensitive ESTimator (MIEST), a machine that can robustly estimate the cosmological parameters, Ωm and σ8, from neural hydrogen maps of simulation models in the CAMELS project—TNG, SIMBA, ASTRID, and EAGLE. An estimator is considered robust if it possesses a consistent predictive power across all simulations, including those used during the training phase. We train our machine using multiple simulation models and ensure that it only extracts common features between the models while disregarding the model-specific features. This allows us to develop a novel model that is capable of accurately estimating parameters across a range of simulation models, without being biased towards any particular model. Upon the investigation of the latent space—a set of summary statistics, we find that the implementation of robustness leads to the blending of latent variables across different models, demonstrating the removal of model-specific features. In comparison to a standard machine lacking robustness, the average performance of MIEST on the unseen simulations during the training phase has been improved by ~17% for Ωm and ~38% for σ8. By using a machine learning approach that can extract robust, yet physical features, we hope to improve our understanding of galaxy formation and evolution in a (subgrid) model-insensitive manner, and ultimately, gain insight into the underlying physical processes responsible for robustness. This is a Learning the Universe publication.
Article



Article
September 17, 2025
Journal Club
10:00–11:00 a.m.
Hancock, 3500

Tri Nguyen



Jimena González
Northwestern University



SkAI Institute
TBA.




We conduct a search for strong gravitational lenses in the Dark Energy Survey (DES) Year 6 imaging data. We implement a pre-trained Vision Transformer (ViT) for our machine learning (ML) architecture and adopt Interactive Machine Learning to construct a training sample with multiple classes to address common types of false positives. Our ML model reduces 236 million DES cutout images to 22,564 targets of interest, including around 85% of previously reported galaxy-galaxy lens candidates discovered in DES. These targets were visually inspected by citizen scientists, who ruled out approximately 90% as false positives. Of the remaining 2,618 candidates, 149 were expert-classified as ‘definite’ lenses and 516 as ‘probable’ lenses, with 147 of these candidates being newly identified. Additionally, we trained a second ViT to find double-source plane lens systems, finding at least one double-source system. Our main ViT excels at identifying galaxy-galaxy lenses, consistently assigning high scores to candidates with high confidence. The top 800 ViT-scored images include around 100 of our ‘definite’ lens candidates. This selection is an order of magnitude higher in purity than previous convolutional neural network-based lens searches and demonstrates the feasibility of applying our methodology for discovering large samples of lenses in future surveys.
    —



Article
September 10, 2025
Works-in-Progress Wednesday
11:00–11:55 a.m.
Hancock, 3500

Billy Moses


Joaquin Vieira
University of Illinois Urbana-Champaign
TBA.
    —
10:00–11:00 a.m.
Vicky Kalogera


Aggelos Katsaggelos
Northwestern University
TBA.
    —
August 29, 2025
Journal Club and Featured Conversation
11:00–11:45 a.m.
Hancock, 3500
Andreas Berlind NSF Astronomy Division
TBA.
    —
10:00–11:00 a.m. Elizabeth Teng Northwestern University
TBA.
    —
May 28, 2025
Works-in-Progress Wednesday, SkAI Hub Orientation, and Journal Club
11:00–11:55 a.m.
Hancock, 4010

Aldana Grichener




Dalya Baron
The University of Arizona




Stanford University
One of the main challenges in modeling massive stars to the onset of core collapse is the computational bottleneck of nucleosynthesis during advanced burning stages. The number of isotopes formed requires solving a large set of fully-coupled stiff ordinary differential equations (ODEs), making the simulations computationally intensive and prone to numerical instability. To overcome this barrier, we design a nuclear neural network (NNN) framework with multiple hidden layers to emulate nucleosynthesis calculations and conduct a proof-of-concept to evaluate its performance. The NNN takes the temperature, density and composition of a burning region as input and predicts the resulting isotopic abundances along with the energy generation and loss rates. We generate training sets for initial conditions corresponding to oxygen core depletion and beyond using large nuclear reaction networks, and compare the predictions of the NNNs to results from a commonly used small net. We find that the NNNs improve the accuracy of the electron fraction by 280–660% and the nuclear energy generation by 250–750%, consistently outperforming the small network across all timesteps. They also achieve significantly better predictions of neutrino losses on relatively short timescales, with improvements ranging from 100–106%. While further work is needed to enhance their accuracy and applicability to different stellar conditions, integrating NNN trained models into stellar evolution codes is promising for facilitating large-scale generation of core-collapse supernova (CCSN) progenitors with higher physical fidelity.





The structure and chemistry of the dusty interstellar medium (ISM) are shaped by complex processes that depend on the local radiation field, gas composition, and dust grain properties. Of particular importance are polycyclic aromatic hydrocarbons (PAHs), which emit strong vibrational bands in the mid-infrared, and play a key role in the ISM energy balance. We recently identified global correlations between PAH band and optical line ratios across three nearby galaxies, suggesting a connection between PAH heating and gas ionization throughout the ISM. In this work, we perform a census of the PAH heating–gas ionization connection using ∼700,000 independent pixels that probe scales of 40–150 pc in 19 nearby star-forming galaxies from the PHANGS survey. We find a universal relation between PAH(11.3 μm/7.7 μm) and ([S II]/Hα) with a slope of ∼0.2 and a scatter of ∼0.025 dex. The only exception is a group of anomalous pixels that show unusually high (11.3 μm/7.7 μm) PAH ratios in regions with old stellar populations and high starlight-to-dust emission ratios. Their mid-infrared spectra resemble those of elliptical galaxies. Active galactic nucleus hosts show modestly steeper slopes, with a ∼10% increase in PAH(11.3 μm/7.7 μm) in the diffuse gas on kiloparsec scales. This universal relation implies an emerging simplicity in the complex ISM, with a sequence that is driven by a single varying property: the spectral shape of the interstellar radiation field. This suggests that other properties, such as gas-phase abundances, gas ionization parameter, and grain charge distribution, are relatively uniform in all but specific cases.
Article




Article
10:00–11:00 a.m.
SkAI Operations Team
Julian Cuevas-Zepeda




The University of Chicago



Northwestern University
TBA.





TBA.
    —




    —
May 21, 2025
Works-in-Progress Wednesday and Colloquium
10:00–11:55 a.m.
Hancock, 35th Floor


Weijian Li


Souvik Chakraborty
Northwestern University



Indian Institute of Technology (IIT) Delhi
TBA.



Operator learning is an emerging area in scientific machine learning which aims to learn mappings between infinite dimensional function spaces. In the first half of the talk, I will delve into the foundations of Wavelet Neural Operator (WNO), a recently developed operator learning algorithm. I will discuss its working principles and its potential applications in complex engineering problems including fracture propagation in materials, tumor detection using USG data and elastography, and climate modelling.
The second half of the talk will focus on what lies beyond neural operators. I will introduce a new scientific machine learning architecture that is loosely motivated from cognitive science. This architecture is a first of its kind foundation model and offers two key advantages: (i) it can simultaneously learn solution operators for multiple parametric PDEs, and (ii) rapid generalization to new parametric PDEs with minimal fine-tuning. We observe that the proposed architecture is robust against catastrophic forgetting and facilitate knowledge transfer across dissimilar tasks. Across a diverse array of mechanics problems, consistent performance enhancements are observed with this architecture compared to task-specific baseline operator learning frameworks.
    —


    —
May 14, 2025
Journal Club
11:00–11:55 a.m.
Hancock, 4010


Jason Sun




Philipp Rajah Moura Srivastava
Northwestern University
The first measurements of the 21 cm brightness temperature power spectrum from the epoch of reionization will very likely be achieved in the near future by radio interferometric array experiments such as the Hydrogen Epoch of Reionization Array (HERA) and the Square Kilometre Array (SKA). Standard MCMC analyses use an explicit likelihood approximation to infer the reionization parameters from the 21 cm power spectrum. In this paper, we present a new Bayesian inference of the reionization parameters where the likelihood is implicitly defined through forward simulations using density estimation likelihood-free inference (DELFI). Realistic effects, including thermal noise and foreground avoidance, are also applied to the mock observations from the HERA and SKA. We demonstrate that this method recovers accurate posterior distributions for the reionization parameters, and it outperforms the standard MCMC analysis in terms of the location and size of credible parameter regions. With the minute-level processing time once the network is trained, this technique is a promising approach for the scientific interpretation of future 21 cm power spectrum observation data. Our code 21cmDELFI-PS is publicly available at this link (https://github.com/Xiaosheng-Zhao/21cmDELFI).



Tomographic three-dimensional 21 cm images from the epoch of reionization contain a wealth of information about the reionization of the intergalactic medium by astrophysical sources. Conventional power spectrum analysis cannot exploit the full information in the 21 cm data because the 21 cm signal is highly non-Gaussian due to reionization patchiness. We perform a Bayesian inference of the reionization parameters where the likelihood is implicitly defined through forward simulations using density estimation likelihood-free inference (DELFI). We adopt a trained 3D convolutional neural network (CNN) to compress the 3D image data into informative summaries (DELFI-3D CNN). We show that this method recovers accurate posterior distributions for the reionization parameters. Our approach outperforms earlier analysis based on two-dimensional 21 cm images. In contrast, a Monte Carlo Markov Chain analysis of the 3D light-cone-based 21 cm power spectrum alone and using a standard explicit likelihood approximation results in less accurate credible parameter regions than inferred by the DELFI-3D CNN, both in terms of the location and shape of the contours. Our proof-of-concept study implies that the DELFI-3D CNN can effectively exploit more information in the 3D 21 cm images than a 2D CNN or power spectrum analysis. This technique can be readily extended to include realistic effects and is therefore a promising approach for the scientific interpretation of future 21 cm observation data.


Modeling of large populations of binary stellar systems is an integral part of many areas of astrophysics, from radio pulsars and supernovae to X-ray binaries, gamma-ray bursts, and gravitational-wave mergers. Binary population synthesis codes that employ self-consistently the most advanced physics treatment available for stellar interiors and their evolution and are at the same time computationally tractable have started to emerge only recently. One element that is still missing from these codes is the ability to generate the complete time evolution of binaries with arbitrary initial conditions using precomputed three-dimensional grids of binary sequences. Here, we present a highly interpretable method, from binary evolution track interpolation. Our method implements simulation generation from irregularly sampled time series. Our results indicate that this method is appropriate for applications within binary population synthesis and computational astrophysics with time-dependent simulations in general. Furthermore, we point out and offer solutions to the difficulty surrounding evaluating the performance of signals exhibiting extreme morphologies akin to discontinuities.
Article



Article


Article
May 7, 2025
Works-in-Progress Wednesday
11:20–11:55 a.m.
Hancock, 35th Floor
Ugur Demir Northwestern University
TBA.
    —
10:40–11:20 a.m. Roxie (Ruoxi) Jiang The University of Chicago
TBA.
    —
10:00–10:40 a.m. Tri Nguyen Northwestern University
TBA.
    —
April 30, 2025
Journal Club
11:00–11:55 a.m.
Hancock, 4010

Elizabeth Teng



Kyle Rocha
Northwestern University
Knowledge about the internal physical structure of stars is crucial to understanding their evolution. The novel binary population synthesis code POSYDON includes a module for interpolating the stellar and binary properties of any system at the end of binary MESA evolution based on a pre-computed set of models. In this work, we present a new emulation method for predicting stellar profiles, i.e., the internal stellar structure along the radial axis, using machine learning techniques. We use principal component analysis for dimensionality reduction and fully-connected feed-forward neural networks for making predictions. We find accuracy to be comparable to that of nearest neighbor approximation, with a strong advantage in terms of memory and storage efficiency. By providing a versatile framework for modeling stellar internal structure, the emulation method presented here will enable faster simulations of higher physical fidelity, offering a foundation for a wide range of large-scale population studies of stellar and binary evolution.



Dusty stellar point sources are a significant stage in stellar evolution and contribute to the metal enrichment of galaxies. These objects can be classified using photometric and spectroscopic observations with color-magnitude diagrams (CMD) and infrared excesses in spectral energy distributions (SED). We employed supervised machine learning spectral classification to categorize dusty stellar sources, including young stellar objects (YSOs) and evolved stars (oxygen- and carbon-rich asymptotic giant branch stars, AGBs), red supergiants (RSGs), and post-AGB (PAGB) stars in the Large and Small Magellanic Clouds, based on spectroscopic labeled data from the Surveying the Agents of Galaxy Evolution (SAGE) project, which used 12 multiwavelength filters and 618 stellar objects. Despite missing values and uncertainties in the SAGE spectral datasets, we achieved accurate classifications. To address small and imbalanced spectral catalogs, we used the Synthetic Minority Oversampling Technique (SMOTE) to generate synthetic data points. Among models applied before and after data augmentation, the Probabilistic Random Forest (PRF), a tuned Random Forest (RF), achieved the highest total accuracy, reaching 89% based on recall in categorizing dusty stellar sources. Using SMOTE does not improve the best model's accuracy for the CAGB, PAGB, and RSG classes; it remains 100%, 100%, and 88%, respectively, but shows variations for OAGB and YSO classes. We also collected photometric labeled data similar to the training dataset, classifying them using the top four PRF models with over 87% accuracy. Multiwavelength data from several studies were classified using a consensus model integrating four top models to present common labels as final predictions.
Article


Article
April 23, 2025
Works-in-Progress Wednesday
11:20–11:55 a.m.
Hancock, 4010
Sneh Pandya Northeastern University
TBA.
    —
10:40–11:20 a.m. Jennifer Li University of Illinois Urbana-Champaign
TBA.
    —
10:00–10:40 a.m. Anarya Ray Northwestern University
TBA.
    —
April 16, 2025
Journal Club
11:00–11:55 a.m.
Hancock, 4010

Anowar J. Shajib



Tri Nguyen
The University of Chicago



Northwestern University
Strong gravitational lensing is a powerful tool for probing the internal structure and evolution of galaxies, the nature of dark matter, and the expansion history of the Universe, among many other scientific applications. For almost all of these science cases, modeling the lensing mass distribution is essential. For that, forward modeling of imaging data to the pixel level is the standard method used for galaxy-scale lenses. However, the traditional workflow of forward lens modeling necessitates a significant amount of human investigator time, requiring iterative tweaking and tuning of the model settings through trial and error. An automated lens modeling pipeline can substantially reduce the need for human investigator time. In this paper, we present \\textsc{dolphin}, an automated lens modeling pipeline that combines artificial intelligence with the traditional forward modeling framework to enable full automation of the modeling workflow. \\textsc{dolphin} uses a neural network model to perform visual recognition of the strong lens components, then autonomously sets up a lens model with appropriate complexity, and fits the model with the modeling engine, lenstronomy. Thanks to the versatility of lenstronomy, dolphin can autonomously model both galaxy-galaxy and galaxy-quasar strong lenses.



The phase space of stellar streams is proposed to detect dark substructure in the Milky Way through the perturbations created by passing subhalos - and thus is a powerful test of the cold dark matter paradigm and its alternatives. Using graph convolutional neural network (GCNN) data compression and simulation-based inference (SBI) on a simulated GD-1-like stream, we improve the constraint on the mass of a [108, 107, 106] M perturbing subhalo by factors of [11, 7, 3] with respect to the current state-of-the-art density power spectrum analysis. We find that the GCNN produces posteriors that are more accurate (better calibrated) than the power spectrum. We simulate the positions and velocities of stars in a GD-1-like stream and perturb the stream with subhalos of varying mass and velocity. Leveraging the feature encoding of the GCNN to compress the input phase space data, we then use SBI to estimate the joint posterior of the subhalo mass and velocity. We investigate how our results scale with the size of the GCNN, the coordinate system of the input and the effect of incomplete observations. Our results suggest that a survey with 10× fewer stars (300 stars) with complete 6-D phase space data performs about as well as a deeper survey (3000 stars) with only 3-D data (photometry, spectroscopy). The stronger constraining power and more accurate posterior estimation motivate further development of GCNNs in combining future photometric, spectroscopic and astrometric stream observations.
Article


Article
April 9, 2025
Works-in-Progress Wednesday
11:00–11:55 a.m.
Hancock, 4010

Aleksandra Ćiprijanović


Sandeep Madireddy
Fermilab



Argonne National Laboratory
TBA.
    —
10:00–11:00 a.m.
Rebecca Willett


Chihway Chang
The University of Chicago
TBA.
    —
April 4, 2025
Colloquium
10:30–11:55 a.m.
Hancock, 4010
Alyssa Goodman Harvard University
I have been lucky enough in my career so far to watch, and I hope help, computational technology change what we can learn about our Universe. Today, in 2025, I somewhat unexpectedly find myself involved in a broad range of AI-based and AI-enhanced efforts designed to speed learning and discovery in astrophysics and in science. In this talk, I will offer glimpses into a handful of ongoing AI-enhanced efforts, each of which is very different from the others, yet which work together in a researcher/educator's life to speed progress. Work to be highlighted includes: (1) automated data-set linking in the “glue” and LIVE-Environments visualization environments; (2) The “Reading Time Machine,” which uses AI to “read” graphics and images and ingest their content into the ADS Literature archive, as “data”; (3) approaches to 3D selection in volumetric data, using both AI and augmented reality (AR); (4) a quest to understand why LLMs are so good at describing infographics, but so terrible at creating them; (5) capabilities of AI for writing code for visualization, in both research and education. The plan of the talk will be to present an overview of each of these efforts in order to inspire broader discussion of whichever topics evolve as most interesting to the assembled audience.
April 02, 2025
Journal Club
11:00–11:55 a.m.
Hancock, 4010

George Winstone


Cliff Johnson
Northwestern University
Gravitational waves, detected a century after they were first theorized, are spacetime distortions caused by some of the most cataclysmic events in the universe, including black hole mergers and supernovae. The successful detection of these waves has been made possible by ingenious detectors designed by human experts. Beyond these successful designs, the vast space of experimental configurations remains largely unexplored, offering an exciting territory potentially rich in innovative and unconventional detection strategies. Here, we demonstrate the application of artificial intelligence (AI) to systematically explore this enormous space, revealing novel topologies for gravitational wave (GW) detectors that outperform current next-generation designs under realistic experimental constraints. Our results span a broad range of astrophysical targets, such as black hole and neutron star mergers, supernovae, and primordial GW sources. Moreover, we are able to conceptualize the initially unorthodox discovered designs, emphasizing the potential of using AI algorithms not only in discovering but also in understanding these novel topologies. We've assembled more than 50 superior solutions in a publicly available Gravitational Wave Detector Zoo which could lead to many new surprising techniques. At a bigger picture, our approach is not limited to gravitational wave detectors and can be extended to AI-driven design of experiments across diverse domains of fundamental physics.



We present a catalogue of 497 galaxy-galaxy strong lenses in the Euclid Quick Release 1 data (63 deg2). In the initial 0.45\\% of Euclid's surveys, we double the total number of known lens candidates with space-based imaging. Our catalogue includes 250 grade A candidates, the vast majority of which (243) were previously unpublished. Euclid's resolution reveals rare lens configurations of scientific value including double-source-plane lenses, edge-on lenses, complete Einstein rings, and quadruply-imaged lenses. We resolve lenses with small Einstein radii ($\\theta_{\\rm E} < \\ang{;;1}$) in large numbers for the first time. These lenses are found through an initial sweep by deep learning models, followed by Space Warps citizen scientist inspection, expert vetting, and system-by-system modelling. Our search approach scales straightforwardly to Euclid Data Release 1 and, without changes, would yield approximately 7000 high-confidence (grade A or B) lens candidates by late 2026. Further extrapolating to the complete Euclid Wide Survey implies a likely yield of over 100000 high-confidence candidates, transforming strong lensing science.
Article


Article
March 26, 2025
Works-in-Progress Wednesday
11:20–11:55 a.m.
Hancock, 4010

Alex Drlica-Wagner


Aravindan Vijayaraghavan
The University of Chicago



Northwestern University
TBA.
    —
10:00–10:50 a.m.
Rina Barber


Daniel Holz
The University of Chicago
TBA.
    —
March 19, 2025
Journal Club
11:00–11:55 a.m.
Hancock, 4010

Tianao Li


Xinfeng Xu
Northwestern University
Diffusion models recently proved to be remarkable priors for Bayesian inverse problems. However, training these models typically requires access to large amounts of clean data, which could prove difficult in some settings. In this work, we present a novel method based on the expectation-maximization algorithm for training diffusion models from incomplete and noisy observations only. Unlike previous works, our method leads to proper diffusion models, which is crucial for downstream tasks. As part of our method, we propose and motivate an improved posterior sampling scheme for unconditional diffusion models. We present empirical evidence supporting the effectiveness of our method.



Mergers of binary neutron stars emit signals in both the gravitational-wave (GW) and electromagnetic spectra. Famously, the 2017 multi-messenger observation of GW170817 led to scientific discoveries across cosmology, nuclear physics and gravity. Central to these results were the sky localization and distance obtained from the GW data, which, in the case of GW170817, helped to identify the associated electromagnetic transient, AT 2017gfo, 11 h after the GW signal. Fast analysis of GW data is critical for directing time-sensitive electromagnetic observations. However, owing to challenges arising from the length and complexity of signals, it is often necessary to make approximations that sacrifice accuracy. Here we present a machine-learning framework that performs complete binary neutron star inference in just 1 s without making any such approximations. Our approach enhances multi-messenger observations by providing: (1) accurate localization even before the merger; (2) improved localization precision by around 30% compared to approximate low-latency methods; and (3) detailed information on luminosity distance, inclination and masses, which can be used to prioritize expensive telescope time. Additionally, the flexibility and reduced cost of our method open new opportunities for equation-of-state studies. Finally, we demonstrate that our method scales to long signals, up to an hour in length, thus serving as a blueprint for data analysis for next-generation ground- and space-based detectors.
Article


Article
March 12, 2025
Colloquium
10:30–11:30 a.m.
Hancock, 4010
Bill Gropp University of Illinois Urbana-Champaign
The end of Dennard or frequency scaling in computer processors nearly twenty years ago has caused a transformation in computing. Innovations in computer architecture have enabled continued improvements in performance, but at the cost of increasing software complexity. GPUs have been key in providing performance for many applications and have enabled the revolution in machine learning and AI. This talk will provide some background on the transformations in computing over the last two decades, describe NCSA and its approach to this revolution in computing, and close with a description of NCSA's efforts in AI.
    —
March 3, 2025
Colloquium
11:00–11:55 a.m.
Hancock, 4010
Renée Hložek University of Toronto
In the sky: the Legacy Survey of Space and Time (LSST) on the Vera C. Rubin Observatory will generate a data deluge: millions of astronomical transients and variable sources will need to be classified from their light curves. I'll discuss the efforts within the Dark Energy Science Collaboration (DESC) to get ready for transient classification through efforts like public Photometric LSST Astronomical Time-series Classification Challenge (PLAsTiCC) and the Extended LSST Astronomical Time-series Classification Challenge (ELAsTiCC) was an expert challenge to LSST broker teams themselves to classify alert streams. I'll place this work in the context of pushing from detections to cosmology. Looking to the brain: I'll present AstroBEATS, a pipeline derived from astronomical image analysis techniques and designed for high-resolution images of the brain. I'll describe how AstroBEATS can be used to study the synaptic firing in the brain and to search for signs of neurodegeneration and describe the processes that generated this interdisciplinary research.
February 12, 2025
Works-in-Progress Wednesday
11:20–11:55 a.m.
Hancock, 4010

Jeff McMahon


Ermin Wei
The University of Chicago


Northwestern University
TBA.
    —
10:45–11:20 a.m.
Vicky Kalogera


Aggelos Katsaggelos
Northwestern University
TBA.
    —
10:00–10:45 a.m.
Emma Alexander

Tjitske Starkenburg
Northwestern University
TBA.
    —
NSF logo

Funding Partners

The SkAI Institute is one of the National Artificial Intelligence Research Institutes funded by the U.S. National Science Foundation and Simons Foundation.

Simons Foundation logo

SkAI Institute

Location-icon outline 875 N. Michigan Ave., Suite 3500
Chicago, IL 60611

 skai-institute@u.northwestern.edu

(Get to the SkAI Hub via the 172 E. Chestnut St. entrance)