Theory of AI for Scientific Computing

Official website of the TASC workshop, to be held June 30, 2025 as part of COLT 2025

About TASC

The Theory of AI for Scientific Computing (TASC) workshop will be held as part of COLT 2025 and aims to advance the theoretical understanding of AI-augmented scientific computing, including capabilities and limitations of recent methods, principled architectures, unique data-generation and training considerations when connecting ML to downstream tasks, the sample complexity potential of transfer learning and active sampling, and ensuring the robustness of deployed systems and the validity of new discoveries. Our goal is to foster new theory that can bridge the gap between rapid methodological developments and their ultimate goals: scientific understanding and deployable computational systems. Join us on June 30, 2025 in Lyon, France for an exciting program of keynotes, posters, awards, and discussions about building principled connections between learning, algorithms, and the physical sciences, identifying promising scientific computing objectives for AI, and formalizing theoretical problems that will inspire continued progress.

Keynote Speakers

Joan Bruna

Joan Bruna

New York University

CS, Data Science, & Mathematics

On Inverse Problems and Diffusion
Aditi Krishnapriyan

Aditi Krishnapriyan

UC Berkeley

Chemical Engineering & EECS

Physics-Informed Machine Learning: Theory and Practice
Houman Owhadi

Houman Owhadi

California Institute of Technology

Computing & Mathematical Sciences

Operator Learning and Computational Homogenization

Workshop Schedule

June 30, 2025 • Mérieux Amphitheater, ENS Lyon • All times in Central European Time (CET)

9:00 AM
Welcoming remarks
9:05 AM
Joan Bruna (NYU)
On Inverse Problems and Diffusion

Abstract: Transport-based models, such as diffusion or flow-matching, have become a leading framework for generative modeling, by reducing the task to modeling conditional expectations under suitable "noise" semigroups. In this talk, we will describe how these models can be adapted beyond generative modeling to the setting of inverse problems. We will focus on two snippets: (i) performing provable posterior sampling in the context of linear inverse problems, and (ii) learning a generative model from corrupted measurements, akin to solving another linear inverse problem, this time in the space of distributions. Joint work with Jiequn Han, Chirag Modi and Eric Vanden-Eijnden.

9:45 AM
Houman Owhadi (Caltech)
Data-Efficient Kernel Methods for Learning Differential Equations and Their Solution Operators: Algorithms and Error Analysis

Abstract: We introduce a novel kernel-based framework for learning differential equations and their solution maps, which is efficient in terms of data requirements (both the number of solution examples and the amount of measurements from each example), as well as computational cost and training procedures. Our approach is mathematically interpretable and supported by rigorous theoretical guarantees in the form of quantitative worst-case error bounds for the learned equations and solution operators. Numerical benchmarks demonstrate significant improvements in computational complexity and robustness, achieving one to two orders of magnitude improvement in accuracy compared to state-of-the-art algorithms. This presentation is based on joint work with Yasamin Jalalian, Juan Felipe Osorio Ramirez, Alexander Hsu, and Bamdad Hosseini. A preprint is available at: https://arxiv.org/abs/2503.01036.

10:25 AM
Coffee break and poster setup
Universal Approximation of Mean-Field Models via Transformers
Authors: Shiba Biswal, Karthik Elamvazhuthi, Rishi Sonthalia

Student-t processes as infinite-width limits of posterior BNNs
Authors: Francesco Caporali, Stefano Favaro, Dario Trevisan

Matrix-Free Two-to-Infinity and One-to-Two Norms Estimation
Authors: Askar Tsyganov, Sergey Samsonov, Maxim Rakhuba

Translation-Equivariance of Normalization Layers and Aliasing in Convolutional Neural Networks
Authors: Jérémy Scanvic, Quentin Barthélemy, Julián Tachella

BWLer: Barycentric Weight Layer Elucidates a Precision-Conditioning Tradeoff for PINNs
Authors: Jerry Weihong Liu, Yasa Baig, Denise H. J. Lee, Rajat Vadiraj Dwaraknath, Atri Rudra, Christopher Re

Query Efficient Structured Matrix Approximation
Authors: Noah Amsel, Pratyush Avi, Tyler Chen, Feyza Duman Keles, Chinmay Hegde, Christopher Musco, Cameron N. Musco, David Persson

Generalized Lie Symmetries in Physics-Informed Neural Operators
Authors: Amy Xiang Wang, Zakhar Shumaylov, Peter Zaika, Ferdia Sherry, Carola-Bibiane Schönlieb

Enhancing Physics-Informed Neural Networks Through Feature Engineering
Authors: Shaghayegh Fazliani, Zachary Frangella, Ya-Chi Chu, Madeleine Udell

On the upper bounds for the matrix spectral norm
Authors: Ryapolov Denis, Maxim Rakhuba, Sergey Samsonov, Alexey Naumov

11:00 AM
Contributed Talk: Xiang Wang
Generalized Lie Symmetries in Physics-Informed Neural Operators

Abstract: Physics-informed neural operators (PINOs) have emerged as powerful tools for learning solution operators of partial differential equations (PDEs). Recent research has demonstrated that incorporating Lie point symmetry information can significantly enhance the training efficiency of PINOs, primarily through techniques like data, architecture, and loss augmentation. In this work, we focus on the latter, highlighting that point symmetries oftentimes result in no training signal, limiting their effectiveness in many problems. To address this, we propose a novel loss augmentation strategy that leverages evolutionary representatives of point symmetries, a specific class of generalized symmetries of the underlying PDE. These generalized symmetries provide a richer set of generators compared to standard symmetries, leading to a more informative training signal. We demonstrate that leveraging evolutionary representatives enhances the performance of neural operators, resulting in improved data efficiency and accuracy during training.

11:10 AM
Contributed Talk: Jerry Liu
BWLer: Barycentric Weight Layer Elucidates a Precision-Conditioning Tradeoff for PINNs

Abstract: Physics-informed neural networks (PINNs) offer a flexible way to solve partial differential equations (PDEs) with machine learning, yet they still fall well short of the machine-precision accuracy many scientific tasks demand. This motivates an investigation into whether the precision ceiling comes from the ill-conditioning of the PDEs themselves or from the typical multi-layer perceptron (MLP) architecture. We introduce the Barycentric Weight Layer (BWLer), which models the PDE solution through barycentric polynomial interpolation. A BWLer can be added on top of an existing MLP (a BWLer-hat) or replace it completely (explicit BWLer), cleanly separating how we represent the solution from how we take its derivatives for the physics loss. Using BWLer, we identify fundamental precision limitations within the MLP: on a simple 1-D interpolation task, even MLPs with O(105) parameters stall around 10-8 relative error -- about eight orders above float64 machine precision -- before any PDE terms are added. In PDE learning, adding a BWLer lifts this ceiling and exposes a tradeoff between achievable accuracy and the conditioning of the PDE loss. For linear PDEs we fully characterize this tradeoff with an explicit error decomposition and navigate it during training with spectral derivatives and preconditioning. Across five benchmark PDEs, adding a BWLer on top of an MLP improves relative error by up to 30x for convection, 10x for reaction, and 1800x for wave equations while remaining compatible with first-order optimizers. Replacing the MLP entirely lets an explicit BWLer reach near-machine-precision on convection, reaction, and wave problems (up to 10 billion times better than prior results) and match the performance of standard PINNs on stiff Burgers’ and irregular-geometry Poisson problems. Together, these findings point to a practical path for combining the flexibility of PINNs with the precision of classical spectral solvers.

11:20 AM
Aditi Krishnapriyan (UC Berkeley)
Machine learning methods for science: physics insights leveraging data at scale

Abstract: Recent advances in deep learning have demonstrated an empirical trend: large-scale, general-purpose models trained on large datasets often outperform specialized models explicitly encoding domain-specific “constraints,” generally known as the “bitter lesson.” At the same time, in scientific modeling, we still require solutions that respect known physical principles. This raises a key question: Can we incorporate physics insights into machine learning methods that still benefit from advances in large-scale training and data?

In this talk, I will explore these questions through several examples. I will briefly discuss physics-informed neural networks (PINNs) to highlight challenges arising from naive incorporation of constraints. I will also discuss approaches that leverage large-scale training while ensuring physical consistency at deployment. Finally, I will discuss combining pre-trained score-based generative models (diffusion and flow matching) with statistical physics insights. By interpreting sampling from generative models as dynamics governed by the Onsager-Machlup action functional, we can efficiently generate physically realistic dynamical pathways (e.g., molecular transitions or protein folding events) without additional training. Throughout, I'll highlight open questions on how physics and mathematical insights can be effectively integrated into methods that naturally improve with increasing data and model scale.

12:00 PM
Poster session
Universal Approximation of Mean-Field Models via Transformers
Authors: Shiba Biswal, Karthik Elamvazhuthi, Rishi Sonthalia

Student-t processes as infinite-width limits of posterior BNNs
Authors: Francesco Caporali, Stefano Favaro, Dario Trevisan

Matrix-Free Two-to-Infinity and One-to-Two Norms Estimation
Authors: Askar Tsyganov, Sergey Samsonov, Maxim Rakhuba

Translation-Equivariance of Normalization Layers and Aliasing in Convolutional Neural Networks
Authors: Jérémy Scanvic, Quentin Barthélemy, Julián Tachella

BWLer: Barycentric Weight Layer Elucidates a Precision-Conditioning Tradeoff for PINNs
Authors: Jerry Weihong Liu, Yasa Baig, Denise H. J. Lee, Rajat Vadiraj Dwaraknath, Atri Rudra, Christopher Re

Query Efficient Structured Matrix Approximation
Authors: Noah Amsel, Pratyush Avi, Tyler Chen, Feyza Duman Keles, Chinmay Hegde, Christopher Musco, Cameron N. Musco, David Persson

Generalized Lie Symmetries in Physics-Informed Neural Operators
Authors: Amy Xiang Wang, Zakhar Shumaylov, Peter Zaika, Ferdia Sherry, Carola-Bibiane Schönlieb

Enhancing Physics-Informed Neural Networks Through Feature Engineering
Authors: Shaghayegh Fazliani, Zachary Frangella, Ya-Chi Chu, Madeleine Udell

On the upper bounds for the matrix spectral norm
Authors: Ryapolov Denis, Maxim Rakhuba, Sergey Samsonov, Alexey Naumov

Call for Papers

We invite submissions at the interface of learning theory, statistics, numerical methods, algorithm design, and the physical sciences. Submissions may be of any length; in particular, we welcome both short poster abstracts and multi-page papers. The topics of the workshop include (but are not limited to) the following:


    Learning-theoretic foundations
    • learning-theoretic and statistical analysis of data-driven solutions to important scientific computing tasks such as solving differential equations, inverse problems, sampling, equation discovery, and beyond
    • mathematical characterizations of settings in which AI-augmented methods can be expected to improve over traditional (AI-free) scientific computing methods
    • end-to-end theoretical studies that consider simultaneously the entire scientific computing pipeline, i.e. not only learning from data but also generating the data itself and integrating the learned model into downstream scientific computing tasks
    • non-i.i.d. settings such as active sampling of ground truth solutions, reinforcement learning (RL) of equations and solvers, and transfer learning between different differential equation families, different solution domains, and different initial or boundary conditions
    • formalizing concrete goals for scientific discovery
    Principled methods for AI-augmented scientific computing
    • theoretically-motivated design of neural architectures and loss functions for scientific computing tasks
    • mathematical connections between generative modeling and tractable solutions of high-dimensional PDEs
    • principled approaches to data-generation, model training, and model deployment
    • statistical machinery for certifying the quality and confidence of AI-augmented algorithms and improving their robustness
    Connections with other subfields of theory
    • sampling
    • learning-augmented algorithms (algorithms with predictions) and data-driven algorithm design
    • optimization
    • randomized numerical linear algebra

Submission criteria:

Papers and abstracts should be submitted as PDF files in any format that has a font size of at least 10 points and margins of at least 1 inch. Submissions are not limited in length, but only the first 8 pages are guaranteed to be reviewed. Accepted submissions will be made public on OpenReview but are non-archival, and we welcome work accepted at previous or upcoming conferences, including COLT 2025 and ICML 2025.


Reviewing and publication:

All submissions will undergo a double-blind peer review process assessing mainly relevance, clarity, and soundness. Reviewing will occur on the OpenReview platform but reviews will not be public. Authors of accepted submissions will be invited to present a poster about it on the day of the workshop (June 30). The organizers will also select up to two submissions to receive best paper and runner-up awards, and their authors will have the opportunity to present short contributed talks.

Paper Submission

Submissions can be made through the OpenReview link below; if registering an account there we recommended using an institutional email to avoid an up to two-week moderation period. Please ensure that you follow the submission guidelines outlined in the Call for Papers.

Submission Link

For any questions regarding the submission process, please contact: tasc.organizers@gmail.com

Important Dates

Paper Submission Deadline

May 16, 2025
May 23, 2025

Notification of Acceptance

June 1, 2025

Camera-Ready Deadline

June 15, 2025

Workshop Date

June 30, 2025

All deadlines are 11:59 PM Anywhere on Earth (AoE)

Organizers

Nick Boffi

Nick Boffi

Carnegie Mellon University

Misha Khodak

Misha Khodak

Princeton University

Jianfeng Lu

Jianfeng Lu

Duke University

Tanya Marwah

Tanya Marwah

Polymathic AI and the Flatiron Institute

Andrej Risteski

Andrej Risteski

Carnegie Mellon University

Registration

Registration for the TASC workshop is automatically included in your COLT 2025 conference registration fee; no separate registration is required. To register for COLT 2025, go to the COLT 2025 registration website and follow the instructions in the "Registration procedure" section. The early bird registration deadline is May 22.