Loading Events

« All Events

  • This event has passed.

Mathematics and Machine Learning Closing Workshop

October 28, 2024 @ 9:00 am - October 30, 2024 @ 5:00 pm

Mathematics and Machine Learning Closing Workshop

Dates: October 28 – Oct. 30, 2024

Location: Room G10, CMSA, 20 Garden Street, Cambridge MA

The closing workshop will provide a forum for discussing the most current research in these areas, including work in progress and recent results from program participants. We will devote one day to frontier topics in interactive theorem proving, such as mathematical library development and AI for mathematical search and theorem proving.

 

Youtube Playlist

Organizers

  • Francois Charton (Meta AI)
  • Michael R. Douglas (Harvard CMSA)
  • Michael Freedman (Harvard CMSA)
  • Fabian Ruehle (Northeastern)
  • Geordie Williamson (Univ. of Sydney)

Speakers

 

Schedule 
(download PDF)

Monday Oct. 28, 2024

9:00–9:30 am
Morning refreshments

9:30–9:45 am
Introductions

9:45–10:45 am
Jürgen Jost, Max Planck Institute for Mathematics in the Sciences

Title: Data visualization with category theory and geometry

Abstract: While data often come in a high-dimensional feature space, they typically exhibit intrinsic constraints and regularities, and they can therefore often be represented on a lower-dimensional, but possibly highly curved Riemannian manifold. Still, for visualization purposes, that dimension still needs to be lowered to 2 or 3. We present the mathematical foundations for such schemes, in particular UMAP, and describe an improved such method.

10:45–11:00 am
Break

11:00 am–12:00 pm
Ankit Anand, Google Deepmind Montreal, Abbas Mehrabian, Google Deepmind Montreal

Title: From Theorem Proving to Disproving: Modern machine learning versus classical heuristic search in automated theorem proving and extremal graph theory

Abstract: Machine learning is widely believed to outperform classical methods, but this is not always the case. Firstly, we describe how we adapted the idea of hindsight experience replay from reinforcement learning to the automated theorem proving domain, so as to use the intermediate data generated during unsuccessful proofs. We show that provers trained in this way can outperform previous machine learning approaches and compete with the state-of-the-art heuristic-based theorem prover E in its best configuration, on the popular benchmarks MPTP2078, M2k and Mizar40. The proofs generated by our algorithm are also almost always significantly shorter than E’s proofs. Based on this paper, which was presented at ICML 2022: https://proceedings.mlr.press/v162/aygun22a.html. Secondly, we study a central extremal graph theory problem inspired by a 1975 conjecture of Erdős, which aims to find graphs with a given size (number of nodes) that maximize the number of edges without having 3- or 4-cycles. We formulate this problem as a sequential decision-making problem and compare AlphaZero, a neural network-guided tree search, with tabu search, a heuristic local search method. Using either method, by introducing a curriculum—jump-starting the search for larger graphs using good graphs found at smaller sizes—we improve the state-of-the-art lower bounds for several sizes. Joint work with Tudor Berariu, Joonkyung Lee, Anurag Murty Naredla, Adam Zsolt Wagner, and other colleagues at Google DeepMind. Based on this paper, which was presented at IJCAI 2024: https://arxiv.org/abs/2311.03583.

12:00–1:30 pm
Lunch Break

1:30–2:30 pm
Fabian Ruehle, Northeastern University, Giorgi Butbaia, University of New Hampshire

Title: Rigorous results 
from ML using RL

Abstract: We explain how to use Reinforcement Learning in Mathematics to obtain provably correct results. After a brief introduction to Reinforcement Learning, I will illustrate the idea using an example from Number Theory, where we solve a Diophantine Equation related to String Theory, and two from Knot Theory. The first knot theory problem is to identify unknots, while the second is concerned with identifying so-called ribbon knots. The latter play an important role in the search for counter-examples to the smooth Poincare conjecture.

2:30–2:45 pm
Break

2:45–3:45 pm
Cengiz Pehlevan, Harvard University

Title: Solvable Models of Scaling and Emergence in Deep Learning

3:45–4:00 pm
Break

4:00–4:30 pm
Matej Balog, Deepmind
via Zoom

Title: FunSearch: Mathematical discoveries from program search with large language models

Abstract: Large language models (LLMs) have demonstrated tremendous capabilities in solving complex tasks, from quantitative reasoning to understanding natural language. However, LLMs sometimes suffer from confabulations (or hallucinations), which can result in them making plausible but incorrect statements. This hinders the use of current large models in scientific discovery. We introduce FunSearch (short for searching in the function space), an evolutionary procedure based on pairing a pretrained LLM with a systematic evaluator. We demonstrate the effectiveness of this approach to surpass the best-known results in important problems, pushing the boundary of existing LLM-based approaches. Applying FunSearch to a central problem in extremal combinatorics—the cap set problem—we discover new constructions of large cap sets going beyond the best-known ones, both in finite dimensional and asymptotic cases. This shows that it is possible to make discoveries for established open problems using LLMs. We showcase the generality of FunSearch by applying it to an algorithmic problem, online bin packing, finding new heuristics that improve on widely used baselines. In contrast to most computer search approaches, FunSearch searches for programs that describe how to solve a problem, rather than what the solution is. Beyond being an effective and scalable strategy, discovered programs tend to be more interpretable than raw solutions, enabling feedback loops between domain experts and FunSearch, and the deployment of such programs in real-world applications.

4:30–5:00 pm
Edgar Costa, MIT

Title: Machine learning L-functions

Abstract: We report on multiple experiments related to L-functions data. L-functions are complex functions that encode significant information about number theory and algebraic geometry, playing a crucial part in the Langlands program, a foundational set of conjectures connecting number theory with other mathematical domains. We focused on two L-function datasets. The first includes about 250k rational L-functions of small arithmetic complexity with diverse origins. Multiple dimensionality reduction techniques were used to analyze invariants and behavioral trends, focusing on how differing origins impact the results. The second dataset is composed of L-functions associated with Maass forms. Although these L-functions are non-rational, they also share the low arithmetic complexity of the first dataset. The crux of our investigation here is determining whether this set manifests similar characteristics to the first one. Based on this exploration, we propose a simple heuristic method to deduce their Fricke sign, an unknown invariant for 40% of the data. This is joint work with: Joanna Biere, Giorgi Butbaia, Alyson Deines, Kyu-Hwan Lee, David Lowry-Duda, Tom Oliver, Tamara Veenstra, and Yidi Qi.

 

 

Tuesday Oct. 29, 2024 

9:15–9:45 am
Morning refreshments

9:45–10:45 am
Yang-Hui He, London Institute for Mathematical Sciences 
Via Zoom

Title: AI assisted mathematics

Abstract: We argue how AI can assist mathematics in three ways: theorem-proving, conjecture formulation, and language processing. Inspired by initial experiments in geometry and string theory in 2017, we summarize how this emerging field has grown over the past years, and show how various machine-learning algorithms can help with pattern detection across disciplines ranging from algebraic geometry to representation theory, to combinatorics, and to number theory.  At the heart of the program is the question how does AI help with theoretical discovery, and the implications for the future of mathematics.

10:45–11:00 am
Break

11:00 –11:30
Angelica Babei

Title: Learning Euler factors of elliptic curves with transformers

Abstract: The L-function of an elliptic curve is at the core of the BSD conjecture, and its Euler factors encode important arithmetic information about the curve. For example, understanding these Euler factors using machine learning techniques has recently led to discovering the phenomenon of murmurations. In this talk, we present some results on learning Euler factors based on 1. other nearby factors, and 2. the Weierstrass equation of the curve. 

11:30–12:00 pm
David Lowry-Duda, ICERM

Title: Exploring patterns in number theory with deep learning: a case study with the Möbius and squarefree indicator functions

Abstract: We report on experiments using neural networks and Int2Int, the integer sequence to integer sequence transformer made by François Charton for this CMSA program. We initially study the Möbius function. This function appears as the coefficients of the reciprocal of the Riemann zeta function and is famously hard to understand. Predicting the Möbius function is closely related to predicting the squarefree indicator function, leading us to perform similar experiments there. Finally, we’ll discuss how varying the input representation and model affects the strength of the predictions and allows us to explain most (but not all) of the predictive strength and behavior.

12:00–1:30 pm
Lunch

1:30–2:30 pm
Amaury Hayat, Ecole des Ponts Paristech, Melanie Matchett Wood, Harvard University, Alex Davies, DeepMind, Jeremy Avigad, Carnegie Mellon University

Title: Machine learning and theorem proving

Abstract: Recent successes in machine learning have raised hopes that neural networks could one day assist mathematicians in proving theorems. This raises the question of an appropriate setting to apply machine learning methods to theorem proving. Formal languages, such as Lean, provide automatic verification of mathematical proofs and thus offer a natural environment. Nevertheless, some challenges emerge, particularly because these languages are often designed to verify correctness rather than find a solution, while mathematicians often perform reasoning steps to do both at the same time. This talk will present recent applications of machine learning methods to theorem proving in Lean, highlight current challenges, and explore what these developments might mean for the future of mathematics.

 

2:30–2:45 pm
Break

2:45–3:45 pm
Adam Wagner, Worcester Polytechnic Institute, Kit Fraser-Taliente, University of Oxford, Gergely Bérczi, Aarhus University, Thomas Harvey, MIT

Title: Sparse subgraphs of the d-cube with diameter d

Abstract: Erdos et al studied spanning subgraphs of the $d$-cube which have the same diameter $d$ as the cube itself. They asked the following natural question: what is the maximum number of edges one can delete from the $d$-dimensional hypercube, without increasing its diameter? We will discuss how we can use PatternBoost, a simple machine learning algorithm that alternates local and global optimization steps, to find good constructions for this problem

3:45–4:00 pm
Break

4:00–4:30 pm
Petros Koumoutsakos, Harvard University

4:30–5:00 pm

Stéphane Mallat,  Flatiron/College de France

Title: Image Generation by Score Diffusion and the Renormalisation Group

Abstract: Score based diffusions generate impressive models of images, sounds and complex physical systems. Are they generalising or memorising? How can deep network estimate high-dimensional scores without curse of dimensionality? This talk shows that generalisation does occur for deep network estimation of scores, with enough training data.  The ability to avoid the curse of dimensionality seems to rely on multiscale properties revealed by a renormalisation group decomposition coming from statistical physics. Applications to models of turbulences will be introduced and discussed.

 

Wednesday Oct. 30, 2024

9:15–9:45 am
Morning refreshments

9:45–10:45 am
Bin Dong, Beijing International Center for Mathematical Research
(via Zoom) 

Title: AI for Mathematics: From Digitization to Intelligentization

Abstract: This presentation explores the synergistic relationship between AI and mathematics, beginning with a brief historical overview of their mutually beneficial interactions. It then examines notable existing work in AI for mathematics, highlighting their achievements and limitations.  Next, I will share preliminary findings from the ongoing AI4M research project at Peking University, including our work on creating high-quality mathematical datasets through formalization (digitization), and our future plans for developing intelligent applications using these datasets. The presentation concludes with a forward-looking perspective on the opportunities and challenges within this exciting interdisciplinary field.

10:45–11:00 am 
Break

11:00 am–12:00 pm 
Eric Vanden-Eijnden, Courant/NYU
via Zoom

Title: Generative modeling with flows and diffusions, with applications to scientific computing.

Abstract: Generative models based on dynamical transport have recently led to significant advances in unsupervised learning. At mathematical level, these models are primarily designed around the construction of a map between two probability distributions that transform samples from the first into samples from the second.  While these methods were first introduced in the context of image generation, they have found a wide range of applications, including in scientific computing where they offer interesting ways to reconsider complex problems once thought intractable because of the curse of dimensionality. In this talk, I will discuss the mathematical underpinning of generative models based on flows and diffusions, and show how a better understanding of their inner workings can help improve their design. These results indicate how to structure the transport to best reach complex target distributions while maintaining computational efficiency, both at learning and sampling stages.  I will also discuss applications of generative AI in scientific computing, in particular in the context of application with models and no data (as opposed to the more standard data andno model), such as Monte Carlo sampling, with applications to the statistical mechanics and Bayesian inference, as well as the numerical integration and interpretation of random dynamical systems driven out of equilibrium.

12:00–1:30 pm 
Lunch

1:30–2:30 pm 
Kyu-Hwan Lee, University of Connecticut

Title: Discovering New Mathematical Structures with Machine Learning

Abstract: Can machine learning help discover new mathematical structures? In this talk, I will present two case studies: murmurations in number theory and loadings of partitions related to Kronecker coefficients in representation theory and combinatorics. The focus will be on the paradigm of examining mathematical objects collectively, rather than individually, to create datasets suitable for machine learning experiments and interpretations.

2:30–2:45 pm 
Break

2:45–3:45 pm 
James Halverson, Northeastern University

Title: Learning the Topological Invariance of Knots

Abstract: This talk focuses on using machine learning for the defining problem in knot theory, the classification of knots up to ambient space isotopy. We will train transformers and convolutional neural networks to distinguish topologically inequivalent knots, given only representatives of the classes and no a priori knowledge of topological invariants. In this scheme, we find that equivalent knots are well-clustered in the embedding space of the neural network, and a trained decoder maps effectively from the embedding space back to knot space. Preliminary results will be presented on a new approach to resolving the Jones unknot conjecture.

3:45–4:00 pm
Break

4:00–5:00 pm
Tristan Buckmaster, New York University, Javier Gomez-Serrano, Brown University
via Zoom

 

 

Image by Sue Side. https://www.sueside.com/

 

Details

Start:
October 28, 2024 @ 9:00 am
End:
October 30, 2024 @ 5:00 pm
Event Category:

Venue

CMSA Room G10
CMSA, 20 Garden Street
Cambridge, MA 02138 United States
+ Google Map
Phone:
6174967132