During 2023–24, the CMSA will host a seminar on New Technologies in Mathematics, organized by Michael Douglas and Samy Jelassi. This seminar will take place on Wednesdays from 2:00 pm–3:00 pm (Eastern Time). The meetings will take place in Room G10 at the CMSA, 20 Garden Street, Cambridge MA 02138, and some meetings will take place virtually on Zoom or be held in hybrid formats. To learn how to attend, please fill out this form, or contact Michael Douglas (mdouglas@cmsa.fas.harvard.edu).

The schedule will be updated as talks are confirmed.

Seminar videos can be found at the CMSA Youtube site: New Technologies in Mathematics Playlist

Formal Mathematics Statement Curriculum Learning

https://youtu.be/4zINaGrPc9M Speaker: Stanislas Polu, OpenAI Title: Formal Mathematics Statement Curriculum Learning Abstract: We explore the use of expert iteration in the context of language modeling applied to formal mathematics. We show that at same compute budget, expert iteration, by which we mean proof search interleaved with learning, dramatically outperforms proof search only.  We also observe that […]

Memorizing Transformers

Virtual

https://youtu.be/5AoOpFFjW28 Speaker: Yuhuai Wu, Stanford and Google Title: Memorizing Transformers Abstract: Language models typically need to be trained or fine-tuned in order to acquire new knowledge, which involves updating their weights. We instead envision language models that can simply read and memorize new data at inference time, thus acquiring new knowledge immediately. In this talk, I […]

Breaking the one-mind-barrier in mathematics using formal verification

CMSA Room G10 CMSA, 20 Garden Street, Cambridge, MA, United States

https://youtu.be/D7dqadF5k9Q New Technologies in Mathematics Seminar Speaker: Johan Commelin, Mathematisches Institut, Albert-Ludwigs-Universität Freiburg Title: Breaking the one-mind-barrier in mathematics using formal verification Abstract: In this talk I will argue that formal verification helps break the one-mind-barrier in mathematics. Indeed, formal verification allows a team of mathematicians to collaborate on a project, without one person understanding all parts of […]

Statistical mechanics of neural networks: From the geometry of high dimensional error landscapes to beating power law neural scaling

CMSA Room G10 CMSA, 20 Garden Street, Cambridge, MA, United States

https://youtu.be/SQlPI07LvUc New Technologies in Mathematics Speaker: Surya Ganguli, Stanford University Title: Statistical mechanics of neural networks: From the geometry of high dimensional error landscapes to beating power law neural scaling Abstract: Statistical mechanics and neural network theory have long enjoyed fruitful interactions.  We will review some of our recent work in this area and then focus on […]

Minerva: Solving Quantitative Reasoning Problems with Language Models

CMSA Room G10 CMSA, 20 Garden Street, Cambridge, MA, United States

https://youtu.be/HUTWime3d6w New Technologies in Mathematics Seminar Speaker: Guy Gur-Ari, Google Research Title: Minerva: Solving Quantitative Reasoning Problems with Language Models Abstract: Quantitative reasoning tasks which can involve mathematics, science, and programming are often challenging for machine learning models in general and for language models in particular. We show that transformer-based language models obtain significantly better performance […]

Towards Faithful Reasoning Using Language Models

CMSA Room G10 CMSA, 20 Garden Street, Cambridge, MA, United States

New Technologies in Mathematics Seminar Speaker: Antonia Creswell, DeepMind Title: Towards Faithful Reasoning Using Language Models Abstract: Language models are showing impressive performance on many natural language tasks, including question-answering. However, language models – like most deep learning models – are black boxes. We cannot be sure how they obtain their answers. Do they reason […]

From Engine to Auto

CMSA Room G10 CMSA, 20 Garden Street, Cambridge, MA, United States

https://youtu.be/h3jcSg359E8 New Technologies in Mathematics Seminar Speakers: João Araújo, Mathematics Department, Universidade Nova de Lisboa and Michael Kinyon, Department of Mathematics, University of Denver Title: From Engine to Auto Abstract: Bill McCune produced the program EQP that deals with first order logic formulas and in 1996 managed to solve Robbins' Conjecture. This very powerful tool reduces […]

How do Transformers reason? First principles via automata, semigroups, and circuits

CMSA Room G10 CMSA, 20 Garden Street, Cambridge, MA, United States

https://youtu.be/g8zdumOAWzw New Technologies in Mathematics Seminar Speaker: Cyril Zhang, Microsoft Research Title: How do Transformers reason? First principles via automata, semigroups, and circuits Abstract: The current "Transformer era" of deep learning is marked by the emergence of combinatorial and algorithmic reasoning capabilities in large sequence models, leading to dramatic advances in natural language understanding, program synthesis, […]

How to steer foundation models?

CMSA Room G10 CMSA, 20 Garden Street, Cambridge, MA, United States

https://youtu.be/ztk5TPYTKZA New Technologies in Mathematics Seminar Speaker: Jimmy Ba, University of Toronto Title: How to steer foundation models? Abstract: By conditioning on natural language instructions, foundation models and large language models (LLMs) have displayed impressive capabilities as general-purpose computers. However, task performance depends significantly on the quality of the prompt used to steer the model. […]

Toolformer: Language Models Can Teach Themselves to Use Tools

Virtual

https://youtu.be/UID_oXuN-0Y New Technologies in Mathematics Seminar Speaker: Timo Schick, Meta AI Title: Toolformer: Language Models Can Teach Themselves to Use Tools Abstract: Language models exhibit remarkable abilities to solve new tasks from just a few examples or textual instructions, especially at scale. They also, paradoxically, struggle with basic functionality, such as arithmetic or factual lookup, where […]

Modern Hopfield Networks for Novel Transformer Architectures

Virtual

https://youtu.be/5LXiQUsnHrI New Technologies in Mathematics Seminar Speaker: Dmitry Krotov, IBM Research - Cambridge Title: Modern Hopfield Networks for Novel Transformer Architectures Abstract: Modern Hopfield Networks or Dense Associative Memories are recurrent neural networks with fixed point attractor states that are described by an energy function. In contrast to conventional Hopfield Networks, which were popular in […]

The TinyStories Dataset: How Small Can Language Models Be And Still Speak Coherent

Virtual

https://youtu.be/wTQH6mRDXhw New Technologies in Mathematics Seminar Speaker: Ronen Eldan, Microsoft Research Title: The TinyStories Dataset: How Small Can Language Models Be And Still Speak Coherent Abstract: While generative language models exhibit powerful capabilities at large scale, when either the model or the number of training steps is too small, they struggle to produce coherent and fluent […]