BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//CMSA - ECPv6.15.20//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-WR-CALNAME:CMSA
X-ORIGINAL-URL:https://cmsa.fas.harvard.edu
X-WR-CALDESC:Events for CMSA
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/New_York
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20190310T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20191103T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20200308T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20201101T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20210314T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20211107T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20220313T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20221106T060000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20211027T140000
DTEND;TZID=America/New_York:20211027T150000
DTSTAMP:20260511T093358
CREATED:20240214T093929Z
LAST-MODIFIED:20240517T193105Z
UID:10002641-1635343200-1635346800@cmsa.fas.harvard.edu
SUMMARY:Why explain mathematics to computers?
DESCRIPTION:Speaker: Patrick Massot\, Laboratoire de Mathématiques d’Orsay and CNRS \nTitle: Why explain mathematics to computers? \nAbstract: A growing number of mathematicians are having fun explaining mathematics to computers using proof assistant softwares. This process is called formalization. In this talk I’ll describe what formalization looks like\, what kind of things it teaches us\, and how it could even turn out to be useful (in our usual sense of “useful”). This will not be a talk about foundations of mathematics\, and I won’t assume any prior knowledge about formalization.
URL:https://cmsa.fas.harvard.edu/event/10-27-2021-new-technologies-in-mathematics-seminar/
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-NTM-Seminar-10.27.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20211013T140000
DTEND;TZID=America/New_York:20211013T150000
DTSTAMP:20260511T093358
CREATED:20240214T093531Z
LAST-MODIFIED:20240515T204354Z
UID:10002637-1634133600-1634137200@cmsa.fas.harvard.edu
SUMMARY:Computer-Aided Mathematics and Satisfiability
DESCRIPTION:Speaker: Marijn Heule\, Carnegie Mellon University \nTitle: Computer-Aided Mathematics and Satisfiability \nAbstract: Progress in satisfiability (SAT) solving has made it possible to determine the correctness of complex systems and answer long-standing open questions in mathematics. The SAT solving approach is completely automatic and can produce clever though potentially gigantic proofs. We can have confidence in the correctness of the answers because highly trustworthy systems can validate the underlying proofs regardless of their size. We demonstrate the effectiveness of the SAT approach by presenting some recent successes\, including the solution of the Boolean Pythagorean Triples problem\, computing the fifth Schur number\, and resolving the remaining case of Keller’s conjecture. Moreover\, we constructed and validated a proof for each of these results. The second part of the talk focuses on notorious math challenges for which automated reasoning may well be suitable. In particular\, we discuss our progress on applying SAT solving techniques to the chromatic number of the plane (Hadwiger-Nelson problem)\, optimal schemes for matrix multiplication\, and the Collatz conjecture.
URL:https://cmsa.fas.harvard.edu/event/10-13-2021-new-technologies-in-mathematics-seminar/
CATEGORIES:New Technologies in Mathematics Seminar
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20211006T150000
DTEND;TZID=America/New_York:20211006T160000
DTSTAMP:20260511T093358
CREATED:20240214T092921Z
LAST-MODIFIED:20240517T200621Z
UID:10002630-1633532400-1633536000@cmsa.fas.harvard.edu
SUMMARY:New results in Supergravity via ML Technology
DESCRIPTION:Speaker: Thomas Fischbacher\, Google \nTitle: New results in Supergravity via ML Technology \nAbstract: The infrastructure built to power the Machine Learning revolution has many other uses beyond Deep Learning. Starting from a general architecture-level overview over the lower levels of Google’s TensorFlow machine learning library\, we review how this has recently helped us to find all the stable vacua of SO(8) Supergravity in 3+1 dimensions\, has allowed major progress on other related questions about M theory\, and briefly discuss other applications in field theory and beyond.
URL:https://cmsa.fas.harvard.edu/event/10-6-2021-new-technologies-in-mathematics-seminar/
CATEGORIES:New Technologies in Mathematics Seminar
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210929T150000
DTEND;TZID=America/New_York:20210929T160000
DTSTAMP:20260511T093358
CREATED:20240214T092650Z
LAST-MODIFIED:20240517T200354Z
UID:10002626-1632927600-1632931200@cmsa.fas.harvard.edu
SUMMARY:Constructions in combinatorics via neural networks
DESCRIPTION:Speaker: Adam Wagner\, Tel Aviv University \nTitle: Constructions in combinatorics via neural networks \nAbstract: Recently\, significant progress has been made in the area of machine learning algorithms\, and they have quickly become some of the most exciting tools in a scientist’s toolbox. In particular\, recent advances in the field of reinforcement learning have led computers to reach superhuman level play in Atari games and Go\, purely through self-play. In this talk I will give a very basic introduction to neural networks and reinforcement learning algorithms. I will also indicate how these methods can be adapted to the ““game” of trying to find a counterexample to a mathematical conjecture\, and show some examples where this approach was successful.
URL:https://cmsa.fas.harvard.edu/event/9-29-2021-new-technologies-in-mathematics-seminar/
CATEGORIES:New Technologies in Mathematics Seminar
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210915T150000
DTEND;TZID=America/New_York:20210915T160000
DTSTAMP:20260511T093358
CREATED:20240214T091846Z
LAST-MODIFIED:20240517T200145Z
UID:10002618-1631718000-1631721600@cmsa.fas.harvard.edu
SUMMARY:Why abstraction is the key to intelligence\, and what we’re still missing
DESCRIPTION:Speaker: Francois Chollet\, Google \nTitle: Why abstraction is the key to intelligence\, and what we’re still missing \nAbstract: This talk provides a personal perspective on the way forward towards more human-like and more intelligent artificial systems. Traditionally\, symbolic and probabilistic methods have dominated the domains of concept formation\, abstraction\, and automated reasoning. More recently\, deep learning-based approaches have led to significant breakthroughs\, including successes in games and combinatorial search tasks. However\, the resulting systems are still limited in scope and capabilities — they remain brittle\, data-hungry\, and their generalization capabilities are limited. We will address a set of questions: why is conceptual abstraction essential for intelligence? What is the nature of abstraction\, and its relationship to generalization? What kind of abstraction can deep learning models generate\, and where do they fail? What are the methods that are currently successful in generating strong conceptual abstraction? Finally\, we will consider how to leverage a hybrid approach to reinforce the strength of different approaches while compensating for their respective weaknesses.
URL:https://cmsa.fas.harvard.edu/event/9-15-2021-new-technologies-in-mathematics-seminar/
CATEGORIES:New Technologies in Mathematics Seminar
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210909T140000
DTEND;TZID=America/New_York:20210909T150000
DTSTAMP:20260511T093358
CREATED:20240214T092250Z
LAST-MODIFIED:20240517T200424Z
UID:10002621-1631196000-1631199600@cmsa.fas.harvard.edu
SUMMARY:The complexity of matrix multiplication approached via algebraic geometry and representation theory
DESCRIPTION:Speaker: JM Landsberg\, Texas A&M \nTitle: The complexity of matrix multiplication approached via algebraic geometry and representation theory \nAbstract: In 1968 V. Strassen discovered the way we usually multiply matrices is not the most efficient possible\, and after considerable work by many authors\, it is generally conjectured by computer scientists that as the size of matrices becomes large\, it becomes almost as easy to multiply them as it is to add them. I will give a brief history of the problem\, explain how this conjecture is naturally understood in the framework of classical algebraic geometry and representation theory\, and conclude by describing recent advances using more sophisticated tools from algebraic geometry. For most of the talk\, no knowledge of algebraic geometry or representation theory will be needed.
URL:https://cmsa.fas.harvard.edu/event/9-22-2021-new-technologies-in-mathematics-seminar/
CATEGORIES:New Technologies in Mathematics Seminar
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210421T150000
DTEND;TZID=America/New_York:20210421T160000
DTSTAMP:20260511T093358
CREATED:20240126T065403Z
LAST-MODIFIED:20240515T192602Z
UID:10001388-1619017200-1619020800@cmsa.fas.harvard.edu
SUMMARY:Homotopy type theory and the quest for extensionality
DESCRIPTION:Speaker: Michael Shulman – Dept. of Mathematics\, University of San Diego \nTitle: Homotopy type theory and the quest for extensionality \nAbstract: Over the past decades\, dependent type theory has proven to be a powerful framework for verified software and formalized mathematics.  However\, its treatment of equality has always been somewhat uncomfortable.  Recently\, homotopy type theory has made progress towards a more useful notion of equality\, which natively implements both isomorphism-invariance in mathematics and representation-independence in programming. This progress is based on ideas from abstract homotopy theory and higher category theory\, and with the development of cubical type theories it can be implemented as a true programming language.  In this talk\, I will survey these developments and their potential applications\, and suggest some directions for further improvement. \n 
URL:https://cmsa.fas.harvard.edu/event/4-21-2021-new-tech-in-math-seminar/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-04.21.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210414T150000
DTEND;TZID=America/New_York:20210414T160000
DTSTAMP:20260511T093358
CREATED:20240126T064644Z
LAST-MODIFIED:20240517T194855Z
UID:10001384-1618412400-1618416000@cmsa.fas.harvard.edu
SUMMARY:A Bayesian neural network predicts the dissolution of compact planetary systems
DESCRIPTION:Speaker: Miles Cranmer – Princeton University \nTitle: A Bayesian neural network predicts the dissolution of compact planetary systems \nAbstract: Despite over three hundred years of effort\, no solutions exist for predicting when a general planetary configuration will become unstable. I will discuss our deep learning architecture (arxiv:2101.04117) which pushes forward this problem for compact systems. While current machine learning algorithms in this area rely on scientist-derived instability metrics\, our new technique learns its own metrics from scratch\, enabled by a novel internal structure inspired from dynamics theory. The Bayesian neural network model can accurately predict not only if\, but also when a compact planetary system with three or more planets will go unstable. Our model\, trained directly from short N-body time series of raw orbital elements\, is more than two orders of magnitude more accurate at predicting instability times than analytical estimators\, while also reducing the bias of existing machine learning algorithms by nearly a factor of three. Despite being trained on three-planet configurations\, the model demonstrates robust generalization to five-planet systems\, even outperforming models designed for that specific set of integrations. I will also discuss some work on recovering symbolic representations of such models using arxiv:2006.11287.
URL:https://cmsa.fas.harvard.edu/event/4-14-2021-new-technologies-in-mathematics/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-04.14.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210407T150000
DTEND;TZID=America/New_York:20210407T160000
DTSTAMP:20260511T093358
CREATED:20240126T073259Z
LAST-MODIFIED:20240515T193258Z
UID:10001398-1617807600-1617811200@cmsa.fas.harvard.edu
SUMMARY:Type theory from the perspective of artificial intelligence
DESCRIPTION:Speaker: David McAllester – Toyota Technological Institute at Chicago \nTitle: Type theory from the perspective of artificial intelligence \nAbstract: This talk will discuss dependent type theory from the perspective of artificial intelligence and cognitive science. From an artificial intelligence perspective it will be argued that type theory is central to defining the “game” of mathematics — an action space and reward structure for pure mathematics. From a cognitive science perspective type theory provides a model of the grammar of the colloquial (natural) language of mathematics. Of particular interest is the notion of a signature-axiom structure class and the three fundamental notions of equality in mathematics — set-theoretic equality between structure elements\, isomorphism between structures\, and Birkoff and Rota’s notion of cryptomorphism between structure classes. This talk will present a version of type theory based on set-theoretic semantics and the 1930’s notion of structure and isomorphism given by the Bourbaki group of mathematicians. It will be argued that this “Bourbaki type theory” (BTT) is more natural and accessible to classically trained mathematicians than Martin-Löf type theory (MLTT). BTT avoids the Curry-Howard isomorphism and axiom J of MLTT. The talk will also discuss BTT as a model of MLTT. The BTT model is similar to the groupoid model in that propositional equality is interpreted as isomorphism but different in various details. The talk will also briefly mention initial thoughts in defining an action space and reward structure for a game of mathematics.
URL:https://cmsa.fas.harvard.edu/event/4-7-2021-new-technologies-in-mathematics-seminar/
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-04.07.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210331T150000
DTEND;TZID=America/New_York:20210331T160000
DTSTAMP:20260511T093358
CREATED:20240126T083143Z
LAST-MODIFIED:20240515T195507Z
UID:10001412-1617202800-1617206400@cmsa.fas.harvard.edu
SUMMARY:Doing Mathematics with Simple Types: Infinitary Combinatorics in Isabelle/HOL
DESCRIPTION:Speaker: Lawrence Paulson\, University of Cambridge Computer Laboratory  \nTitle: Doing Mathematics with Simple Types: Infinitary Combinatorics in Isabelle/HOL  \nAbstract: Are proof assistants relevant to mathematics? One approach to this question is to explore the breadth of mathematical topics that can be formalised. The partition calculus was introduced by Erdös and R. Rado in 1956 as the study of “analogues and extensions of Ramsey’s theorem”. Highly technical results were obtained by Erdös-Milner\, Specker and Larson (among many others) for the particular case of ordinal partition relations\, which is concerned with countable ordinals and order types. Much of this material was formalised last year (with the assistance of Džamonja and Koutsoukou-Argyraki). Some highlights of this work will be presented along with general observations about the formalisation of mathematics\, including ZFC\, in simple type theory. \n\n\n\n\n\n\n\n\n 
URL:https://cmsa.fas.harvard.edu/event/3-31-2021-new-tech-in-math/
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-03.31.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210324T150000
DTEND;TZID=America/New_York:20210324T160000
DTSTAMP:20260511T093358
CREATED:20240126T083019Z
LAST-MODIFIED:20240515T195432Z
UID:10001411-1616598000-1616601600@cmsa.fas.harvard.edu
SUMMARY:Word and Graph Embeddings for Machine Learning
DESCRIPTION:Speaker: Steve Skiena\, Dept. of Computer Science and AI Insititute\, Stony Brook University \nTitle: Word and Graph Embeddings for Machine Learning \nAbstract: DeepWalk is an approach we have developed to construct vertex embeddings: vector representations of vertices which be applied to a very general class of problems in data mining and information retrieval. DeepWalk exploits an appealing analogy between sentences as sequences of words and random walks as sequences of vertices to transfer deep learning (unsupervised feature learning) techniques from natural language processing to network analysis. It has become extremely popular\, having been cited by over 4600 research papers since its publication at KDD 2014. In this talk\, I will introduce the notion of graph embeddings\, and demonstrate why they make such powerful features for machine learning applications. I will focus on more recent efforts concerning (1) fast embedding methods for very large networks\, (2) techniques for embedding dynamic graphs\, and (3) embedding spaces as models for knowledge generation. \n  \n 
URL:https://cmsa.fas.harvard.edu/event/3-24-2021-new-tech-in-math-seminar/
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-03.24.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210310T150000
DTEND;TZID=America/New_York:20210310T160000
DTSTAMP:20260511T093358
CREATED:20240126T083441Z
LAST-MODIFIED:20240517T194750Z
UID:10001414-1615388400-1615392000@cmsa.fas.harvard.edu
SUMMARY:The Ramanujan Machine: Using Algorithms for the Discovery of Conjectures on Mathematical Constants
DESCRIPTION:Speaker: Ido Kaminer\, Technion – Israel Institute of Technology\, Faculty of Electrical Engineering \nTitle: The Ramanujan Machine: Using Algorithms for the Discovery of Conjectures on Mathematical Constants \nAbstract: In the past\, new conjectures about fundamental constants were discovered sporadically by famous mathematicians such as Newton\, Euler\, Gauss\, and Ramanujan. The talk will present a different approach – a systematic algorithmic approach that discovers new mathematical conjectures on fundamental constants. We call this approach “the Ramanujan Machine”. The algorithms found dozens of well-known formulas as well as previously unknown ones\, such as continued fraction representations of π\, e\, Catalan’s constant\, and values of the Riemann zeta function. Part of the conjectures were in retrospect simple to prove\, whereas others remained so far unproved. We will discuss these puzzles and wider open questions that arose from this algorithmic investigation – specifically\, a newly-discovered algebraic structure that seems to generalize all the known formulas and connect between fundamental constants. We will also discuss two algorithms that proved useful in finding conjectures: a variant of the meet-in-the-middle algorithm and a gradient descent algorithm tailored to the recurrent structure of continued fractions. Both algorithms are based on matching numerical values; consequently\, they conjecture formulas without providing proofs or requiring prior knowledge of the underlying mathematical structure. This way\, our approach reverses the conventional usage of sequential logic in formal proofs; instead\, using numerical data to unveil mathematical structures and provide leads to further mathematical research.
URL:https://cmsa.fas.harvard.edu/event/3-10-2021-new-tech-in-math/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-03.10.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210303T150000
DTEND;TZID=America/New_York:20210303T160000
DTSTAMP:20260511T093358
CREATED:20240126T084416Z
LAST-MODIFIED:20240517T194704Z
UID:10001419-1614783600-1614787200@cmsa.fas.harvard.edu
SUMMARY:Neural Theorem Proving in Lean using Proof Artifact Co-training and Language Models
DESCRIPTION:Speaker: Jason Rute\, CIBO Technologies \nTitle: Neural Theorem Proving in Lean using Proof Artifact Co-training and Language Models \nAbstract: Labeled data for imitation learning of theorem proving in large libraries of formalized mathematics is scarce as such libraries require years of concentrated effort by human specialists to be built. This is particularly challenging when applying large Transformer language models to tactic prediction\, because the scaling of performance with respect to model size is quickly disrupted in the data-scarce\, easily-overfitted regime. We propose PACT ({\bf P}roof {\bf A}rtifact {\bf C}o-{\bf T}raining)\, a general methodology for extracting abundant self-supervised data from kernel-level proof terms for co-training alongside the usual tactic prediction objective. We apply this methodology to Lean\, an interactive proof assistant which hosts some of the most sophisticated formalized mathematics to date. We instrument Lean with a neural theorem prover driven by a Transformer language model and show that PACT improves theorem proving success rate on a held-out suite of test theorems from 32% to 48%.
URL:https://cmsa.fas.harvard.edu/event/3-3-2021-new-tech-in-math/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-03.03.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210224T150000
DTEND;TZID=America/New_York:20210224T160000
DTSTAMP:20260511T093358
CREATED:20240126T085540Z
LAST-MODIFIED:20240517T194101Z
UID:10001422-1614178800-1614182400@cmsa.fas.harvard.edu
SUMMARY:A Mathematical Language
DESCRIPTION:  \nSpeaker: Thomas Hales\, Univ. of Pittsburgh Dept. of Mathematics \nTitle: A Mathematical Language \nAbstract: A controlled natural language for mathematics is an artificial language that is designed in an explicit way with precise computer-readable syntax and semantics.  It is based on a single natural language (which for us is English) and can be broadly understood by mathematically literate English speakers.  This talk will describe the design of a controlled natural language for mathematics that has been influenced by the Lean theorem prover\, by TeX\, and by earlier controlled natural languages. The semantics are provided by dependent type theory.
URL:https://cmsa.fas.harvard.edu/event/2-24-2021-new-technologies-in-mathematics/
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-02.24.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210210T150000
DTEND;TZID=America/New_York:20210210T160000
DTSTAMP:20260511T093358
CREATED:20240126T092338Z
LAST-MODIFIED:20240515T195824Z
UID:10001434-1612969200-1612972800@cmsa.fas.harvard.edu
SUMMARY:A Mathematical Exploration of Why Language Models Help Solve Downstream Tasks
DESCRIPTION:Speaker: Nikunj Saunshi\, Dept. of Computer Science\, Princeton University \nTitle: A Mathematical Exploration of Why Language Models Help Solve Downstream Tasks \nAbstract: Autoregressive language models pretrained on large corpora have been successful at solving downstream tasks\, even with zero-shot usage. However\, there is little theoretical justification for their success. This paper considers the following questions: (1) Why should learning the distribution of natural language help with downstream classification tasks? (2) Why do features learned using language modeling help solve downstream tasks with linear classifiers? For (1)\, we hypothesize\, and verify empirically\, that classification tasks of interest can be reformulated as next word prediction tasks\, thus making language modeling a meaningful pretraining task. For (2)\, we analyze properties of the cross-entropy objective to show that eps-optimal language models in cross-entropy (log-perplexity) learn features that are O(sqrt{eps}) good on such linear classification tasks\, thus demonstrating mathematically that doing well on language modeling can be beneficial for downstream tasks. We perform experiments to verify assumptions and validate our theoretical results. Our theoretical insights motivate a simple alternative to the cross-entropy objective that performs well on some linear classification tasks. \n  \n 
URL:https://cmsa.fas.harvard.edu/event/2-10-2021-new-tech-in-math/
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-02.10.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210127T150000
DTEND;TZID=America/New_York:20210127T160000
DTSTAMP:20260511T093358
CREATED:20240126T092449Z
LAST-MODIFIED:20240515T191649Z
UID:10001435-1611759600-1611763200@cmsa.fas.harvard.edu
SUMMARY:Knowledge graph representation: From recent models towards a theoretical understanding
DESCRIPTION:Speaker: Carl Allen and Ivana Balažević – University of Edinburgh School of Informatics \nTitle: Knowledge graph representation: From recent models towards a theoretical understanding \nAbstract: Knowledge graphs (KGs)\, or knowledge bases\, are large repositories of facts in the form of triples (subject\, relation\, object)\, e.g. (Edinburgh\, capital_of\, Scotland). Many models have been developed to succinctly represent KGs such that known facts can be recalled (question answering) and\, more impressively\, previously unknown facts can be inferred (link prediction). Subject and object entities are typically represented as vectors in R^d and relations as mappings (e.g. linear transformations) between them. Such representation can be interpreted as positioning entities in a space such that relations are implied by their relative locations. In this talk we give an overview of knowledge graph representation including select recent models; and\, by drawing a connection to word embeddings\, explain a theoretical model for how semantic relationships can correspond to geometric structure.
URL:https://cmsa.fas.harvard.edu/event/1-27-2021-new-tech-in-math-seminar/
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-01.27.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210120T150000
DTEND;TZID=America/New_York:20210120T160000
DTSTAMP:20260511T093358
CREATED:20240126T093733Z
LAST-MODIFIED:20240515T191339Z
UID:10001440-1611154800-1611158400@cmsa.fas.harvard.edu
SUMMARY:Language Modeling for Mathematical Reasoning
DESCRIPTION:Speaker: Christian Szegedy \nTitle: Language Modeling for Mathematical Reasoning \nAbstract: In this talk\, I will summarize the current state of the art of transformer based language models and give examples on non-trivial reasoning task language models can solve in higher order logic reasoning. I will also discuss how to inject injective bias into transformer networks via pretraining on very simple synthetic tasks and representing graph structures for transformer networks. \n 
URL:https://cmsa.fas.harvard.edu/event/1-20-2021-new-tech-in-math/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-01.20.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210113T150000
DTEND;TZID=America/New_York:20210113T160000
DTSTAMP:20260511T093358
CREATED:20240126T093843Z
LAST-MODIFIED:20240517T201031Z
UID:10001441-1610550000-1610553600@cmsa.fas.harvard.edu
SUMMARY:AI and Theorem Proving
DESCRIPTION:Speaker: Josef Urban\, Czech Technical University \nTitle: AI and Theorem Proving \nAbstract: The talk will discuss the main approaches that combine machine learning with automated theorem proving and automated formalization. This includes learning to choose relevant facts for “hammer” systems\, guiding the proof search of tableaux and superposition automated provers by interleaving learning and proving (reinforcement learning) over large ITP libraries\, guiding the application of tactics in interactive tactical systems\, and various forms of lemmatization and conjecturing. I will also show some demos of the systems\, and discuss autoformalization approaches such as learning probabilistic grammars from aligned informal/formal corpora\, combining them with semantic pruning\, and using neural methods to learn direct translation from Latex to formal mathematics.
URL:https://cmsa.fas.harvard.edu/event/1-13-2021-new-technologies-in-mathematics/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-01.13.21.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20201209T150000
DTEND;TZID=America/New_York:20201209T160000
DTSTAMP:20260511T093358
CREATED:20240127T015429Z
LAST-MODIFIED:20240515T200218Z
UID:10001466-1607526000-1607529600@cmsa.fas.harvard.edu
SUMMARY:Machine learning and su(3) structures on six manifolds
DESCRIPTION:Speaker: James Gray – Virginia Tech \nTitle: Machine learning and su(3) structures on six manifolds \nAbstract: In this talk we will discuss the application of Machine Learning techniques to obtain numerical approximations to various metrics of SU(3) structure on six manifolds. More precisely\, we will be interested in SU(3) structures whose torsion classes make them suitable backgrounds for various string compactifications. A variety of aspects of this topic will be covered. These will include learning moduli dependent Ricci-Flat metrics on Calabi-Yau threefolds and obtaining numerical approximations to torsional SU(3) structures. \n 
URL:https://cmsa.fas.harvard.edu/event/12-9-2020-new-tech-in-math/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-12.09.20.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20201118T150000
DTEND;TZID=America/New_York:20201118T160000
DTSTAMP:20260511T093358
CREATED:20240127T020145Z
LAST-MODIFIED:20240515T200420Z
UID:10001472-1605711600-1605715200@cmsa.fas.harvard.edu
SUMMARY:Universes as Big Data\, or Machine-Learning Mathematical Structures
DESCRIPTION:Speaker: Yang-Hui He\, Oxford University\, City University of London and Nankai University \nTitle: Universes as Big Data\, or Machine-Learning Mathematical Structures \nAbstract: We review how historically the problem of string phenomenology lead theoretical physics first to algebraic/differetial geometry\, and then to computational geometry\, and now to data science and AI. With the concrete playground of the Calabi-Yau landscape\, accumulated by the collaboration of physicists\, mathematicians and computer scientists over the last 4 decades\, we show how the latest techniques in machine-learning can help explore problems of physical and mathematical interest\, from geometry\, to group theory\, to combinatorics and number theory. \n  \n 
URL:https://cmsa.fas.harvard.edu/event/11-18-2020-new-tech-in-math/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-11.18.20.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20201111T150000
DTEND;TZID=America/New_York:20201111T160000
DTSTAMP:20260511T093358
CREATED:20240127T021159Z
LAST-MODIFIED:20240515T200604Z
UID:10001480-1605106800-1605110400@cmsa.fas.harvard.edu
SUMMARY:Towards AI for mathematical modeling of complex biological systems: Machine-learned model reduction\, spatial graph dynamics\, and symbolic mathematics
DESCRIPTION:Speaker: Eric Mjolsness\, Departments of Computer Science and Mathematics\, UC Irvine \nTitle: Towards AI for mathematical modeling of complex biological systems: Machine-learned model reduction\, spatial graph dynamics\, and symbolic mathematics \nAbstract: The complexity of biological systems (among others) makes demands on the complexity of the mathematical modeling enterprise that could be satisfied with mathematical artificially intelligence of both symbolic and numerical flavors. Technologies that I think will be fruitful in this regard include (1) the use of machine learning to bridge spatiotemporal scales\, which I will illustrate with the “Dynamic Boltzmann Distribution” method for learning model reduction of stochastic spatial biochemical networks and the “Graph Prolongation Convolutional Network” approach to course-graining the biophysics of microtubules; (2) a meta-language for stochastic spatial graph dynamics\, “Dynamical Graph Grammars”\, that can represent structure-changing processes including microtubule dynamics and that has an underlying combinatorial theory related to operator algebras; and (3) an integrative conceptual architecture of typed symbolic modeling languages and structure-preserving maps between them\, including model reduction and implementation maps. \n  \n  \n 
URL:https://cmsa.fas.harvard.edu/event/11-11-2020-new-technologies-in-mathematics/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-11.11.20-1.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20201104T150000
DTEND;TZID=America/New_York:20201104T160000
DTSTAMP:20260511T093358
CREATED:20240127T021940Z
LAST-MODIFIED:20240515T200835Z
UID:10001486-1604502000-1604505600@cmsa.fas.harvard.edu
SUMMARY:Some exactly solvable models for machine learning via Statistical physics
DESCRIPTION:Speaker: Florent Krzakala\, EPFL \nTitle: Some exactly solvable models for machine learning via Statistical physics \nAbstract: The increasing dimensionality of data in the modern machine learning age presents new challenges and opportunities. The high dimensional settings allow one to use powerful asymptotic methods from probability theory and statistical physics to obtain precise characterizations and develop new algorithmic approaches. Statistical mechanics approaches\, in particular\, are very well suited for such problems. Will give examples of recent works in our group that build on powerful methods of statistical physics of disordered systems to analyze some relevant questions in machine learning and neural networks\, including overparameterization\, kernel methods\, and the behavior gradient descent algorithm in a high dimensional non-convex landscape.
URL:https://cmsa.fas.harvard.edu/event/11-4-2020-new-technologies-in-math/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-11.04.20.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20201028T150000
DTEND;TZID=America/New_York:20201028T160000
DTSTAMP:20260511T093358
CREATED:20240127T031359Z
LAST-MODIFIED:20240515T201157Z
UID:10001502-1603897200-1603900800@cmsa.fas.harvard.edu
SUMMARY:Generalization bounds for rational self-supervised learning algorithms\, or "Understanding generalizations requires rethinking deep learning"
DESCRIPTION:Speakers: Boaz Barak and Yamini Bansal\, Harvard University Dept. of Computer Science \nTitle: Generalization bounds for rational self-supervised learning algorithms\, or “Understanding generalizations requires rethinking deep learning” \nAbstract: The generalization gap of a learning algorithm is the expected difference between its performance on the training data and its performance on fresh unseen test samples. Modern deep learning algorithms typically have large generalization gaps\, as they use more parameters than the size of their training set. Moreover the best known rigorous bounds on their generalization gap are often vacuous. In this talk we will see a new upper bound on the generalization gap of classifiers that are obtained by first using self-supervision to learn a complex representation of the (label free) training data\, and then fitting a simple (e.g.\, linear) classifier to the labels. Such classifiers have become increasingly popular in recent years\, as they offer several practical advantages and have been shown to approach state-of-art results. We show that (under the assumptions described below) the generalization gap of such classifiers tends to zero as long as the complexity of the simple classifier is asymptotically smaller than the number of training samples. We stress that our bound is independent of the complexity of the representation that can use an arbitrarily large number of parameters. Our bound assuming that the learning algorithm satisfies certain noise-robustness (adding small amount of label noise causes small degradation in performance) and rationality (getting the wrong label is not better than getting no label at all) conditions that widely (and sometimes provably) hold across many standard architectures. We complement this result with an empirical study\, demonstrating that our bound is non-vacuous for many popular representation-learning based classifiers on CIFAR-10 and ImageNet\, including SimCLR\, AMDIM and BigBiGAN. The talk will not assume any specific background in machine learning\, and should be accessible to a general mathematical audience. Joint work with Gal Kaplun. \n 
URL:https://cmsa.fas.harvard.edu/event/10-28-2020-new-technologies-in-mathematics-seminar/
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-10.28.20.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20201014T150000
DTEND;TZID=America/New_York:20201014T160000
DTSTAMP:20260511T093358
CREATED:20240201T021720Z
LAST-MODIFIED:20240515T192014Z
UID:10001522-1602687600-1602691200@cmsa.fas.harvard.edu
SUMMARY:Triple Descent and a Fine-Grained Bias-Variance Decomposition
DESCRIPTION:Speaker: Jeffrey Pennington\, Google Brain \nTitle: Triple Descent and a Fine-Grained Bias-Variance Decomposition \nAbstract: Classical learning theory suggests that the optimal generalization performance of a machine learning model should occur at an intermediate model complexity\, striking a balance between simpler models that exhibit high bias and more complex models that exhibit high variance of the predictive function. However\, such a simple trade-off does not adequately describe the behavior of many modern deep learning models\, which simultaneously attain low bias and low variance in the heavily overparameterized regime. Recent efforts to explain this phenomenon theoretically have focused on simple settings\, such as linear regression or kernel regression with unstructured random features\, which are too coarse to reveal important nuances of actual neural networks. In this talk\, I will describe a precise high-dimensional asymptotic analysis of Neural Tangent Kernel regression that reveals some of these nuances\, including non-monotonic behavior deep in the overparameterized regime. I will also present a novel bias-variance decomposition that unambiguously attributes these surprising observations to particular sources of randomness in the training procedure.
URL:https://cmsa.fas.harvard.edu/event/10-14-2020-new-technologies-seminar/
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-10.14.20.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20200923T150000
DTEND;TZID=America/New_York:20200923T160000
DTSTAMP:20260511T093358
CREATED:20240209T014918Z
LAST-MODIFIED:20240515T201416Z
UID:10001784-1600873200-1600876800@cmsa.fas.harvard.edu
SUMMARY:Self-induced regularization from linear regression to neural networks
DESCRIPTION:Speaker: Andrea Montanari\, Departments of Electrical Engineering and Statistics\, Stanford \nTitle: Self-induced regularization from linear regression to neural networks \nAbstract: Modern machine learning methods –most noticeably multi-layer neural networks– require to fit highly non-linear models comprising tens of thousands to millions of parameters. Despite this\, little attention is paid to the regularization mechanism to control model’s complexity. Indeed\, the resulting models are often so complex as to achieve vanishing training error: they interpolate the data. Despite this\, these models generalize well to unseen data : they have small test error. I will discuss several examples of this phenomenon\, beginning with a simple linear regression model\, and ending with two-layers neural networks in the so-called lazy regime. For these examples precise asymptotics could be determined mathematically\, using tools from random matrix theory. I will try to extract a unifying picture. A common feature is the fact that a complex unregularized nonlinear model becomes essentially equivalent to a simpler model\, which is however regularized in a non-trivial way. [Based on joint papers with: Behrooz Ghorbani\, Song Mei\, Theodor Misiakiewicz\, Feng Ruan\, Youngtak Sohn\, Jun Yan\, Yiqiao Zhong] \n  \n 
URL:https://cmsa.fas.harvard.edu/event/9-23-2020-new-tech-in-mathematics-seminar/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-09.23.20.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20200916T150000
DTEND;TZID=America/New_York:20200916T160000
DTSTAMP:20260511T093358
CREATED:20240209T021453Z
LAST-MODIFIED:20240515T183741Z
UID:10001799-1600268400-1600272000@cmsa.fas.harvard.edu
SUMMARY:Graph Representation Learning: Recent Advances and Open Challenges
DESCRIPTION:Speaker: William Hamilton\, McGill University and MILA \nTitle: Graph Representation Learning: Recent Advances and Open Challenges \nAbstract: Graph-structured data is ubiquitous throughout the natural and social sciences\, from telecommunication networks to quantum chemistry. Building relational inductive biases into deep learning architectures is crucial if we want systems that can learn\, reason\, and generalize from this kind of data. Recent years have seen a surge in research on graph representation learning\, most prominently in the development of graph neural networks (GNNs). Advances in GNNs have led to state-of-the-art results in numerous domains\, including chemical synthesis\, 3D-vision\, recommender systems\, question answering\, and social network analysis. In the first part of this talk I will provide an overview and summary of recent progress in this fast-growing area\, highlighting foundational methods and theoretical motivations. In the second part of this talk I will discuss fundamental limitations of the current GNN paradigm and propose open challenges for the theoretical advancement of the field. \n 
URL:https://cmsa.fas.harvard.edu/event/9-16-2020-new-technologies-seminar/
LOCATION:Virtual
CATEGORIES:New Technologies in Mathematics Seminar
ATTACH;FMTTYPE=image/png:https://cmsa.fas.harvard.edu/media/CMSA-New-Technologies-in-Mathematics-09.16.20-1.png
END:VEVENT
END:VCALENDAR