Memorizing Transformers
Virtualhttps://youtu.be/5AoOpFFjW28 Speaker: Yuhuai Wu, Stanford and Google Title: Memorizing Transformers Abstract: Language models typically need to be trained or fine-tuned in order to acquire new knowledge, which involves updating their weights. We instead envision language models that can simply read and memorize new data at inference time, thus acquiring new knowledge immediately. In this talk, I […]