Pretraining, Instruction Tuning, Alignment, Specialization: On the Source of Large Language Model Abilities
Speaker
Yao Fu, University of Edinburgh
Time
2023-02-14 12:22:00 ~ 2023-02-14 12:22:00
Location
电院4号楼E谷TED讲座厅
Host
林洲汉
Abstract
Recently, the field has been greatly impressed and inspired by Large Language Models (LLMs) like GPT-3.5. The LLMs' multi-dimensional abilities are significantly beyond many NLP researchers’ and practitioners’ expectations and thus reshaping the research paradigm of NLP. A natural question is how LLMs get there, and where these fantastic abilities come from. In this talk we try to dissect the strong LLM abilities and trace them to their sources, hoping to give a comprehensive roadmap about the evolution of LLMs.
Bio
Yao Fu is a Ph.D. student at University of Edinburgh and a student researcher at Allen Institute for AI. Previously he finished his M.S. in Columbia University and B.S. in Peking University. Yao studies large scale probabilistic generative models for human language. His publication covers topics of large language models, emergent abilities, and complex reasoning.