🎓Jifan Zhang
🏛️University of Wisconsin-Madison
📅Tuesday, October 29th, 2024 | 12:30 PM - 1:30 PM (EDT)
📍C341 Classroom in Van Leer
📚Talk Title: Learning from Black-Box General Intelligences
📎Slides
📝Abstract
General intelligences, both human and artificial (e.g. LLMs), offer remarkable flexibility in handling diverse tasks. However, directly leveraging these general intelligences at scale is prohibitively expensive. This raises the key question of how we can efficiently train lightweight, specialized models for specific applications by learning from and distilling the knowledge of black-box general intelligences. In this talk, I will discuss the label-efficient learning paradigms that have been developed over the past two decades, covering techniques in active learning, semi-supervised learning, and transfer learning. I will highlight scenarios and approaches that have proven empirically effective for label-efficient learning, including fine-tuning large pretrained models, uncertainty sampling and handling class imbalance. I will conclude by discussing the challenges and growing importance of label-efficient learning in an open-world scenario. This talk will provide an overview of the key ideas, results, and open problems in learning efficiently from black-box general intelligences.
👤Bio
Jifan Zhang is a Ph.D. candidate in computer science at the University of Wisconsin, working with Robert Nowak. He obtained his M.S. and B.S. degrees in computer science from the University of Washington. During that time, he was advised by Kevin Jamieson, Lalit Jain, Tanner Schmidt, Dieter Fox, and Zachary Tatlock. His research focuses on both applied and theoretical perspectives of Machine Learning, primarily on alignment of LLMs, humor generation with LLMs, and efficient distillation of black box intelligence.