Advith's Calendar

Natural Language Processing Seminar: Chi Han, "Opening and Reprogramming The Black-Box of Large Language Models."

Event Type
Seminar/Symposium
Sponsor
Siebel School of Computing and Data Science
Location
2405 Siebel Center
Virtual
Join online
Date
Dec 10, 2025   1:00 pm  
Speaker
Chi Han
Contact
Allison Mette
E-Mail
agk@illinois.edu
Originating Calendar
Siebel School Speakers Calendar
Abstract:
While large language models (LLMs) continue to revolutionize natural language processing and AI applications, tracing, locating, and precisely addressing their limitations remains a major challenge. The development of future language models would greatly benefit from a structural understanding of LLMs. This presentation brings together several recent papers that systematically explain the internal representations of LLMs from both theoretical and empirical perspectives. These works offer preliminary characterizations of the roles and adaptation of internal components: (1) how positional representations hold clues for resolving the context length limitations of LLMs, (2) how word representations can be used as steers for generation, and (3) how cross-modal representations can be best aligned for scientific discovery. Together, they provide insights into addressing inherent limitations of LLMs in a principled and efficient way, and point to a promising future of developing a modular “brain atlas” for LLMs.


Bio:
Chi Han is currently a final-year Computer Science Ph.D. student in the NLP group at the University of Illinois Urbana-Champaign (UIUC), under the advisory of Prof. Heng Ji. Before joining UIUC, he was an undergraduate student at Tsinghua University, China, in the Yao Class program. He visited the CoCoSci Lab at the Massachusetts Institute of Technology (MIT) during his undergraduate studies. He has first-authored papers in top conferences, including NeurIPS, ICLR, ACL, and NAACL, and received first-authored outstanding paper awards in NAACL 2024 and ACL 2024, and received IBM PhD Fellowship and Amazon AICE PhD Fellowship. His research interests are centered around a theoretical understanding of representations in foundation models (FMs), with the aim of providing insights and tools for efficient, controllable, and interpretable foundation models.
link for robots only