Skip to main content

Toward A Foundation of Reprogramming Large Language Models by Chi Han

Event Details

Date
Friday, February 27, 2026
Time
3-4 p.m.
Location
Description

Abstract: Large language models (LLMs) have achieved strong performance in domains covered by pre-training corpora. However, once pre-trained, their internal computations are largely treated as 
fixed black-boxes, and systematically modifying them remains an open challenge. As a result, 
when LLMs exhibit limitations in open-world settings, such as extreme-scale contexts, novel 
scenario generation, and scientific discovery, we lack principled methods to intervene at the 
internal mechanistic level where these failures originate. In this talk, I present a quest toward a 
principled foundation for LLM reprogramming by uncovering governing principles of their 
internal computations and developing methods for post-hoc reprogramming of LLMs.
My approach integrates two complementary components: theoretical analyses of how LLMs 
process context and generate decisions, and empirical methods that reprogram internal 
architectures with minimal additional resources. First, I present LM-Infinite, which provides a 
mechanistic explanation of context-length failure and enables generalization to contexts 
exceeding 200 million tokens without retraining. Second, in LM-Steer, I show that word 
embeddings act as controllable steers for generation, enabling efficient, interpretable, and 
transferable control over LLM outputs. Third, extending beyond language, I introduce a 
modular chemical language model that incorporates domain-grounded representations to 
support synthesis-aware molecular reasoning and drug discovery. Together, my research sheds 
light on principled mechanisms for making LLMs more interpretable, adaptable, and reliable 
across scientific disciplines and real-world settings.

Bio: Chi Han is a final-year Ph.D. candidate in Computer Science at the University of Illinois 
Urbana-Champaign (UIUC), where he works in the NLP group under the supervision of Prof. 
Heng Ji. He received his undergraduate degree from Tsinghua University, China, through the 
Yao Class program. His research has led to first-author publications at top venues including 
NeurIPS, ICLR, TMLR, ACL, and NAACL, which received Outstanding Paper Awards at NAACL 
2024 and ACL 2024, as well as the Best Demo Award (1st place) at the NSF Summit for AI 
Institutes Leadership (SAIL). His research has been supported by the IBM PhD Fellowship, 
Amazon AICE PhD Fellowship, the Mavis Future Faculty Fellowship, and Capital One seeding 
grant. His research interests focus on developing theoretical foundations and empirical 
methods for post-hoc reprogramming of large language models (LLMs). His work addresses 
intrinsic limitations of LLMs across multiple domains, including context length, decisionmaking transparency, generation control, and scientific discovery, and achieves strong 
performance across downstream tasks

Cost
Free
Accessibility

We value inclusion and access for all participants and are pleased to provide reasonable accommodations for this event. Please email junjie.hu@wisc.edu to make a disability-related accommodation request. Reasonable effort will be made to support your request.

Tags