Skip to main content
UNIVERSITY OF ILLINOIS URBANA-CHAMPAIGN
Toggle navigation
Research
Dropdown menu toggle
Artificial Intelligence
Arts and Humanities
Astrophysics
Digital Agriculture
Earth and Environment
Engineering
Health Sciences
Project Highlights
Expertise
Dropdown menu toggle
Compute Resources
Data Analytics
Facilities
Innovative Systems
Integrated Cyberinfrastructure
Program Administration
Software and Applications
User Services
Visualization
People
Dropdown menu toggle
Leadership
Directorates
Staff Directory
News & Events
Dropdown menu toggle
News
Calendar
Press Room
Tours
About
Dropdown menu toggle
Careers
Fellowships & Internships
Industry Partner Program
Institutional Partnerships
Diversity
History
Giving
Contact
Search
Search
Toggle navigation
Calendar
National Center for Supercomputing Applications WordPress Master Calendar
Share on Facebook
Tweet
Email
add to calendar
contact
add an event
View Full Calendar
NCSA staff who would like to submit an item for the calendar can email
newsdesk@ncsa.illinois.edu
.
Horizon-Length Prediction: Advancing Fill-in-the-Middle Capabilities for Code Generation with Lookahead Planning
Event Type
Seminar/Symposium
Sponsor
PL/FM/SE
Location
0222 Siebel Center and Zoom
Virtual
Date
Oct 11, 2024 2:00 - 3:00 pm
Speaker
Yifeng Ding, UIUC
Contact
Kristin Irle
E-Mail
kirle@illinois.edu
Phone
217-244-0229
Views
24
Originating Calendar
Siebel School Speakers Calendar
Abstract
: Fill-in-the-Middle (FIM) has become integral to code language models, enabling generation of missing code given both left and right contexts. However, the current FIM training paradigm, which reorders original training sequences and then performs regular next-token prediction (NTP), often leads to models struggling to generate content that aligns smoothly with the surrounding context. Crucially, while existing works rely on rule-based post-processing to circumvent this weakness, such methods are not practically usable in open-domain code completion tasks as they depend on restrictive, dataset-specific assumptions (e.g., generating the same number of lines as in the ground truth). Moreover, model performance on FIM tasks deteriorates significantly without these unrealistic assumptions.
We hypothesize that NTP alone is insufficient for models to learn effective planning conditioned on the distant right context, a critical factor for successful code infilling. To overcome this, we propose Horizon-Length Prediction (HLP), a novel training objective that teaches models to predict the number of remaining middle tokens (i.e., horizon length) at each step. HLP advances FIM with lookahead planning, enabling models to inherently learn infilling boundaries for arbitrary left and right contexts without relying on dataset-specific post-processing. Our evaluation across different models and sizes shows that HLP significantly improves FIM performance by up to 24% relatively on diverse benchmarks, across file-level and repository-level, and without resorting to unrealistic post-processing methods. Furthermore, the enhanced planning capability gained through HLP boosts model performance on code reasoning. Importantly, HLP only incurs negligible training overhead and no additional inference cost, ensuring its practicality for real-world scenarios.
link for robots only
Back to top