ASE 2024 (series) / Student Research Competition /
Calibration of Large Language Models for Code Summarization
Tue 29 Oct 2024 13:00 - 13:15 at Bondi - SRC Presentations
High-quality code summaries help developers understand code, but they are often missing or require significant effort to write. Automatic code summarization can generate summaries that resemble human-written ones, with Large Language Models (LLMs) achieving state-of-the-art performance. Yet, they often fail and produce summaries unlike what a human would say. Can we gauge the likelihood that a LLM-generated summary is sufficiently similar to what a human would write? We examine whether the token-level confidences of several LLMs indicate this notion of summary quality. We suggest a method to produce well-calibrated likelihoods that a LLM-generated summary is similar to a human-written one.
Tue 29 OctDisplayed time zone: Pacific Time (US & Canada) change
Tue 29 Oct
Displayed time zone: Pacific Time (US & Canada) change
13:00 - 13:45 | |||
13:00 15mTalk | Calibration of Large Language Models for Code Summarization Student Research Competition Yuvraj Virk UC Davis | ||
13:15 15mTalk | Can Large Language Models Comprehend Code Stylometry? Student Research Competition | ||
13:30 15mTalk | Semi-Automated Verification of Interior Unsafe Code Encapsulation in Real-World Rust Systems Student Research Competition Zihao Rao Fudan University |