Task-Aware Reduction for Scalable LLM–Database Systems
Large Language Models (LLMs) are increasingly applied to data-intensive workflows, from database querying to developer observability. Yet the effectiveness of these systems is constrained by the volume, verbosity, and noise of real-world text-rich data such as logs, telemetry, and monitoring streams. Feeding such data directly into LLMs is costly, environmentally unsustainable, and often misaligned with task objectives. Parallel efforts in LLM efficiency have focused on model- or architecture-level optimizations, but the challenge of reducing upstream input verbosity remains under explored. In this paper, we argue for treating the token budget of an LLM as an attention budget and elevating task-aware text reduction as a first-class design principle for language–data systems. We position input-side reduction not as compression, but as attention allocation: prioritizing information most relevant to downstream tasks. We outline open research challenges for building benchmarks, designing adaptive reduction pipelines, and integrating token-budget–aware preprocessing into database and retrieval systems. Our vision is to channel scarce attention resources toward meaningful signals in noisy, data-intensive workflows, enabling scalable, accurate, and sustainable LLM–data integration.
Wed 12 NovDisplayed time zone: Eastern Time (US & Canada) change
15:00 - 16:30 | WKS-13: LLMs Meet Databases: Next-Generation Data Systems (Part 3)LMD at Room 3 Chair(s): Anastasios Kementsietsidis Google DeepMind | ||
15:00 30mTalk | LLM-Driven Event Log Generation from Forensic Cases: A Comparative Study of ChatGPT, Claude, and Gemini LMD | ||
15:30 30mTalk | Task-Aware Reduction for Scalable LLM–Database Systems LMD Marcus Barnes University of Toronto, Taher A. Ghaleb Trent University, Safwat Hassan University of Toronto, Canada Pre-print | ||
16:00 30mTalk | ScenarioBench: Trace-Grounded Compliance Evaluation for Text-to-SQL and RAG LMD Pre-print | ||