Enhancing LLM Capabilities
Beyond Scaling Up
(Nov. 15, 2024; EMNLP'24 Tutorial)
Contact: wenpeng@psu.edu
Penn State
UC Davis
Penn State
ASU
USC
UPenn & Oracle
Abstract:
General-purpose large language models (LLMs) are progressively expanding both in scale and access to unpublic training data. This has led to notable progress in a variety of AI problems. Nevertheless, two questions exist: i) Is scaling up the sole avenue of extending the capabilities of LLMs? ii) Instead of developing general-purpose LLMs, how to endow LLMs with specific knowledge? This tutorial targets researchers and practitioners who are interested in capability extension of LLMs that go beyond scaling up. To this end, we will discuss several lines of research that follow that direction, including: (i) optimizing input prompts to fully exploit LLM potential, (ii) enabling LLMs to self-improve responses through various feedback signals, (iii) updating or editing the internal knowledge of LLMs when necessary, (iv) leveraging incidental structural supervision from target tasks, and (v) defending against potential attacks and threats from malicious users. At last, we will conclude the tutorial by outlining directions for further investigation.
Dan Roth
Progress of current LLMs
Scaling trend of LLMs
From general-purpose LLMs to domain-specific LLMs
Rui Zhang
Search-based prompt optimization
Text gradient–based prompt optimization
Gradient-based prompt optimization
Wenpeng Yin
Prompting for LLM Intrinsic Self-improvement
Finetuning for LLM Intrinsic Self-improvement
Fei Wang
Examine the issues caused by unreliable knowledge, such as hallucinations
Remedy LLMs’ internal knowledge by integrating external information in a training-free manner
LLM knowledge editing with lightweight tuning
Ben Zhou
Symbolic constraints as structures (e.g., human-written, mathematical constraints, and compiler constraints)
Structures from decomposing the target problem
Procedural structures that come from cognitive or problem-solving processes, such as DSP, ReAct, and RAP.
Muhao Chen
Introducing inference-time threats (e.g., prompt injection, malicious task instructions, jailbreaking attacks, adversarial demonstrations, and training-free backdoor attacks)
Defense techniques (e.g., prompt robustness estimation, demonstration-based defense, and ensemble debiasing)
Dan Roth
Augmentation via MoE, Agents/RAG/Expert models
Retrieval Augmented Generation