Language Modeling with Few-shot Language Feedback

Loading...
Thumbnail Image
Author(s)
Richardson, Christopher
Advisor(s)
Editor(s)
Associated Organization(s)
Series
Supplementary to:
Abstract
The objective of the proposed research is to develop methods to leverage few-shot language feedback to improve language models on various tasks. The motivation for this work stems from the overarching goal in the artificial intelligence (AI) field of achieving alignment - that is, AI that advances the intended objectives of its users. Recent progress with instruction-tuned language models has given rise to capable chatbot agents that can solve myriad tasks described in natural language and engage with users in a conversational setting. Despite these advances, language models still face fundamental challenges in aligning with human objectives. In particular, they must continuously adapt and learn from human users. Current models can respond to direct feedback conversationally, but have limited abilities to generalize from that feedback to new situations and unseen data. This is the problem of few-shot language feedback. We seek to better understand the capabilities and limits of current language models in generalizing from language feedback and to develop methods to improve responses in the few-shot language feedback scenario.
Sponsor
Date
2025-04-23
Extent
Resource Type
Text
Resource Subtype
Dissertation
Rights Statement
Rights URI