Large Language Models (LLMs) have demonstrated remarkable human-level natural language generation capabilities. However, their potential to generate misinformation, often called the hallucination problem, poses a significant risk to their deployment. A common approach to address this issue is to retrieve relevant knowledge and fine-tune the LLM with the knowledge in its input. Unfortunately, this method incurs high training costs and may cause catastrophic forgetting for multi-tasking models. To overcome these limitations, this thesis propose a knowledge-constrained decoding method called KCTS (Knowledge-Constrained Tree Search), which guides a frozen LM to generate text aligned with the reference knowledge at each decoding step using a knowledge classifier score and a future reward-aware optimization algorithm, MCTS (Monte-Carlo Tree Search). To adapt the sequence-level knowledge classifier to token-level guidance, a novel token-level hallucination detection method called RIPA (Reward Inflection Point Approximation) is also proposed. The empirical results on knowledge-grounded dialogue and abstractive summarization demonstrate the strength of KCTS as a plug-and-play, model-agnostic decoding method that can effectively reduce hallucinations in natural language generation.
| Date of Award | 2024 |
|---|
| Original language | English |
|---|
| Awarding Institution | - The Hong Kong University of Science and Technology
|
|---|
| Supervisor | Yangqiu SONG (Supervisor) |
|---|
Knowledge grounded generation with inference-time optimization and constrained decoding
CHOI, S. (Author). 2024
Student thesis: Master's thesis