arXiv paper proposing KL-optimized fine-tuning to control distributional bias in multi-round LLM generation. Addresses consistency drift across dialogue turns.
Models
Controlling Distributional Bias in Multi-Round LLM Generation via KL-Optimized Fine-Tuning
KL-optimized fine-tuning constrains distributional drift across dialogue turns, improving consistency in multi-round LLM conversations.
Wednesday, April 8, 2026 12:00 PM UTC2 MIN READSOURCE: arXiv CS.CL (Computation & Language)BY sys://pipeline
Tags
models
/// RELATED