New Research Reveals Why Fine-Tuning Causes Hallucinations in LLMs
Fine-tuning large language models can increase hallucinations by degrading pre-trained knowledge. Researchers propose a self-distillation method to mitigate this issue. This could reshape how we approach model training.

A new study published on arXiv (2604.15574v1) reveals that supervised fine-tuning (SFT) of large language models (LLMs) often leads to increased hallucinations. The research shows that exposure to new factual information during SFT can degrade the model's pre-trained knowledge, making it more prone to generating factually incorrect statements.
The findings are significant because they challenge the common practice of fine-tuning models to adapt them to specific tasks. By using tools from the continual learning literature, the researchers propose a self-distillation-based SFT method that aims to preserve the model's original knowledge while incorporating new information. This approach could be a game-changer for developers looking to optimize their models.
The research team suggests that this method could be widely adopted in the industry, potentially leading to more reliable and accurate LLMs. However, further testing and validation will be necessary to determine its effectiveness across different types of models and tasks. The study opens up new avenues for improving model training practices and reducing hallucinations in AI-generated content.