Belief Graphs Boost Weak LLMs in Multi-Agent Reasoning
A new study shows that belief graphs significantly improve weaker LLMs in cooperative reasoning tasks, while stronger models see minimal benefits. The research highlights the importance of integration architecture in leveraging these graphs effectively.

Researchers have discovered that explicit belief graphs can enhance the performance of weaker large language models (LLMs) in cooperative multi-agent reasoning tasks. The study, conducted through over 3,000 controlled trials using the cooperative card game Hanabi, compared four LLM families. The findings reveal that belief graphs are only beneficial for weaker models when integrated as prompt context, particularly in tasks requiring 2nd-order Theory of Mind (80% vs 10%, p<0.0001, OR=36.0).
The key takeaway is that the integration architecture determines the value of belief graphs. For strong models, these graphs are merely decorative when used as prompt context. However, when belief graphs are used to gate action selection through ranked shortlists, they become structurally beneficial, indicating a more effective use of these tools in cooperative reasoning scenarios.
The research opens new avenues for optimizing LLM performance in multi-agent settings. Future studies could explore how different integration methods affect belief graph efficacy across various tasks and model strengths. This could lead to more tailored approaches for leveraging belief graphs in real-world applications, particularly where cooperative reasoning is crucial.