In the field of large language models (LLMs), multi-agent debates (MAD) pose a significant challenge due to their high computational costs. They involve multiple agents communicating with one another, all referencing each other's solutions. Despite attempts to improve LLM performance through Chain-of-Thought (CoT) prompting and self-consistency, these methods are still limited by the increased complexity…
