
The researchers behind the BoT project are from Peking University, UC Berkeley, and Stanford University. To find more information about their work, you can check out the paper and GitHub repository associated with the project. Additionally, you can follow them on Twitter and join their Telegram Channel, Discord Channel, and LinkedIn Group for updates and discussions on their research.

Single-query and multi-query reasoning systems in Large Language Models (LLMs) have certain limitations. For single-query reasoning systems, it is impractical to manually design them task by task as they typically rely on prior assumptions or relevant exemplars of reasoning processes1. On the other hand, multi-query reasoning systems are computationally intensive because they recursively expand reasoning paths to find a unique intrinsic structure for each task. Both types of systems are limited by their reasoning structures and exemplars, failing to derive general and high-level guidelines or thoughts from previously completed tasks, which could be useful for improving efficiency and accuracy when solving similar problems4.

The two primary methods used to enhance the reasoning performance of Large Language Models (LLMs) as mentioned in the article are:
More effective prompting methods: This involves designing better prompts to guide the LLMs in solving problems. Prompt engineering has been used to improve the performance of LLMs in various tasks.
Increasing the model size: By increasing the number of parameters in the LLMs, their reasoning performance can be boosted. This is because larger models have more capacity to learn and reason about complex problems.
These methods have been successful in improving the reasoning abilities of LLMs, such as GPT-4, PaLM, and LLaMA. However, they have certain limitations, such as the need for manual design of prompts and the computational intensity of larger models. To overcome these limitations, the researchers have proposed a novel approach called the Buffer of Thoughts (BoT), which enhances the reasoning accuracy, efficiency, and resilience of LLMs across a wide range of tasks.