Ensembling LLM’s “The Blender LLM”

Mahesh Patel
3 min readJul 1, 2023

--

https://arxiv.org/abs/2306.02561

The Blender LLM paper introduces a novel #ensembling framework for large language models (LLMs). The framework, called Blender, leverages the diverse strengths of multiple LLMs to attain consistently high performance on a variety of tasks.

Blender works by first training a meta-model on a dataset of human-generated text and code. This meta-model learns to identify the strengths and weaknesses of each LLM in the ensemble. When a new task is presented to Blender, it uses the meta-model to select the most appropriate LLMs to contribute to the final output.

https://arxiv.org/abs/2306.02561

Blender LLM uses two training methods:

  • PairRanker: This method trains a meta-model to distinguish between good and bad outputs from different LLMs. The meta-model is trained on a dataset of human-generated text and code, and it learns to identify the strengths and weaknesses of each LLM in the ensemble.
  • GenFuser: This method fuses the outputs of the top-ranked LLMs from PairRanker into a single, improved output. GenFuser uses a technique called “attention” to combine the strengths of the different LLMs and mitigate their weaknesses.

The two training methods work together to produce a Blender LLM that can consistently outperform individual LLMs on a variety of tasks.

https://arxiv.org/abs/2306.02561

Blender has been shown to outperform individual LLMs on a variety of tasks, including text #summarization, question answering, and code generation. It is also more robust to adversarial attacks than individual LLMs.

The Blender paper makes the following contributions:

  • It proposes a novel ensembling framework for LLMs that can consistently outperform individual LLMs on a variety of tasks.
  • It introduces a meta-model that learns to identify the strengths and weaknesses of LLMs and selects the most appropriate LLMs to contribute to the final output.
  • It demonstrates the effectiveness of Blender on a variety of tasks, including text summarization, question answering, and code generation.
  • It shows that Blender is more robust to adversarial attacks than individual LLMs.

Here are some additional details about the two training methods:

  • PairRanker: PairRanker is a supervised learning method. This means that it is trained on a dataset of labeled data, where each data point consists of a pair of outputs from different LLMs and a label indicating which output is better. The meta-model learns to predict the label for each data point, and it is trained using a loss function that minimizes the difference between the predicted label and the actual label.
  • GenFuser: GenFuser is an unsupervised learning method. This means that it is trained on a dataset of unlabeled data, where each data point consists of the output of a single LLM. GenFuser learns to combine the outputs of different LLMs by using a technique called “attention.” Attention allows GenFuser to focus on the most relevant parts of the different outputs and to combine them in a way that produces a single, improved output.

The Blender paper is a significant contribution to the field of LLMs. It provides a new way to ensemble LLMs that can consistently improve their performance on a variety of tasks. The paper also introduces a novel meta-model that can identify the strengths and weaknesses of LLMs and select the most appropriate LLMs to contribute to the final output.

--

--

Mahesh Patel
Mahesh Patel

No responses yet