Meta Introduces Self-Taught Evaluator to Train LLMs

www.analyticsdrift.com

Image source: Analytics Drift

On 20th August 2024, Meta announced a new method of evaluating large language model (LLM) performance, which reduces manual effort.

Image source: Meta

Introduction to the new LLM evaluation technology

The current method to train accurate LLM evaluators relies on human-annotated data. This requires time, money, and specialized training, often creating a bottleneck in rapid development.

Image source: Meta

Challenges of using the traditional method

The Self-Taught Evaluators eliminate the requirement of human-labeled data by working on the principle of LLM-as-a-judge, which generates a reasoning chain for accurate responses.

Image source: Canva

Addressing LLM evaluation challenges

Self-taught evaluators use a seed LLM and unlabeled instructions to generate training data. It iteratively fine-tunes performance by adding examples with a correct reasoning chain to the training data.

Image source: Meta

Self-Taught Evaluator methodology in assessing LLM accuracy

The Self-Taught Evaluator surpassed some models trained on human-labeled data, enhancing the accuracy on benchmarks like MT-Bench and RewardBench.

Image source: Canva

Judging Self-Taught Evaluator’s performance

The Self-Taught Evaluator heavily relies on the initial seed model. It becomes necessary for you to choose seed and base models that are relevant to your data and specific requirements.

Image source: Canva

Efficient Utilization of Self-Taught Evaluator