SOTA metrics for evaluating Retrieval Augmented Generation (RAG)
ragas is a framework that helps you evaluate your Retrieval Augmented Generation (RAG) pipelines. RAG denotes a class of LLM applications that use external data to augment the LLM’s context. There are existing tools and frameworks that help you build these pipelines but evaluating it and quantifying your pipeline performance can be hard.. This is were ragas (RAG Assessment) comes in
ragas provides you with the tools based on the latest research for evaluating LLM generated text to give you insights about your RAG pipeline. ragas can be integrated with your CI/CD to provide continuous check to ensure performance.
pip install ragas
if you want to install from source
git clone https://github.com/explodinggradients/ragas && cd ragas
pip install -e .
This is a small example program you can run to see ragas in action!
from datasets import load_dataset
from ragas.metrics import (
Evaluation,
rouge1,
bert_score,
entailment_score,
) # import the metrics you want to use
# load the dataset
ds = load_dataset("explodinggradients/eli5-test", split="test_eli5").select(range(100))
# init the evaluator, this takes in the metrics you want to use
# and performs the evaluation
e = Evaluation(
metrics=[rouge1, bert_score, entailment_score,],
batched=False,
batch_size=30,
)
# run the evaluation
results = e.eval(ds["ground_truth"], ds["generated_text"])
print(results)
If you want a more in-depth explanation of core components, check out our quick-start notebook
- Levenshtein distance the number of single character edits (additional, insertion, deletion) required to change your generated text to ground truth text.
- Levenshtein ratio is obtained by dividing the Levenshtein distance by sum of number of characters in generated text and ground truth. This type of metrics is suitable where one works with short and precise texts.
N-gram based metrics as name indicates uses n-grams for comparing generated answer with ground truth. It is suitable to extractive and abstractive tasks but has its limitations in long free form answers due to the word based comparison.
-
ROGUE (Recall-Oriented Understudy for Gisting Evaluation):
- ROUGE-N measures the number of matching ‘n-grams’ between generated text and ground truth. These matches do not consider the ordering of words.
- ROUGE-L measures the longest common subsequence (LCS) between generated text and ground truth. This means is that we count the longest sequence of tokens that is shared between both
-
BLEU (BiLingual Evaluation Understudy)
It measures precision by comparing clipped n-grams in generated text to ground truth text. These matches do not consider the ordering of words.
Model based methods uses language models combined with NLP techniques to compare generated text with ground truth. It is well suited for free form long or short answer types.
-
BertScore
Bert Score measures the similarity between ground truth text answers and generated text using SBERT vector embeddings. The common choice of similarity measure is cosine similarity for which values range between 0 to 1. It shows good correlation with human judgement.
-
EntailmentScore
Textual entailment to measure factual consistency in generated text given ground truth. Score can range from 0 to 1 with latter indicating perfect factual entailment for all samples. Entailment score is highly correlated with human judgement.
-
$Q^2$ Best used to measure factual consistencies between ground truth and generated text. Scores can range from 0 to 1. Higher score indicates better factual consistency between ground truth and generated answer. Employs QA-QG paradigm followed by NLI to compare ground truth and generated answer.
$Q^2$ score is highly correlated with human judgement.⚠️ time and resource hungry metrics.
📜 Checkout citations for related publications.