Share via


Microsoft.Extensions.AI.Evaluation.NLP Namespace

Classes

BLEUEvaluator

An IEvaluator that evaluates the quality of a response produced by an AI model by comparing it to a reference response using the BLEU (Bilingual Evaluation Understudy) algorithm. It is often used to evaluate the quality of machine translation or text generation tasks.

BLEUEvaluatorContext

Contextual information that the BLEUEvaluator uses to compute the BLEU score for a response.

F1Evaluator

An IEvaluator that evaluates the quality of a response produced by an AI model by comparing it to a reference response using the F1 scoring algorithm. F1 score is the ratio of the number of shared words between the generated response and the reference response.

F1EvaluatorContext

Contextual information that the F1Evaluator uses to compute the F1 score for a response.

GLEUEvaluator

An IEvaluator that evaluates the quality of a response produced by an AI model by comparing it to a reference response using the GLEU (Google-BLEU) algorithm. The GLEU evaluator measures the similarity between the generated response and one or more reference responses using n-gram overlap.

GLEUEvaluatorContext

Contextual information that the GLEUEvaluator uses to compute the GLEU score for a response.