Understanding LLM Evaluation Metrics is crucial for maximizing the potential of large language models. LLM evaluation Metrics help measure a model’s accuracy, relevance, and overall effectiveness using various benchmarks and criteria. By systematically evaluating these models, developers can identify strengths, address weaknesses, and refine them for real-world applications. This process ensures that LLMs meet high standards of performance, fairness, and user satisfaction while continuously improving their capabilities.
In the field of AI development, the significance of LLM evaluation cannot be emphasised enough. Large language models (LLMs) must be evaluated to make sure they are accurate, dependable, and meet user expectations. This improves user satisfaction and confidence.
Key Benefits of LLM Evaluation
Quality Assurance: Regular evaluations ensure that LLMs maintain high standards of output quality, which is crucial for applications where accuracy is paramount.
User-Centric Development: By incorporating user feedback into the evaluation process, developers can create models that better meet the needs and preferences of their target audience.
Benchmarking Progress: Evaluation metrics allow teams to track improvements over time, providing a clear picture of how model updates and training efforts translate into enhanced performance.
Risk Mitigation: Evaluating LLMs helps identify potential biases or ethical concerns in model outputs, enabling organizations to address these issues proactively and reduce the risk of negative consequences.
Below we will look into the LLM evaluation metrics division:
Accuracy Metrics: Measure the correctness of the model’s outputs against a set of ground truth answers, often using precision, recall, and F1 scores.
Lexical Similarity: Assesses how closely the generated text matches reference texts, typically using metrics like BLEU or ROUGE to evaluate word overlap.
Relevance and Informativeness: Evaluates whether the model’s responses are pertinent to the query and provide valuable information, often assessed through human judgment or relevance scores.
Bias and Fairness: Analyzes the model’s outputs for potential biases and ensures equitable treatment across different demographics, focusing on ethical implications.
Efficiency: Measures the computational resources required for the model to generate outputs, including response time and resource consumption.
LLM Based: Refers to metrics specifically designed for evaluating large language models, considering their unique characteristics and capabilities in generating human-like text.
Understanding Accuracy Metrics
Below we will look into the accuracy metrics in detail:
1. Perplexity
Perplexity is an important metric used to evaluate language models. It essentially measures how well a model predicts the next word in a sentence or sequence. In simpler terms, perplexity tells us how “surprised” or “uncertain” the model is when it encounters new text.
When a model is confident about predicting the next word, the perplexity will be low. Conversely, if the model is unsure or predicts many different possible next words, the perplexity will be high.
How Perplexity is Calculated?
To calculate perplexity, we look at the likelihood of the model generating the correct sequence of words. The formula is:
Where:
represents the probability of the iii-th word given the previous words in the sentence.
N is the total number of words in the sequence.
The model computes the log probabilities of each word, averages them, negates the result, and then exponentiates it to get the perplexity.
Example to Understand Perplexity
Let’s make it clearer with an example. Imagine the sentence “I am learning about perplexity.” Suppose the model assigns the following probabilities:
To find the perplexity, you would:
Calculate the log of each probability:
Sum these log probabilities.
Average the log probabilities by dividing by the number of words in the sentence.
Finally, apply the exponentiation to get the perplexity.
What Does Perplexity Tell Us?
The main takeaway is that lower perplexity is better. A low perplexity means the model is confident and accurate in predicting the next word. On the other hand, a high perplexity suggests that the model is uncertain or “guessing” more when predicting the next word.
For example, if the model predicts the next word with high certainty, it will have a low perplexity score. If it’s not sure about the next word and considers many options, the perplexity will be higher.
Why Perplexity is Important?
Perplexity is valuable because it provides a simple, interpretable measure of how well a language model is performing. The lower the perplexity, the better the model is at predicting the next word in a sequence. However, while perplexity is useful, it’s not the only metric to assess a model. It’s often combined with other metrics, like accuracy or human evaluations, to get a fuller picture of a model’s performance.
Limitations of Perplexity
Next-word prediction, not comprehension: Perplexity measures how well a model predicts the next word, not its understanding of meaning or context. Low perplexity doesn’t guarantee meaningful or coherent text.
Vocabulary and tokenization dependence: Perplexity is influenced by vocabulary size and tokenization methods, making comparisons across different models and settings difficult.
Bias towards frequent words: Perplexity can be lowered by accurately predicting common words, even if the model struggles with less frequent but semantically important terms.
2. Cross Entropy Loss
Cross entropy loss is a way to quantify how far the predicted probability distribution is from the actual distribution. It is used in classification tasks, including language modeling, where the model predicts a probability distribution over the next word or token in a sequence.
Mathematically, cross entropy loss for a single prediction is defined as:
Where:
p(xi) is the true probability distribution of the i-th word (often represented as one-hot encoding for classification tasks),
q(xi) is the predicted probability distribution of the i-th word,
The summation is over all possible words iii in the vocabulary.
For a language model, this equation can be applied over all words in a sequence to calculate the total loss.
How Cross Entropy Loss Works?
Let’s break this down:
True Distribution: This represents the actual word (or token) that occurred in the data. For example, if the actual word in a sentence is “dog”, the true distribution will have a probability of 1 for “dog” and 0 for all other words (in one-hot encoding).
Predicted Distribution: This is the probability distribution predicted by the model for each word in the vocabulary. For example, the model might predict that there’s a 60% chance the next word is “dog”, 30% chance it’s “cat”, and 10% for other words.
Logarithm: The log function helps turn multiplication into addition, and it also emphasizes small probabilities. This way, if the model assigns a high probability to the correct word, the loss is low. If the model assigns a low probability to the correct word, the loss will be higher.
Example of Cross Entropy Loss
Imagine a simple vocabulary with only three words: [“dog”, “cat”, “fish”]. Suppose the actual next word in a sentence is “dog”. The true probability distribution for “dog” will look like this:
Now, let’s say the model predicts the following probabilities for the next word:
The cross entropy loss can be calculated as:
Substitute the values:
Since the terms for “cat” and “fish” are multiplied by 0, they vanish, so:
Using a calculator:
So, the cross entropy loss in this case is approximately 0.2218. This loss would be smaller if the model predicted “dog” with higher confidence (a higher probability), and larger if it predicted a word that was far from the correct one.
Why is Cross Entropy Loss Important?
Cross entropy loss is critical because it directly penalizes the model when its predictions deviate from the true values. It’s commonly used in training models for classification tasks, including language models, because:
It gives a clear measure of how far off the model is from the correct predictions.
It encourages the model to improve its probability estimates by adjusting the weights during training, helping the model get better over time.
It’s mathematically convenient for optimization, especially when using gradient-based methods like stochastic gradient descent (SGD).
In language models, cross entropy loss is used to train the model by minimizing the difference between the predicted word probabilities and the actual words. This helps the model generate more accurate predictions over time.
Limitations of Cross Entropy Loss
Word-level prediction, not understanding: Cross-entropy loss optimizes for accurate next-word prediction, not genuine language understanding. Minimizing loss doesn’t guarantee the model grasps meaning or context.
Data distribution dependence: Cross-entropy is sensitive to the training data. Biased or noisy data can lead to models that perform well on training data but poorly generalize.
Frequent word bias: Cross-entropy can be dominated by frequent word predictions, potentially masking poor performance on less common but crucial vocabulary.
Understanding Lexical Similarity Metrics
Now we will look into the understanding of Lexical similarity metrics in detail below:
3. BLEU
The BLEU score is a widely used metric for evaluating the quality of text generated by machine translation models. It’s a way to measure how closely the machine-generated translation matches human translations. Despite being designed for machine translation, BLEU can also be applied to other natural language processing (NLP) tasks where the goal is to generate sequences of text, such as text summarization or caption generation.
BLEU stands for Bilingual Evaluation Understudy and is primarily used to evaluate machine-generated translations by comparing them to one or more reference translations created by humans. The BLEU score ranges from 0 to 1, where a higher score indicates that the machine-generated text is closer to human-produced text in terms of n-gram (word sequence) matching.
N-grams are consecutive sequences of words. For example, for the sentence “The cat is on the mat”, the 2-grams (or bigrams) would be: [“The cat”, “cat is”, “is on”, “on the”, “the mat”].
How BLEU Score is Calculated?
BLEU evaluates the precision of n-grams in the generated text compared to reference translations. It uses the following steps:
Precision Calculation: BLEU computes the precision for different n-grams, such as unigrams (1-grams), bigrams (2-grams), trigrams (3-grams), etc. The precision for an n-gram is defined as the ratio of the number of n-grams that appear in both the generated text and the reference text to the total number of n-grams in the generated text.
Brevity Penalty: BLEU also includes a brevity penalty to avoid favoring shorter translations. If the machine-generated translation is shorter than the reference translation, the brevity penalty reduces the BLEU score. The brevity penalty is calculated as follows:
Final BLEU Score: Finally, the BLEU score is calculated by combining the precision for multiple n-grams (usually from 1 to 4) and applying the brevity penalty:
Example of BLEU Calculation
Let’s walk through a simple example to understand how BLEU works.
Reference Sentence: “The cat is on the mat.”
Generated Sentence: “A cat is on the mat.”
Unigram Precision: We first calculate the unigram (1-gram) precision. Here, the unigrams in the reference are [“The”, “cat”, “is”, “on”, “the”, “mat”], and in the generated sentence, they are [“A”, “cat”, “is”, “on”, “the”, “mat”]. Common unigrams between the reference and generated sentence are: [“cat”, “is”, “on”, “the”, “mat”]. So, the unigram precision is:
Bigram Precision: Next, we calculate the bigram (2-gram) precision. The bigrams in the reference sentence are: [“The cat”, “cat is”, “is on”, “on the”, “the mat”], and in the generated sentence, they are: [“A cat”, “cat is”, “is on”, “on the”, “the mat”]. Common bigrams between the reference and generated sentence are: [“cat is”, “is on”, “on the”, “the mat”]. So, the bigram precision is:
Brevity Penalty: Since the generated sentence is shorter than the reference sentence, we apply the brevity penalty. Assuming the length of the reference is 6 and the length of the generated sentence is 5, the brevity penalty would be:
Final BLEU Score: Now, we combine the unigram and bigram precision and apply the brevity penalty: After calculating the logs and the exponentiation, we get the final BLEU score.
Why is BLEU Important?
BLEU is important because it provides an automated, reproducible way to evaluate machine-generated text. It offers several advantages:
Consistency: It gives a consistent metric across different systems and datasets.
Efficiency: BLEU allows for quick, automated evaluation, which is useful during model development or hyperparameter tuning.
Comparison: BLEU helps compare different translation models or other sequence generation models, as it’s based on a clear, quantitative evaluation.
Limitations of BLEU
N-gram overlap, not semantics: BLEU solely measures overlapping n-grams between generated and reference text, ignoring meaning. High BLEU doesn’t guarantee semantic similarity or correct information.
Exact word matching, penalizes paraphrasing: BLEU’s reliance on exact word matches penalizes valid paraphrasing and synonymous substitutions, even if meaning is preserved.
Insensitive to word order within n-grams: While n-grams capture some local word order, BLEU doesn’t fully account for it. Rearranging words within an n-gram can impact the score even if meaning is largely maintained.
4. ROUGE
ROUGE is a set of metrics used to evaluate automatic text generation tasks, such as summarization and machine translation. Unlike BLEU, which is precision-based, ROUGE focuses on recall by comparing the overlap of n-grams (sequences of words) between the generated text and a set of reference texts. The goal is to assess how much information from the reference text is captured in the generated output.
ROUGE is widely used to evaluate models in tasks like text summarization, abstractive summarization, and image captioning, among others.
Types of ROUGE Metrics
ROUGE includes multiple variants, each focusing on different types of evaluation. The most common ROUGE metrics are:
ROUGE-N: This measures the overlap of n-grams (i.e., unigrams, bigrams, trigrams, etc.) between the generated and reference texts.
ROUGE-1 is the unigram (1-gram) overlap.
ROUGE-2 is the bigram (2-gram) overlap.
ROUGE-L: This calculates the longest common subsequence (LCS) between the generated and reference texts. It measures the longest sequence of words that appear in both the generated and reference texts in the same order.
ROUGE-S: This measures the overlap of skip-bigrams, which are pairs of words in the same order but not necessarily adjacent to each other.
ROUGE-W: This is a weighted version of ROUGE-L, which gives different weights to the different lengths of the common subsequences.
ROUGE-SU: This combines ROUGE-S and ROUGE-1 to also consider the unigrams in the skip-bigrams.
ROUGE-Lsum: This variant measures the longest common subsequence in a sentence-summary combination, often used in document summarization tasks.
How ROUGE is Calculated?
The basic calculation of ROUGE involves comparing recall for n-grams (how much of the reference n-grams are captured in the generated n-grams). Here’s how you can think of the core calculations:
Additionally, there are variations that also calculate precision and F1 score, which combine recall and precision to provide a balance between how much the generated text matches and how much of it is relevant.
Precision: Measures the percentage of n-grams in the generated text that match those in the reference.
F1 Score: This is the harmonic mean of precision and recall and is often used to provide a balanced evaluation metric.
Example of ROUGE Calculation
Let’s break down how ROUGE would work in a simple example.
Reference Text: “The quick brown fox jumps over the lazy dog.”
Generated Text: “A fast brown fox jumps over the lazy dog.”
ROUGE-1 (Unigram) Precision
We first find the unigrams in both the reference and the generated text:
There are 6 matching bigrams, and there are 8 bigrams in the reference and 8 in the generated text.
Why ROUGE is Important?
ROUGE is particularly valuable for tasks like automatic text summarization, where we need to ensure that the generated summary captures key information from the original document. It is highly popular because it evaluates recall, which is crucial in tasks where missing important content would hurt the result.
Key reasons why ROUGE is important:
Recall-Based: ROUGE prioritizes recall, ensuring that the model generates content that matches reference content as closely as possible.
Evaluates Meaning: ROUGE is designed to evaluate how much information the generated text contains in comparison to the reference, making it useful for summarization tasks.
Widely Used: Many NLP research papers use ROUGE as the go-to metric, making it a standard for evaluating summarization systems.
Limitations of ROUGE
Despite its popularity, ROUGE has its drawbacks:
Doesn’t Account for Paraphrasing: ROUGE doesn’t capture semantic meaning as well as human evaluation. Two sentences may have the same meaning but use different words or sentence structures, which ROUGE may penalize.
Ignores Fluency: ROUGE focuses on n-gram overlap but doesn’t account for grammatical correctness or fluency of the generated text.
5. METEOR
It stands for Metric for Evaluation of Translation with Explicit Ordering, and it was introduced to improve the limitations of previous evaluation methods, particularly for machine translation tasks. METEOR considers multiple factors beyond just n-gram precision:
Exact word matching: The system’s translation is compared with reference translations, where exact word matches increase the score.
Synonym matching: Synonyms are counted as matches, making METEOR more flexible in evaluating translations that convey the same meaning but use different words.
Stemming: The metric accounts for variations in word forms by reducing words to their root forms (e.g., “running” to “run”).
Word order: METEOR penalizes word order mismatches, since the order of words is often important in translation.
Paraphrasing: METEOR is designed to handle paraphrasing, where different words or structures are used to express the same idea.
How METEOR is Calculated?
METEOR is calculated using a combination of precision, recall, and a number of penalties for mismatches in word order, stemming, and synonymy. Here’s a general breakdown of how METEOR is calculated:
Exact word matches: METEOR calculates how many exact word matches there are between the generated and reference text. The more matches, the higher the score.
Synonym matches: METEOR allows for synonyms (i.e., words with similar meanings) to be counted as matches. For example, “good” and “excellent” could be treated as a match.
Stemming: Words are reduced to their root form. For example, “playing” and “played” would be treated as the same word after stemming.
Precision and Recall: METEOR calculates the precision and recall of the matches:
Precision: The proportion of matched words in the generated text to the total number of words in the generated text.
Recall: The proportion of matched words in the generated text to the total number of words in the reference text.
The F1 score is then calculated as the harmonic mean of precision and recall.
Penalty for word order: To account for the importance of word order, METEOR applies a penalty to translations that have a large deviation from the reference word order. This penalty reduces the score for translations with major word order mismatches.
Final METEOR Score: The final METEOR score is a weighted combination of the precision, recall, synonym matching, stemming, and word order penalties. The formula is: The Penalty term depends on the number of word order mismatches and the length of the generated sentence, and it ranges from 0 to 1.
Example of METEOR Calculation
Let’s walk through an example of how METEOR would work in a simple scenario:
Reference Translation: “The cat is on the mat.”
Generated Translation: “A cat sits on the mat.”
Step 1: Exact Word Matches
The words that match exactly between the reference and the generated text are:
“cat”, “on”, “the”, “mat”.
There are 4 exact word matches.
Step 2: Synonym Matching
The word “sits” in the generated sentence can be considered a synonym for “is” in the reference sentence.
So, “sits” and “is” are treated as a match.
Step 3: Stemming
Both “sits” and “is” would be reduced to their root forms during stemming. The root form of “sits” is “sit”, which is similar to “is” (as they both represent the same action in this context). However, in practice, METEOR would treat these as synonyms (this is an approximation).
Step 4: Calculate Precision and Recall
Precision: The total number of word matches (including synonyms) divided by the total number of words in the generated translation.
Recall: The total number of word matches divided by the total number of words in the reference translation.
Step 5: Calculate F1 Score
The F1 score is the harmonic mean of precision and recall:
Step 6: Apply Penalty
In this example, the word order between the reference and generated translations is slightly different. However, the penalty for word order is typically small if the differences are minimal, so the final penalty might be 0.1.
Step 7: Final METEOR Score
Finally, the METEOR score is calculated by applying the penalty:
Thus, the METEOR score for this translation would be 0.72.
Why METEOR is Important?
METEOR is a more flexible evaluation metric than BLEU because it takes several important linguistic aspects into account, such as:
Synonym matching: This helps to recognize that different words with the same meaning should be treated as equivalent.
Word order: METEOR penalizes significant differences in word order, which is crucial in tasks like machine translation.
Stemming: By reducing words to their base form, METEOR reduces the impact of morphological differences.
These features make METEOR a better choice for evaluating machine translations, especially when considering natural language that may have more variation than a strict n-gram matching approach.
Limitations of METEOR
While METEOR is more flexible than BLEU, it still has some limitations:
Complexity: METEOR is more complex to compute than BLEU because it involves stemming, synonym matching, and calculating word order penalties.
Performance on Short Texts: METEOR can sometimes give higher scores to short translations that match a lot of content in a small number of words, potentially overestimating the quality of a translation.
Subjectivity of Synonym Matching: Deciding what words are synonyms can sometimes be subjective and context-dependent, making METEOR’s evaluation a bit inconsistent in some cases.
Understanding Relevance and Informativeness Metrics
We will now explore relevance and informativeness metrics:
6. BERTScore
BERTScore is based on the idea that the quality of text generation should not only depend on exact word matches but also on the semantic meaning conveyed by the generated text. It uses the powerful pre-trained BERT model, which encodes words in a contextual manner—i.e., it captures the meaning of words in context rather than in isolation.
How BERTScore Works?
Embedding Generation: First, BERTScore generates contextual embeddings for each token (word or subword) in both the generated and reference texts using the pre-trained BERT model. These embeddings capture the meaning of words in the context of the sentence.
Cosine Similarity: For each token in the generated text, BERTScore calculates the cosine similarity with the tokens in the reference text. Cosine similarity measures how similar two vectors (embeddings) are. The closer the cosine similarity value is to 1, the more semantically similar the tokens are.
Precision, Recall, and F1 Score: BERTScore computes three core values—precision, recall, and F1 score—based on the cosine similarity values:
Precision: Measures how much of the generated text aligns with the reference text in terms of semantic similarity. It calculates the average cosine similarity of each generated token to the most similar token in the reference.
Recall: Measures how much of the reference text is captured in the generated text. It calculates the average cosine similarity of each reference token to the most similar token in the generated text.
F1 Score: This is the harmonic mean of precision and recall, providing a balanced score between the two.
The basic BERTScore formula for precision and recall is:
Where:
Finally, the F1 Score is calculated as:
Example of BERTScore Calculation
Let’s walk through a simple example:
Reference Text: “The quick brown fox jumped over the lazy dog.”
Generated Text: “A fast brown fox leapt over the lazy dog.”
Generate Embeddings: Both the reference and generated sentences are passed through BERT, and contextual embeddings for each word are extracted.
Calculate Cosine Similarities: For each token in the generated sentence, calculate the cosine similarity to the tokens in the reference sentence:
For example, the token “fast” in the generated sentence will be compared to the tokens “quick” and “brown” in the reference sentence. The cosine similarity between “fast” and “quick” may be high, as they are semantically similar.
Compute Precision and Recall: After calculating the similarities, compute the precision and recall for the generated text based on how well the tokens align with the reference.
Compute F1 Score: Finally, calculate the F1 score as the harmonic mean of precision and recall.
For this example, BERTScore would likely assign high similarity to words like “brown”, “fox”, “lazy”, and “dog”, and would penalize the difference between “quick” and “fast” as well as “jumped” and “leapt”. The generated sentence may still be considered high quality due to semantic equivalence, even though there are some lexical differences.
Why BERTScore is Important?
BERTScore has several advantages, particularly in evaluating the semantic relevance and informativeness of the generated text:
Contextual Understanding: Since BERT generates contextual embeddings, it can understand word meanings in context, which helps in capturing semantic similarity even if the exact words are different.
Handles Synonyms: Unlike traditional n-gram-based metrics, BERTScore recognizes synonyms and paraphrases, which is critical in tasks like machine translation or text generation, where different wordings can express the same idea.
Handles Word Order: BERTScore accounts for word order to some extent, especially when measuring the overall semantic meaning of the sentence. This is more accurate than simple word overlap measures.
More Informative: BERTScore focuses on both relevance (precision) and informativeness (recall), which makes it better suited for tasks where both factors matter, such as summarization or translation.
Limitations of BERTScore
While BERTScore is a powerful metric, it also has some limitations:
Computationally Expensive: Since BERTScore uses the BERT model to generate embeddings, it can be computationally expensive, especially when dealing with large datasets or long sentences.
Dependence on Pre-trained Models: BERTScore relies on the pre-trained BERT model. The quality of BERTScore can be influenced by how well the pre-trained model generalizes to the specific task or domain, and it may not always perform optimally for tasks that differ significantly from the data BERT was trained on.
Interpretability: While BERTScore is more advanced than traditional metrics, it may be harder to interpret because it does not give explicit insight into which words or phrases in the generated text are responsible for high or low scores.
Lack of Sentence Fluency Evaluation: BERTScore evaluates semantic similarity but doesn’t account for fluency or grammatical correctness. A sentence could have a high BERTScore but still sound awkward or ungrammatical.
7. MoverScore
MoverScore leverages word embeddings to calculate how far apart two sets of words (the reference and the generated texts) are in terms of semantic meaning. The core idea is that, instead of merely counting the overlap between words (as in BLEU or ROUGE), MoverScore looks at the distance between the words in a continuous semantic space.
It’s inspired by earth mover’s distance (EMD), a measure of the minimal cost to move a set of distributions to match another set. In the case of MoverScore, the “distribution” is the set of word embeddings for the words in the sentences, and the “cost” is the semantic distance between words in the embeddings.
How MoverScore Works?
Word Embeddings: First, both the reference and generated sentences are converted into word embeddings using pre-trained models like Word2Vec, GloVe, or BERT. These embeddings represent words as vectors in a high-dimensional space, where semantically similar words are positioned closer to each other.
Matching Words: Next, MoverScore calculates the semantic distance between each word in the generated text and the words in the reference text. The basic idea is to measure how far words in the generated text are from the words in the reference text, in terms of their embeddings.
Earth Mover’s Distance (EMD): The Earth Mover’s Distance is used to calculate the minimal cost of transforming the set of word embeddings in the generated sentence into the set of word embeddings in the reference sentence. EMD provides a measure of the “effort” required to move the words in one sentence to match the words in the other sentence, based on their semantic meaning.
MoverScore Calculation: The MoverScore is calculated by computing the EMD between the word embeddings of the generated sentence and the reference sentence. The lower the cost of “moving” the embeddings from the generated text to the reference text, the better the generated text is considered to match the reference text semantically. The formula for MoverScore is typically expressed as: Here, EMD is the earth mover’s distance between the generated and reference sentence embeddings, and the denominator is the maximum possible EMD, which serves as a normalization factor.
Example of MoverScore Calculation
Let’s consider a simple example to demonstrate how MoverScore works:
Reference Sentence: “The cat sat on the mat.”
Generated Sentence: “A cat is resting on the carpet.”
Generate Word Embeddings: Both the reference and generated sentences are passed through a pre-trained model to obtain word embeddings. The words “cat” and “resting”, for example, would have embeddings that represent their meanings in the context of the sentence.
Calculate Semantic Distance: Next, the semantic distance between the words in the generated sentence and the reference sentence is computed. For instance, the word “resting” in the generated sentence might have a close embedding to “sat” in the reference sentence because both describe similar actions (the cat is in a resting position as opposed to sitting).
Calculate Earth Mover’s Distance (EMD): The EMD is then calculated to measure the minimal “cost” required to match the embeddings from the generated sentence to the embeddings in the reference sentence. If “cat” and “cat” are the same word, there is no cost to move them, but the distance for other words like “mat” vs. “carpet” will be non-zero.
Final MoverScore: Finally, the MoverScore is calculated by normalizing the EMD with respect to the maximum possible distance and inverting it. A lower EMD means a higher MoverScore, indicating the generated sentence is semantically closer to the reference sentence.
Why MoverScore is Important?
MoverScore provides several advantages over traditional metrics like BLEU, ROUGE, and METEOR:
Semantic Focus: MoverScore focuses on the meaning of the words, not just their exact matches. It evaluates the semantic similarity between the generated and reference texts, which is crucial for tasks where the wording may differ, but the meaning remains the same.
Context-Aware: By using word embeddings (such as those from BERT or Word2Vec), MoverScore is context-aware. This means it can recognize that two different words may have similar meanings in a given context, and it captures that similarity.
Handles Paraphrasing: MoverScore is particularly useful in tasks where paraphrasing is common (e.g., summarization, translation). It doesn’t penalize minor word changes that still convey the same meaning, unlike BLEU or ROUGE, which may fail to account for such differences.
Limitations of MoverScore
While MoverScore is a powerful metric, it also has some limitations:
Computational Complexity: MoverScore requires computing the earth mover’s distance, which can be computationally expensive, especially for long sentences or large datasets.
Dependency on Word Embeddings: The quality of MoverScore depends on the quality of the word embeddings used. If the embeddings are not trained on relevant data or fail to capture nuances in a specific domain, the MoverScore may not accurately reflect the quality of the generated text.
Not Language-Agnostic: Since MoverScore relies on word embeddings, it is generally not language-agnostic. The embeddings used must be specific to the language of the text being evaluated, which may limit its applicability in multilingual settings.
Lack of Fluency or Grammar Assessment: MoverScore evaluates semantic similarity but does not consider fluency or grammatical correctness. A sentence that is semantically similar to the reference might still be ungrammatical or awkward.
8. Undertsanding Bias Score
Bias Score is a metric used to measure the degree of bias in natural language processing (NLP) models, particularly in text generation tasks. It aims to assess whether a model produces output that disproportionately favors certain groups, attributes, or perspectives while disadvantaged others. Bias in AI models, especially in large language models (LLMs), has gained significant attention due to its potential to perpetuate harmful stereotypes or reinforce societal inequalities.
In general, the higher the Bias Score, the more biased the model’s outputs are considered to be. Bias can manifest in various forms, including:
Stereotyping: Associating certain characteristics (e.g., professions, behaviors, or roles) with specific genders, races, or other groups.
Exclusion: Ignoring or marginalizing certain groups or perspectives.
Disproportionate Representation: Presenting certain groups in a more favorable or negative light than others.
How Bias Score Works?
The process of calculating the Bias Score involves several steps, which may vary depending on the exact implementation. However, most approaches follow a general framework that involves identifying sensitive attributes and evaluating the extent to which the model’s output exhibits bias towards those attributes.
Identify Sensitive Attributes: The first step in calculating Bias Score is identifying which sensitive attributes or groups are of concern. This may include gender, ethnicity, religion, or other demographic characteristics.
Model Output Analysis: The model’s output, whether text, predictions, or generated content, is analyzed for biased language or associations related to sensitive attributes. For example, when the model generates text or completes sentences based on specific prompts, the output is examined for gendered or racial biases.
Bias Detection: The next step involves detecting potential bias in the output. This could include checking for stereotypical associations (e.g., “nurse” being associated predominantly with females or “engineer” with males). The model’s outputs are analyzed for disproportionate representation or negative stereotyping of certain groups.
Bias Score Calculation: Once bias has been detected, the Bias Score is calculated by comparing the degree of bias in the model’s output against a reference or baseline. This could involve comparing the frequency of biased terms in the output to the expected distribution of those terms. The score might be normalized or scaled to produce a value that reflects the extent of bias, typically on a scale from 0 to 1, where 0 indicates no bias and 1 indicates extreme bias.
Example of Bias Score Calculation
Let’s go through an example:
Sensitive Attribute: Gender (Male and Female)
Generated Sentence: “The scientist is a man who conducts experiments.”
Identify Sensitive Attributes: The sensitive attribute in this example is gender, as we are concerned with whether the profession “scientist” is associated with a male gender.
Bias Detection: In the generated sentence, the term “man” is associated with the role of “scientist.” This could be viewed as biased because it reinforces a stereotype that scientists are primarily male.
Bias Score Calculation: The Bias Score is calculated by measuring how often the model associates the word “man” with the “scientist” role. This is then compared to a balanced baseline where “scientist” is equally linked to both male and female terms.The formula could look something like: If the model predominantly associates “scientist” with male pronouns or references (e.g., “man”), the Bias Score would be higher, indicating a higher degree of gender bias.
Why Bias Score is Important
Detecting Harmful Bias: Bias Score helps identify whether an NLP model is reinforcing harmful stereotypes or social biases. Detecting such biases is important to ensure that the generated text does not inadvertently harm certain groups or perpetuate societal inequalities.
Improving Fairness: By measuring the Bias Score, developers can identify areas where a model needs improvement in terms of fairness. This metric can guide the modification of training data or model architecture to reduce bias and improve the overall ethical standards of AI systems.
Accountability: As AI systems are increasingly deployed in real-world applications, including hiring, law enforcement, and healthcare, ensuring fairness and accountability is critical. Bias Score helps organizations assess whether their models produce outputs that are fair and unbiased, helping to prevent discriminatory outcomes.
Limitations of Bias Score
Context Sensitivity: Bias Score calculations can sometimes be context-sensitive, meaning that a model’s output might be biased in one scenario but not in another. For example, some terms might be biased in a general sense but not in a particular context, making it difficult to provide a definitive Bias Score across all situations.
Data Dependence: The Bias Score depends heavily on the data used for evaluation. If the reference dataset used to determine bias is flawed or unbalanced, it could lead to inaccurate measurements of bias.
Quantitative Measure: While Bias Score is a quantitative metric, bias itself is a complex and multifaceted concept. The metric might not capture all the nuances of bias in a model’s output, such as subtle cultural biases or implicit biases that are not easily identified in a simple analysis.
False Positives/Negatives: Depending on how the Bias Score is calculated, there could be false positives (labeling neutral outputs as biased) or false negatives (failing to identify bias in certain outputs). Ensuring that the metric captures genuine bias without overfitting is an ongoing challenge.
9. Understanding Fairness Score
Fairness Score measures how a model treats different groups or individuals. It ensures no group is unfairly favored. This metric is crucial for AI and machine learning models. Biased decisions in these systems can have serious consequences. They can impact hiring, lending, criminal justice, and healthcare.
The Fairness Score is used to measure the degree of fairness in a model’s predictions or outputs, which can be defined in various ways depending on the specific task and context. It aims to quantify how much the model’s performance varies across different demographic groups, such as gender, race, age, or socioeconomic status.
Types of Fairness Metrics
Before understanding the Fairness Score, it is essential to note that fairness in machine learning can be measured in different ways. The Fairness Score can be calculated using various fairness metrics depending on the chosen definition of fairness. Some of the commonly used fairness metrics are:
Demographic Parity (Group Fairness): This metric checks whether the model’s predictions are equally distributed across different groups. For example, in a hiring model, demographic parity would ensure that candidates from different gender or racial groups are selected at equal rates.
Equalized Odds (Individual Fairness): Equalized odds ensures that the model’s performance (e.g., true positive rate and false positive rate) is the same across different groups. This metric ensures that the model does not make different types of errors for different demographic groups.
Equality of Opportunity: This is a variation of equalized odds, where the focus is solely on ensuring equal true positive rates for different groups. It is especially relevant in cases where the model’s decision to classify individuals as positive or negative has critical real-world consequences, such as in the criminal justice system.
Conditional Use Accuracy Equality: This metric measures whether the model has the same accuracy within each group defined by the sensitive attribute. It aims to ensure that the model’s accuracy does not disproportionately favor one group over another.
Individual Fairness: This approach checks whether similar individuals receive similar predictions. The model should treat similar individuals equally, regardless of sensitive attributes like gender or race.
How Fairness Score Works?
The calculation of the Fairness Score depends on the fairness metric being used. Here’s a general approach:
Identify Sensitive Attributes: Sensitive attributes (e.g., gender, race, age) must first be identified. These are the attributes you want to evaluate for fairness.
Evaluate Model Performance Across Groups: The model’s performance is then analyzed for each subgroup defined by these sensitive attributes. For example, if gender is a sensitive attribute, you would compare the model’s performance for male and female groups separately.
Compute the Fairness Score: The Fairness Score is typically calculated by measuring the disparity in performance metrics (e.g., accuracy, false positive rate, or true positive rate) between different groups. The greater the disparity, the lower the Fairness Score. For example, if a model performs well for one group but poorly for another group, the Fairness Score would be low, signaling a bias or unfairness. Conversely, if the model performs equally well for all groups, the Fairness Score will be high, indicating fairness.
Where:
GGG is the set of all groups defined by sensitive attributes (e.g., male, female, white, Black).
Performance of group g is the model’s performance metric (e.g., accuracy, precision) for group ggg.
Average Performance is the overall performance metric across all groups.
The Fairness Score ranges from 0 (indicating extreme unfairness) to 1 (indicating perfect fairness).
Example of Fairness Score Calculation
Let’s consider a binary classification model for hiring that uses gender as a sensitive attribute. Suppose the model is evaluated on two groups: males and females.
Male Group:
Accuracy: 85%
True Positive Rate: 90%
False Positive Rate: 5%
Female Group:
Accuracy: 75%
True Positive Rate: 70%
False Positive Rate: 10%
Now, to calculate the Fairness Score, we can evaluate the disparity in performance between the two groups. Let’s say we are interested in accuracy as the performance metric.
Calculate the disparity in accuracy:
Male Group Accuracy: 85%
Female Group Accuracy: 75%
Disparity = 85% – 75% = 10%
Calculate the Fairness Score:
In this case, the Fairness Score is 0.9, indicating a relatively high degree of fairness. However, a score closer to 1 would signify better fairness, and a score closer to 0 would indicate a high level of unfairness or bias.
Why Fairness Score is Important?
Ethical AI Development: The Fairness Score helps ensure that AI models are not causing harm to vulnerable or underrepresented groups. By quantifying fairness, developers can ensure that AI systems operate equitably, adhering to ethical standards.
Regulatory Compliance: In many industries, such as finance, healthcare, and hiring, fairness is a legal requirement. For example, algorithms used in hiring should not discriminate based on gender, race, or other protected characteristics. The Fairness Score can help ensure that models comply with these regulations.
Reducing Harm: A model with a low Fairness Score may be causing disproportionate harm to certain groups. By identifying and addressing biases early on, developers can mitigate the negative impact of AI systems.
Limitations of Fairness Score
Trade-offs Between Fairness and Accuracy: In some cases, achieving fairness can come at the expense of accuracy. For example, improving fairness for one group may result in a drop in overall performance. This trade-off needs to be carefully managed.
Context Dependence: Fairness is not always a one-size-fits-all concept. What is considered fair in one context might not be considered fair in another. The definition of fairness can vary depending on societal norms, the specific application, and the groups being evaluated.
Complexity of Sensitive Attributes: Sensitive attributes such as race or gender are not always clear-cut. There are many ways in which these attributes can manifest or be perceived, and these complexities may not always be captured by a single Fairness Score.
Bias in Fairness Metrics: Ironically, fairness metrics themselves can be biased depending on how they are designed or how data is collected. Ensuring that the fairness metrics are fair and unbiased is an ongoing challenge.
10. Understanding Toxicity Detection
Toxicity Detection is a metric used to evaluate the harmfulness of text generated by language models, especially when applied in natural language processing (NLP) tasks. It focuses on identifying whether the output produced by an AI system contains inappropriate, offensive, or harmful content. The goal of toxicity detection is to ensure that language models generate content that is safe, respectful, and non-harmful.
Toxicity detection has become an essential aspect of evaluating language models, particularly in scenarios where AI models are used to generate content in open-ended contexts, such as social media posts, chatbots, content moderation systems, or customer service applications. Since AI-generated content can inadvertently or intentionally promote hate speech, offensive language, or harmful behavior, toxicity detection is vital to reduce the negative impact of such models.
Types of Toxicity
Toxicity can manifest in several ways, and understanding the various types of toxicity is crucial for evaluating the performance of toxicity detection systems. Some common types of toxicity include:
Hate Speech: Text that expresses hatred or promotes violence against a person or group based on attributes like race, religion, ethnicity, sexual orientation, or gender.
Abuse: Verbal attacks, threats, or any other form of abusive language directed at individuals or groups.
Harassment: Repeated, targeted behavior meant to disturb, intimidate, or degrade others, including cyberbullying.
Offensive Language: Mildly offensive words or phrases that are generally socially unacceptable, such as curse words or slurs.
Discrimination: Language that shows prejudice against or unfair treatment of people based on certain characteristics like gender, race, or age.
How Toxicity Detection Works?
Toxicity detection typically relies on machine learning models that are trained to recognize harmful language in text. These models analyze the output and score it based on how likely it is to contain toxic content. The general approach involves:
Data Annotation: Toxicity detection models are trained on datasets containing text that is labeled as either toxic or non-toxic. These datasets include examples of harmful and non-harmful language, often manually labeled by human annotators. The training data helps the model learn patterns of toxic language, including slang, offensive terms, and harmful sentiment.
Feature Extraction: The model extracts various features from the text, such as word choice, sentence structure, sentiment, and context, to identify potentially toxic content. These features may include:
Explicit Words: Offensive or abusive terms like slurs or profanity.
Sentiment: Detecting whether the overall sentiment of the text is hostile or degrading.
Context: Toxicity can depend on the context, so the model often considers the surrounding words to evaluate intent and level of harm.
Classification: The model classifies the text as either toxic or non-toxic. Typically, the classification task involves assigning a binary label (toxic or not) or a continuous toxicity score to the text. The score reflects how likely it is that the text contains harmful language.
Thresholding: Once the model generates a toxicity score, a threshold is set to determine whether the content is toxic enough to require intervention. For instance, if the toxicity score exceeds a predefined threshold, the model may flag the output for review or moderation.
Post-processing: In many cases, additional filtering or moderation steps are used to automatically filter out the most harmful content based on toxicity scores. These systems may be integrated into platforms for automated content moderation.
Example of Toxicity Detection in Practice
Let’s take an example where a language model generates the following text:
Generated Text 1: “I can’t believe how stupid this person is!”
Generated Text 2: “You’re such an idiot, and you’ll never succeed!”
Now, toxicity detection systems would analyze these two sentences for harmful language:
Sentence 1: The word “stupid” might be considered mildly offensive, but it does not contain hate speech or abuse. The toxicity score could be low.
Sentence 2: The word “idiot” and the overall tone of the sentence indicate verbal abuse and offensive language. This sentence would likely receive a higher toxicity score.
A toxicity detection system would evaluate both sentences and assign a higher score to the second one, signaling that it’s more harmful than the first. Depending on the threshold set, the second sentence might be flagged for review or discarded.
Toxicity Score Calculation
The Toxicity Score is usually calculated based on the model’s output for a given piece of text. This score can be represented as a probability or a continuous value between 0 and 1, where:
A score close to 0 indicates that the content is non-toxic or safe.
A score close to 1 indicates high levels of toxicity.
For example, if a model is trained on a large dataset containing toxic and non-toxic sentences, the model can be tasked with predicting the probability that a new sentence is toxic. This can be represented as:
If the model predicts a probability of 0.8 for a given sentence, it means that the sentence has an 80% chance of being toxic.
Why Toxicity Detection is Important?
Preventing Harmful Content: Language models that generate text for social media platforms, customer support, or chatbots must be evaluated for toxicity to prevent the spread of harmful content, including hate speech, harassment, and abusive language.
Maintaining Community Standards: Toxicity detection helps platforms enforce their community guidelines by automatically filtering out inappropriate or offensive content, promoting a safe online environment for users.
Ethical Responsibility: Language models must be responsible in how they interact with people. Toxicity detection is crucial for ensuring that models do not perpetuate harmful stereotypes, encourage violence, or violate ethical standards.
Legal Compliance: In some industries, there are legal requirements regarding the content that AI models generate. For example, chatbots used in customer service or healthcare must avoid producing offensive or harmful language to comply with regulations.
Limitations of Toxicity Detection
Context Sensitivity: Toxicity can be highly context-dependent. A word or phrase that is offensive in one context may be acceptable in another. For example, “idiot” might be considered offensive when directed at a person, but it could be used humorously in certain situations.
False Positives and Negatives: Toxicity detection models can sometimes flag non-toxic content as toxic (false positives) or fail to detect toxic content (false negatives). Ensuring the accuracy of these models is challenging, as toxicity can be subtle and context-specific.
Cultural Differences: Toxicity may vary across cultures and regions. What is considered offensive in one culture may be acceptable in another. Models need to be sensitive to these cultural differences, which can be difficult to account for in training data.
Evolution of Language: Language and societal norms change over time. Words that were once considered acceptable may become offensive, or vice versa. Toxicity detection systems need to adapt to these evolving linguistic trends to remain effective.
Understanding Efficiency Metric
After exploring about so many metrics now it is time to learn about efficiency metrics in detail below:
11. Latency
Latency is a critical efficiency metric in the evaluation of large language models (LLMs), referring to the amount of time it takes for a model to generate a response after receiving an input. In simpler terms, latency measures how quickly a system can process data and return an output. For language models, this would be the time taken from when a user inputs a query to when the model produces the text response.
In applications like real-time chatbots, virtual assistants, or interactive systems, low latency is essential to provide smooth and responsive user experiences. High latency, on the other hand, can result in delays, causing frustration for users and diminishing the effectiveness of the system.
Key Factors Affecting Latency
Several factors can influence the latency of an LLM:
Model Size: Larger models (e.g., GPT-3, GPT-4) require more computational resources, which can increase the time needed to process input and generate a response. Larger models typically have higher latency due to the complexity of their architecture and the number of parameters they contain.
Hardware: The hardware on which the model is running can significantly affect latency. Running a model on a high-performance GPU or TPU will generally result in lower latency compared to using a CPU. Additionally, cloud-based systems may have more overhead due to network latency.
Batch Processing: If multiple requests are processed simultaneously in batches, it may reduce the overall time for each individual request, improving latency. However, this is highly dependent on the server infrastructure and the model’s ability to handle concurrent requests.
Optimization Techniques: Techniques such as model pruning, quantization, and knowledge distillation can reduce the size of the model without significantly sacrificing performance, leading to reduced latency. Also, approaches like mixed-precision arithmetic and model caching can help speed up inference.
Input Length: The length of the input text can affect latency. Longer inputs require more time for the model to process, as the model has to consider more tokens and context to generate an appropriate response.
Network Latency: When LLMs are hosted on cloud servers, network latency (the delay in data transmission over the internet) can also play a role in overall latency. A slow internet connection or server congestion can add delay to the time it takes for data to travel back and forth.
Measuring Latency
Latency is typically measured as the inference time, which is the time taken for a model to process an input and generate an output. There are several ways to measure latency:
End-to-End Latency: The time taken from when the user submits the input to when the response is displayed, including all preprocessing and network delays.
Model Inference Latency: This is the time taken specifically by the model to process the input and generate a response. It excludes any preprocessing or postprocessing steps.
Average Latency: The average latency across multiple inputs or requests is often calculated to provide a more general view of system performance.
Percentiles of Latency: Often, the 99th percentile or 95th percentile latency is measured to understand the performance of the system under stress or heavy load. This tells you how fast 99% or 95% of responses are generated, excluding outliers that might skew the average. Where the 99th percentile means that 99% of the requests have lower latency than this value.
Why Latency is Important in LLM Evaluation?
User Experience: For real-time applications like chatbots, virtual assistants, and interactive AI systems, latency directly impacts user experience. Users expect responses in milliseconds or seconds, and delays can cause frustration or reduce the usability of the system.
Real-Time Applications: Many LLMs are used in environments where real-time responses are critical. Examples include live customer support, automated content moderation, and voice assistants. High latency can undermine the utility of these systems and cause users to disengage.
Scalability: In production environments, latency can affect the scalability of a system. If the model has high latency, it may struggle to handle a large number of requests simultaneously, leading to bottlenecks, slowdowns, and potential system crashes.
Throughput vs. Latency Trade-Off: Latency is often balanced with throughput, which refers to the number of requests a system can handle in a given period. High throughput typically means lower latency, but this is not always the case, especially in systems that cannot handle a large number of requests simultaneously. Optimizing for one may come at the cost of the other.
Optimizing Latency in LLMs
To optimize latency while maintaining performance, there are several techniques that can be used:
Model Pruning: This technique involves removing unnecessary neurons or weights from a trained model, reducing its size and improving inference speed without sacrificing too much accuracy.
Quantization: By reducing the precision of the weights in a model (e.g., using 16-bit floating-point numbers instead of 32-bit), it is possible to reduce the computational cost and increase the inference speed.
Distillation: Knowledge distillation involves transferring the knowledge from a large, complex model to a smaller, more efficient model. The smaller model retains much of the performance of the larger one but is faster and less resource-intensive.
Caching: For models that generate responses based on similar queries, caching previous responses can help reduce latency for repeated queries.
Batching: Processing multiple requests at once (batching) can help reduce latency by allowing the system to utilize hardware resources more efficiently, especially in environments with high request volumes.
Edge Computing: Moving models closer to the user by deploying them on edge devices or local servers can reduce latency associated with network transmission times.
Example of Latency Impact
Consider two language models with different latencies in a chatbot application:
Model A (Low Latency): Responds in 100 ms.
Model B (High Latency): Responds in 2 seconds.
For users interacting with these chatbots in a real-time conversation, the response time of Model A will provide a smoother, more engaging experience. In contrast, Model B would create noticeable delays, causing potential frustration for the user.
If these models were deployed in a customer service application, Model B‘s high latency could result in lower customer satisfaction and increased wait times. Model A, with its faster response time, would likely lead to higher customer retention and a more positive experience.
12. Computational Efficiency
Computational efficiency can be measured in various ways, depending on the specific aspect of resource usage being considered. In general, it refers to how efficiently a model can produce the desired output using the least amount of computational resources. For LLMs, the most common resources involved are:
Memory Usage: The amount of memory required to store model parameters, intermediate results, and other necessary data during inference.
Processing Power (Compute): The number of calculations or floating-point operations (FLOPs) required to process an input and generate an output.
Energy Consumption: The amount of energy consumed by the model during training and inference, which can be a major factor in large-scale deployments.
Key Aspects of Computational Efficiency
Model Size: Larger models, like GPT-3, contain billions of parameters, which require significant computational power to operate. Reducing the size of a model while maintaining performance is one way to improve its computational efficiency. Smaller models or more efficient architectures are typically faster and consume less power.
Training and Inference Speed: The time it takes for a model to complete tasks such as training or generating text is an important measure of computational efficiency. Faster models can process more requests within a given time frame, which is essential for applications requiring real-time or near-real-time responses.
Memory Usage: Efficient use of memory is crucial, especially for large models. Reducing memory consumption helps prevent bottlenecks during model training or inference, enabling deployment on devices with limited memory resources.
Energy Efficiency: Energy consumption is an important aspect of computational efficiency, particularly in cloud computing environments where resources are shared. Optimizing models for energy efficiency reduces costs and the environmental impact of AI systems.
Measuring Computational Efficiency
Several metrics are used to evaluate computational efficiency in LLMs:
FLOPs (Floating Point Operations): This measures the number of operations required by a model to process an input. The fewer FLOPs a model uses, the more computationally efficient it is. For example, a model with fewer FLOPs may run faster and consume less power. FLOPs=Operations per second
Parameter Efficiency: This refers to how effectively the model uses its parameters. Efficient models maximize performance with a smaller number of parameters, which directly impacts their computational efficiency. Model Size=Number of Parameters
Smaller, optimized models require less memory and processing power, making them more efficient.
Latency: This measures the amount of time the model takes to produce a response after receiving an input. Lower latency translates to higher computational efficiency, especially in real-time applications. Latency=Time taken to process and generate output
Throughput: Throughput refers to the number of tasks or predictions the model can handle in a specific amount of time. Higher throughput means the model is more efficient at processing multiple inputs in parallel, which is important in large-scale deployments.
Why Computational Efficiency is Important?
Cost Reduction: Computational resources, such as GPUs or cloud services, can be expensive, especially when dealing with large-scale models. Optimizing computational efficiency reduces the cost of running models, which is essential for commercial applications.
Scalability: As demand for LLMs increases, computational efficiency ensures that models can scale effectively without requiring disproportionately high computational resources. This is critical for cloud-based services or applications that need to handle millions of users.
Energy Consumption: The energy usage of AI models, particularly large ones, can be significant. By improving computational efficiency, it is possible to reduce the environmental impact of running these models, making them more sustainable.
Real-Time Applications: Low-latency and high-throughput performance are especially important for applications like chatbots, virtual assistants, or real-time translation, where delays or interruptions can harm user experience. Efficient models can meet the demanding needs of these applications.
Model Deployment: Many real-world applications of LLMs, such as on mobile devices or edge computing platforms, have strict computational constraints. Computationally efficient models can be deployed in such environments without requiring excessive computational resources.
Optimizing Computational Efficiency
Several techniques can be employed to optimize the computational efficiency of LLMs:
Model Compression: This involves reducing the size of a model without significantly affecting its performance. Techniques like quantization, pruning, and knowledge distillation can make models smaller and faster.
Distributed Computing: Using multiple machines or GPUs to handle different parts of the model or different tasks can improve computational efficiency by distributing the load. This is particularly useful in training large models.
Efficient Model Architectures: Research into new model architectures, such as transformers with fewer parameters or sparsely activated models, can lead to more efficient models that require less computational power.
Parallel Processing: Leveraging parallel processing techniques, where tasks are broken down into smaller parts and processed simultaneously, can speed up inference times and reduce overall computational costs.
Hardware Acceleration: Using specialized hardware like GPUs, TPUs, or FPGAs can greatly improve the efficiency of training and inference, as these devices are optimized for parallel processing and large-scale computations.
Fine-Tuning: Rather than training a large model from scratch, fine-tuning pre-trained models on specific tasks can reduce the computational cost and improve efficiency, as the model already has learned general patterns from large datasets.
Example of Computational Efficiency
Consider two versions of a language model:
Model A: A large model with 175 billion parameters, taking 10 seconds to generate a response and consuming 50 watts of power.
Model B: A smaller, optimized version with 30 billion parameters, taking 3 seconds to generate a response and consuming 20 watts of power.
In this case, Model B would be considered more computationally efficient because it generates output faster and consumes less power, even though it still performs well for most tasks.
Understanding LLM Based Metrics
Below we will understand LLM based metrics:
13. LLM as a Judge
LLM as a Judge is the process where large language models are used to assess the quality of outputs generated by another instance of an AI system, typically in the context of natural language processing (NLP) tasks. Rather than relying solely on traditional metrics (like BLEU, ROUGE, etc.), an LLM can be asked to evaluate whether the generated output adheres to predefined rules, structures, or even ethical standards.
For example, an LLM might be tasked with evaluating whether a machine-generated essay is logically coherent, contains biased language, or adheres to specific guidelines (such as word count, tone, or style). LLMs can also be used to assess whether the content reflects factual accuracy or to predict the potential impact or reception of a certain piece of content.
How LLM as a Judge Works?
The process of using LLMs as a judge generally follows these steps:
Task Definition: First, the specific task or evaluation criterion must be defined. This could involve assessing fluency, coherence, relevance, creativity, factual accuracy, or adherence to certain stylistic or ethical guidelines.
Model Prompting: Once the task is defined, the LLM is prompted with the content to evaluate. This could involve providing the model with a piece of text (e.g., a machine-generated article) and asking it to rate or provide feedback based on the criteria outlined earlier.
Model Assessment: The LLM then processes the input and produces an evaluation. Depending on the task, the evaluation might include a score, an analysis, or a recommendation. For example, in a task focused on fluency, the LLM might provide a numerical score representing how fluent and coherent the text is.
Comparison to Ground Truth: The generated assessment is often compared to a baseline or a human evaluation (when available). This helps ensure that the LLM’s judgments align with human expectations and are consistent across different tasks.
Feedback and Iteration: Based on the LLM’s output, adjustments can be made to improve the generated content or the evaluation criteria. This iterative feedback loop helps refine both the generation process and the judging mechanism.
Key Benefits of Using LLM as a Judge
Scalability: One of the primary advantages of using LLMs as judges is their scalability. LLMs can quickly evaluate vast amounts of content, making them ideal for tasks like content moderation, plagiarism detection, or automatic grading of assignments.
Consistency: Human evaluators may have subjective biases or vary in their judgments based on mood, context, or other factors. LLMs, however, can offer consistent evaluations, making them useful for maintaining uniformity across large datasets or tasks.
Efficiency: Using an LLM as a judge is far more time-efficient than manual evaluations, especially when dealing with large volumes of data. This can be particularly helpful in contexts such as content creation, marketing, and customer feedback analysis.
Automation: LLMs can help automate the evaluation of machine-generated content, allowing systems to self-improve and adapt over time. This is useful for fine-tuning models in a variety of tasks, from natural language understanding to generating more human-like text.
Real-Time Evaluation: LLMs can assess content in real-time, providing immediate feedback during the creation or generation of new content. This is valuable in dynamic environments, such as chatbots, customer service, or real-time content moderation.
Common Tasks Where LLMs Act as Judges
Content Quality Evaluation: LLMs can be used to assess the quality of generated text in terms of fluency, coherence, and relevance. For instance, after a model generates a piece of text, an LLM can be tasked with evaluating whether the text flows logically, maintains a consistent tone, and adheres to the guidelines set for the task.
Bias and Fairness Detection: LLMs can be used to identify bias in generated text. This includes detecting gender, racial, or cultural bias that may exist in the content, helping to ensure that AI-generated outputs are neutral and equitable.
Fact-Checking and Accuracy: LLMs can assess whether the generated content is factually accurate. Given their large knowledge base, these models can be asked to evaluate whether specific claims in the text hold true against known facts or data.
Grading and Scoring: In education, LLMs can act as grading systems for assignments, essays, or exams. They can evaluate content based on predefined rubrics, providing feedback on structure, argumentation, and clarity.
Example of LLM as a Judge in Action
Imagine that you have a model that generates product descriptions for an e-commerce site. After generating a product description, you could use an LLM as a judge to assess the quality of the text based on the following criteria:
Relevance: Does the description accurately reflect the product features?
Fluency: Is the text grammatically correct and readable?
Bias Detection: Is the text free from discriminatory language or stereotyping?
Length: Does the description meet the required word count?
The LLM could be prompted to rate the description on a scale of 0 to 10 for each criterion. Based on this feedback, the generated content could be refined or improved.
Why LLM as a Judge is Important?
Enhanced Automation: By automating the evaluation process, LLMs can make large-scale content generation more efficient and accurate. This can reduce human involvement and speed up the content creation process, particularly in industries like marketing, social media, and customer service.
Improved Content Quality: With LLMs acting as judges, organizations can ensure that generated content aligns with the desired tone, style, and quality standards. This is especially critical in customer-facing applications where high-quality content is necessary to maintain a positive brand image.
Bias Mitigation: By incorporating LLMs as judges, companies can identify and eliminate biases from AI-generated content, leading to more ethical and fair outputs. This helps prevent discrimination and promotes inclusivity.
Scalability and Cost-Effectiveness: Using LLMs to judge large amounts of content provides a cost-effective way to scale operations. It reduces the need for manual evaluation and can help businesses meet the growing demand for automated services.
Limitations of LLM as a Judge
Bias in the Judge: While LLMs can be helpful in judging content, they are not immune to the biases present in their training data. If the LLM has been trained on biased datasets, it might inadvertently reinforce harmful stereotypes or unfair evaluations.
Lack of Subjectivity: While LLMs can provide consistency in evaluations, they may lack the nuanced understanding that a human evaluator might have. For instance, LLMs may miss subtle context or cultural references that are important for evaluating content appropriately.
Dependence on Training Data: The accuracy of LLMs as judges is limited by the quality of the data used for their training. If the training data does not cover a wide range of contexts or languages, the LLM’s evaluation might not be accurate or comprehensive.
14. RTS
RTS (Reason Then Score) is a metric used in the evaluation of language models and AI systems, particularly in the context of tasks involving reasoning and decision-making. It emphasizes a two-step process where the model first provides a rationale or reasoning behind its output and then assigns a score or judgment based on that reasoning. The idea is to separate the reasoning process from the scoring process, allowing for more transparent and interpretable AI evaluations.
RTS involves two distinct steps in the evaluation process:
Reasoning: The model is required to explain or justify the reasoning behind its output. This is typically done by generating a set of logical steps, supporting evidence, or explanations that lead to the final answer.
Scoring: Once the reasoning is provided, the model assigns a score to the quality of the response or decision, typically based on the correctness of the reasoning and its alignment with a predefined standard or evaluation criteria.
This two-step approach aims to improve the interpretability and accountability of AI systems, allowing humans to better understand how a model reached a particular conclusion.
How RTS Works?
RTS generally follows these steps:
Task Definition: A specific reasoning task is defined. This could be answering a complex question, making a decision based on a set of criteria, or performing a logic-based operation. The task often involves both understanding context and applying reasoning to generate an output.
Model Reasoning: The model is prompted to explain the reasoning process it used to arrive at a particular conclusion. For example, in a question-answering task, the model might first break down the question and then explain how each part of the question contributes to the final answer.
Model Scoring: After the reasoning process is outlined, the model then evaluates how well it did in answering the question or solving the problem. This scoring could involve providing a numerical rating or assessing the overall correctness, coherence, or relevance of the generated reasoning and final answer.
Comparison to Ground Truth: The final score or evaluation is often compared to human judgments or reference answers. The purpose is to validate the quality of the reasoning and the accuracy of the final output, ensuring that the AI’s decision-making process is aligned with expert standards.
Feedback and Iteration: Based on the score and feedback from human evaluators or comparison to ground truth, the model can be iteratively improved. This feedback loop helps refine both the reasoning and scoring aspects of the AI system.
Key Benefits of RTS (Reason Then Score)
Improved Transparency: RTS helps increase the transparency of AI systems by requiring the model to provide explicit reasoning. This makes it easier for humans to understand why a model arrived at a certain conclusion, helping to build trust in AI outputs.
Accountability: By breaking down the reasoning process and then scoring the output, RTS holds the model accountable for its decisions. This is crucial for high-stakes applications like healthcare, law, and autonomous systems, where understanding the “why” behind a decision is just as important as the decision itself.
Enhanced Interpretability: In complex tasks, RTS allows for a more interpretable approach. For instance, if a model is used to answer a legal question, RTS ensures that the model’s reasoning can be followed step by step, making it easier for a human expert to assess the soundness of the model’s conclusion.
Better Evaluation of Reasoning Skills: By separating reasoning from scoring, RTS provides a more accurate evaluation of a model’s reasoning capabilities. It ensures that the model is not just outputting a correct answer, but is also able to explain how it arrived at that answer.
Common Tasks Where RTS is Used
Complex Question Answering: In question answering tasks, especially those that require multi-step reasoning or the synthesis of information from various sources, RTS can be used to ensure that the model not only provides the correct answer but also explains how it arrived at that answer.
Legal and Ethical Decision Making: RTS can be used in scenarios where AI models are required to make legal or ethical decisions. The model provides its reasoning behind a legal interpretation or an ethical judgment, which is then scored based on correctness and adherence to legal standards or ethical principles.
Logical Reasoning Tasks: In tasks such as puzzles, mathematical reasoning, or logic problems, RTS can help evaluate how well a model applies logic to derive solutions, ensuring that the model not only provides an answer but also outlines the steps it took to arrive at that solution.
Summarization: In text summarization tasks, RTS can be used to evaluate whether the model has effectively summarized the key points of a document and provided a clear reasoning for why it selected certain points over others.
Dialogue Systems: In conversational AI, RTS can be used to evaluate how well a model reasons through a conversation and provides coherent, logically structured responses that align with the user’s needs.
Example of RTS (Reason Then Score) in Action
Consider a scenario where an AI system is tasked with answering a complex question such as:
Question: “What is the impact of climate change on agricultural production?”
Reasoning Step: The model might first break down the question into sub-components such as “climate change,” “agricultural production,” and “impact.” Then, it would explain how climate change affects weather patterns, soil quality, water availability, etc., and how these changes influence crop yields, farming practices, and food security.
Scoring Step: After providing this reasoning, the model would evaluate its answer based on its accuracy, coherence, and relevance. It might assign a score based on how well it covered key aspects of the question and how logically it connected its reasoning to the final conclusion.
Final Score: The final score could be a numerical value (e.g., 0 to 10) reflecting how well the model’s reasoning and answer align with expert knowledge.
Why RTS (Reason Then Score) is Important?
Improves AI Accountability: RTS ensures that AI systems are held accountable for the way they make decisions. By requiring reasoning to be separate from scoring, it provides a clear audit trail of how conclusions are drawn, which is critical for applications like legal analysis and policy-making.
Fosters Trust: Users are more likely to trust AI systems if they can understand how decisions are made. RTS provides transparency into the decision-making process, which can help build trust in the model’s outputs.
Encourages More Thoughtful AI Design: When models are forced to provide reasoning before scoring, it encourages developers to design systems that are capable of deep, logical reasoning and not just surface-level pattern recognition.
Limitations of RTS (Reason Then Score)
Complexity: The two-step nature of RTS can make it more difficult to implement compared to simpler evaluation metrics. Generating reasoning requires more sophisticated models and additional training, which may add complexity to the development process.
Dependence on Context: Reasoning-based tasks often depend heavily on context. A model’s ability to reason well in one domain (e.g., legal text) may not translate to another domain (e.g., medical diagnosis), which can limit the general applicability of RTS.
Potential for Misleading Reasoning: If the model’s reasoning is flawed or biased, the final score may still be high, despite the reasoning being inaccurate. Therefore, it’s important to ensure that the reasoning step is as accurate and unbiased as possible.
15. G-Eval
G-Eval, or Generative Evaluation, is a flexible evaluation metric for generative AI systems that helps assess the overall effectiveness and quality of the generated content. It is often used in tasks like text generation, dialogue systems, summarization, and creative content production. G-Eval aims to provide a more holistic view of how a model performs in terms of both its outputs and its overall behavior during the generation process.
Key elements that G-Eval takes into account include:
Relevance: Whether the generated content is pertinent to the given input, question, or prompt.
Creativity: How original or creative the content is, especially in tasks such as storytelling, poetry, or brainstorming.
Coherence: Whether the generated content maintains a logical flow and makes sense in the context of the input.
Diversity: The ability of the model to generate varied and non-repetitive outputs, especially important for tasks requiring creativity.
Fluency: The grammatical and syntactic quality of the generated content.
Human-likeness: How closely the content resembles human-generated text in terms of style, tone, and structure.
How G-Eval Works?
G-Eval typically involves the following process:
Content Generation: The AI model generates content based on a given input or prompt. This could include text generation, dialogue, creative writing, etc.
Human Evaluation: Human evaluators assess the quality of the generated content based on predefined criteria such as relevance, creativity, coherence, and fluency. This is often done on a scale (e.g., 1 to 5) to rate each of these factors.
Automated Evaluation: Some implementations of G-Eval combine human feedback with automated metrics like perplexity, BLEU, ROUGE, or other traditional evaluation scores to provide a more comprehensive view of the model’s performance.
Comparison to Baselines: The generated content is compared to a baseline or reference content, which could be human-generated text or outputs from another model. This helps determine whether the AI-generated content meets certain standards or expectations.
Iterative Feedback: Based on the evaluation, feedback is provided to refine and improve the generative model. This can be done through fine-tuning, adjusting the model’s hyperparameters, or re-training it with more diverse or specific datasets.
Key Benefits of G-Eval
Holistic Evaluation: Unlike traditional metrics, G-Eval considers multiple dimensions of content quality, allowing for a broader and more nuanced evaluation of generative models.
Alignment with Human Expectations: G-Eval focuses on how well the generated content aligns with human expectations in terms of creativity, relevance, and coherence. This makes it an important tool for applications where human-like quality is essential.
Encourages Creativity: By including creativity as an evaluation criterion, G-Eval helps to push generative models towards more innovative and original outputs, which is valuable in tasks such as storytelling, creative writing, and marketing.
Improved Usability: For real-world applications, it is important to generate content that is not only accurate but also useful and engaging. G-Eval ensures that AI-generated outputs meet practical needs in terms of human relevance, fluency, and coherence.
Adaptability: G-Eval can be applied to various generative tasks, whether for dialogue generation, text summarization, translation, or even creative tasks like music or poetry generation. It is a versatile metric that can be tailored to different use cases.
Common Use Cases for G-Eval
Text Generation: In natural language generation (NLG) tasks, G-Eval is used to assess how well a model generates text that is fluent, relevant, and coherent with the given input or prompt.
Dialogue Systems: For chatbots and conversational AI, G-Eval helps evaluate how natural and relevant the responses are in a dialogue context. It can also assess the creativity and diversity of responses, ensuring that conversations do not become repetitive or monotonous.
Summarization: In automatic summarization tasks, G-Eval can evaluate whether the generated summaries are coherent, concise, and adequately reflect the main points of the original content.
Creative Writing: G-Eval is particularly valuable in evaluating AI models used for creative tasks like storytelling, poetry generation, and scriptwriting. It assesses not only the fluency and coherence of the text but also its originality and creativity.
Content Generation for Marketing: In marketing, G-Eval can help assess AI-generated advertisements, social media posts, or promotional content for creativity, relevance, and engagement.
Example of G-Eval in Action
Let’s say you are using a generative model to write a creative short story based on the prompt: “A group of astronauts discovers an alien species on a distant planet.”
Content Generation: The model generates a short story about the astronauts encountering a peaceful alien civilization, filled with dialogues and vivid descriptions.
Human Evaluation: Human evaluators rate the story on several aspects:
Relevance: Does the story stay on topic and follow the prompt? (e.g., 4/5)
Creativity: How original and creative is the plot and the alien species? (e.g., 5/5)
Coherence: Does the story flow logically from start to finish? (e.g., 4/5)
Fluency: Is the text well-written and grammatically correct? (e.g., 5/5)
Automated Evaluation: The model’s generated text is also evaluated using automated metrics like perplexity to measure fluency and BLEU for any comparisons to a reference text, if available.
Final G-Eval Score: The combined score, considering both human and automated evaluations, gives an overall quality rating of the model’s performance in this task.
Why G-Eval is Important?
Better Model Performance: By providing a more comprehensive evaluation framework, G-Eval encourages the development of more capable generative models that not only generate accurate but also creative, relevant, and coherent content.
Real-World Applications: In many real-world scenarios, especially in fields like marketing, entertainment, and customer service, the quality of AI-generated content is judged not just by accuracy but also by how engaging and useful it is. G-Eval addresses this need by evaluating models on these practical aspects.
Improved Human-AI Interaction: As AI models are increasingly integrated into systems that interact with humans, it is important that these systems produce outputs that are both useful and natural. G-Eval helps ensure that these systems generate content that is human-like and appropriate for various contexts.
Limitations of G-Eval
Subjectivity of Human Evaluation: While G-Eval aims to be holistic, the human evaluation aspect is still subjective. Different evaluators may have varying opinions on what constitutes creativity or relevance, which can introduce inconsistency in the results.
Difficulty in Defining Criteria: The criteria used in G-Eval, such as creativity or relevance, can be difficult to quantify and may require domain-specific definitions or guidelines to ensure consistent evaluation.
Resource Intensive: G-Eval often requires significant human involvement, which can be time-consuming and resource-intensive, especially when applied to large-scale generative tasks.
Conclusion
After reading this article, you now understand the significance of LLM Evaluation Metrics for large language models. You’ve learned about various assessment metrics that evaluate LLMs across tasks like language translation, question answering, text generation, and text summarization. A set of essential standards for evaluation has been presented to you. Additionally, you’ve explored best practices to conduct evaluations effectively. Since LLM Evaluation Metrics remain an active research area, new measurements and benchmarks will continue to emerge as the field evolves.
Harsh Mishra is an AI/ML Engineer who spends more time talking to Large Language Models than actual humans. Passionate about GenAI, NLP, and making machines smarter (so they don’t replace him just yet). When not optimizing models, he’s probably optimizing his coffee intake. 🚀☕
We use cookies essential for this site to function well. Please click to help us improve its usefulness with additional cookies. Learn about our use of cookies in our Privacy Policy & Cookies Policy.
Show details
Powered By
Cookies
This site uses cookies to ensure that you get the best experience possible. To learn more about how we use cookies, please refer to our Privacy Policy & Cookies Policy.
brahmaid
It is needed for personalizing the website.
csrftoken
This cookie is used to prevent Cross-site request forgery (often abbreviated as CSRF) attacks of the website
Identityid
Preserves the login/logout state of users across the whole site.
sessionid
Preserves users' states across page requests.
g_state
Google One-Tap login adds this g_state cookie to set the user status on how they interact with the One-Tap modal.
MUID
Used by Microsoft Clarity, to store and track visits across websites.
_clck
Used by Microsoft Clarity, Persists the Clarity User ID and preferences, unique to that site, on the browser. This ensures that behavior in subsequent visits to the same site will be attributed to the same user ID.
_clsk
Used by Microsoft Clarity, Connects multiple page views by a user into a single Clarity session recording.
SRM_I
Collects user data is specifically adapted to the user or device. The user can also be followed outside of the loaded website, creating a picture of the visitor's behavior.
SM
Use to measure the use of the website for internal analytics
CLID
The cookie is set by embedded Microsoft Clarity scripts. The purpose of this cookie is for heatmap and session recording.
SRM_B
Collected user data is specifically adapted to the user or device. The user can also be followed outside of the loaded website, creating a picture of the visitor's behavior.
_gid
This cookie is installed by Google Analytics. The cookie is used to store information of how visitors use a website and helps in creating an analytics report of how the website is doing. The data collected includes the number of visitors, the source where they have come from, and the pages visited in an anonymous form.
_ga_#
Used by Google Analytics, to store and count pageviews.
_gat_#
Used by Google Analytics to collect data on the number of times a user has visited the website as well as dates for the first and most recent visit.
collect
Used to send data to Google Analytics about the visitor's device and behavior. Tracks the visitor across devices and marketing channels.
AEC
cookies ensure that requests within a browsing session are made by the user, and not by other sites.
G_ENABLED_IDPS
use the cookie when customers want to make a referral from their gmail contacts; it helps auth the gmail account.
test_cookie
This cookie is set by DoubleClick (which is owned by Google) to determine if the website visitor's browser supports cookies.
_we_us
this is used to send push notification using webengage.
WebKlipperAuth
used by webenage to track auth of webenagage.
ln_or
Linkedin sets this cookie to registers statistical data on users' behavior on the website for internal analytics.
JSESSIONID
Use to maintain an anonymous user session by the server.
li_rm
Used as part of the LinkedIn Remember Me feature and is set when a user clicks Remember Me on the device to make it easier for him or her to sign in to that device.
AnalyticsSyncHistory
Used to store information about the time a sync with the lms_analytics cookie took place for users in the Designated Countries.
lms_analytics
Used to store information about the time a sync with the AnalyticsSyncHistory cookie took place for users in the Designated Countries.
liap
Cookie used for Sign-in with Linkedin and/or to allow for the Linkedin follow feature.
visit
allow for the Linkedin follow feature.
li_at
often used to identify you, including your name, interests, and previous activity.
s_plt
Tracks the time that the previous page took to load
lang
Used to remember a user's language setting to ensure LinkedIn.com displays in the language selected by the user in their settings
s_tp
Tracks percent of page viewed
AMCV_14215E3D5995C57C0A495C55%40AdobeOrg
Indicates the start of a session for Adobe Experience Cloud
s_pltp
Provides page name value (URL) for use by Adobe Analytics
s_tslv
Used to retain and fetch time since last visit in Adobe Analytics
li_theme
Remembers a user's display preference/theme setting
li_theme_set
Remembers which users have updated their display / theme preferences
We do not use cookies of this type.
_gcl_au
Used by Google Adsense, to store and track conversions.
SID
Save certain preferences, for example the number of search results per page or activation of the SafeSearch Filter. Adjusts the ads that appear in Google Search.
SAPISID
Save certain preferences, for example the number of search results per page or activation of the SafeSearch Filter. Adjusts the ads that appear in Google Search.
__Secure-#
Save certain preferences, for example the number of search results per page or activation of the SafeSearch Filter. Adjusts the ads that appear in Google Search.
APISID
Save certain preferences, for example the number of search results per page or activation of the SafeSearch Filter. Adjusts the ads that appear in Google Search.
SSID
Save certain preferences, for example the number of search results per page or activation of the SafeSearch Filter. Adjusts the ads that appear in Google Search.
HSID
Save certain preferences, for example the number of search results per page or activation of the SafeSearch Filter. Adjusts the ads that appear in Google Search.
DV
These cookies are used for the purpose of targeted advertising.
NID
These cookies are used for the purpose of targeted advertising.
1P_JAR
These cookies are used to gather website statistics, and track conversion rates.
OTZ
Aggregate analysis of website visitors
_fbp
This cookie is set by Facebook to deliver advertisements when they are on Facebook or a digital platform powered by Facebook advertising after visiting this website.
fr
Contains a unique browser and user ID, used for targeted advertising.
bscookie
Used by LinkedIn to track the use of embedded services.
lidc
Used by LinkedIn for tracking the use of embedded services.
bcookie
Used by LinkedIn to track the use of embedded services.
aam_uuid
Use these cookies to assign a unique ID when users visit a website.
UserMatchHistory
These cookies are set by LinkedIn for advertising purposes, including: tracking visitors so that more relevant ads can be presented, allowing users to use the 'Apply with LinkedIn' or the 'Sign-in with LinkedIn' functions, collecting information about how visitors use the site, etc.
li_sugr
Used to make a probabilistic match of a user's identity outside the Designated Countries
MR
Used to collect information for analytics purposes.
ANONCHK
Used to store session ID for a users session to ensure that clicks from adverts on the Bing search engine are verified for reporting purposes and for personalisation
We do not use cookies of this type.
Cookie declaration last updated on 24/03/2023 by Analytics Vidhya.
Cookies are small text files that can be used by websites to make a user's experience more efficient. The law states that we can store cookies on your device if they are strictly necessary for the operation of this site. For all other types of cookies, we need your permission. This site uses different types of cookies. Some cookies are placed by third-party services that appear on our pages. Learn more about who we are, how you can contact us, and how we process personal data in our Privacy Policy.