Stay Ahead, Stay ONMINE

LettuceDetect: A Hallucination Detection Framework for RAG Applications

Originally published on HuggingFace TL;DR We present LettuceDetect, a lightweight hallucination detector for Retrieval-Augmented Generation (RAG) pipelines. It is an encoder-based model built on ModernBERT, released under the MIT license with ready-to-use Python packages and pretrained models. What: LettuceDetect is a token-level detector that flags unsupported segments in LLM answers. đŸ„Ź How: Trained on RAGTruth (18k examples), leveraging ModernBERT for context lengths up to 4k tokens. 🚀 Why: It addresses (1) the context-window limits in prior encoder-only models, and (2) the high compute costs of LLM-based detectors. ⚖ Highlights: Beats prior encoder-based models (e.g., Luna) on RAGTruth. ✅ Surpasses fine-tuned Llama-2-13B [2] at a fraction of the size, and is highly efficient at inference. âšĄïž Entirely open-source with an MIT license. 🔓 LettuceDetect keeps your RAG framework fresh by spotting rotten parts of your LLM’s outputs. 😊 Quick links Why LettuceDetect? Large Language Models (LLMs) have made considerable advancements in NLP tasks, like GPT-4 [4], the Llama-3 models [5], or Mistral [6] (and many more). Despite the success of LLMs, Hallucinations remain a key obstacle deploying LLMs in high-stakes scenarios (such as in healthcare or legal) [7,8]. Retrieval-Augmented Generation (RAG) attempts to mitigate hallucinations by grounding an LLM’s responses in retrieved documents, providing external knowledge that the model can reference [9]. But even though RAG is a powerful method to reduce hallucinations, LLMs still suffer from hallucinations in these settings [1]. Hallucinations are information in the output that is nonsensical, factually incorrect, or inconsistent with the retrieved context [8]. Ji et al. [10] categorizes hallucinations into: Intrinsic hallucinations: Stemming from the model’s preexisting internal knowledge. Extrinsic hallucinations: Occurring when the answer conflicts with the context or references provided While RAG approaches can mitigate intrinsic hallucinations, they are not immune to extrinsic hallucinations. Sun et al. [11] showed that models tend to prioritize their intrinsic knowledge over the external context. As LLMs remain prone to hallucinations, their applications in critical domains e.g. medical or legal, can be still flawed. Current solutions for hallucination detection Current solutions for hallucination detection can be categorized into different categories based on the approach they take: Prompt-based detectors These methods (e.g., RAGAS, Trulens, ARES) typically leverage zero-shot or few-shot prompts to detect hallucinations. They often rely on large LLMs (like GPT-4) and employ strategies such as SelfCheckGPT [12], LM vs. LM [13], or Chainpoll [14]. While often effective, they can be computationally expensive due to repeated LLM calls. Fine-tuned LLM detectors Large models (e.g., Llama-2, Llama-3) can be fine-tuned for hallucination detection [1,15]. This can yield high accuracy (as shown by the RAGTruth authors using Llama-2-13B or the RAG-HAT work on Llama-3-8B) but is resource-intensive to train and deploy. Inference costs also tend to be high due to their size and slower speeds. Encoder-based detectors Models like Luna [2] rely on a BERT-style encoder (often limited to 512 tokens) for token-level classification. These methods are generally more efficient than running a full LLM at inference but are constrained by short context windows and attention mechanisms optimized for smaller inputs. ModernBERT for long context ModernBERT [3] is a drop-in replacement for BERT and is a state-of-the-art encoder-only transformers architecture that incorporates several modern design improvements over the original BERT model such as it uses Rotary Positional Embeddings (RoPe) to handle sequences of up to 8,192 tokens, unpadding optimization to eliminate wasted computation on padding tokens, and GeGLU activation layers for enhanced expressiveness and alternating attention for more efficient attention computation. LettuceDetect capitalizes on ModernBERT’s extended context window to build a token-level classifier for hallucination detection. This approach sidesteps many limitations of older BERT-based models (e.g., short context bounds) and avoids the inference overhead of large LLM-based detectors. Our experiments show that LettuceDetect outperforms other encoder-based systems while remaining competitive with fine-tuned LLM detectors at a fraction of their computational cost. Data RAGTruth is the first large-scale open-source (MIT License) benchmark specifically designed to evaluate hallucination detection in Retrieval-Augmented Generation (RAG) settings. It contains 18,000 annotated examples spanning multiple tasks: Question answering (QA): Sampled from the MS MARCO dataset, where up to three documents are retrieved for each question, and an LLM is prompted to produce an answer. Data-to-text generation: Based on the Yelp Open Dataset, where LLMs generate reviews for sampled businesses. News summarization: Uses random documents from CNN/DailyMail, prompting LLMs to provide summaries. Each example in RAGTruth includes: A question and prompt, One or more reference passages or context documents, An LLM-generated response, Human annotations at the span level indicating whether any part of the answer is factually unsupported or contradictory to the reference passages. Below is an illustrative sample from RAGTruth, showing an instance of hallucination: QUESTION How to explain quartiles? REFERENCE Passage 2:Second quartile (Q2) which is more commonly known as median splits the data in half (50%).Median divides the data into a lower half and an upper half.Third quartile (Q3), also known as upper quartile, splits lowest 75% (or highest 25%) of data.(
) RESPONSE The first quartile (Q1) splits the lowest 25% of the data, while the second quartile (Q2) splits the data into two equal halves, with the median being the middle value of the lower half. Finally, the third quartile (Q3) splits the highest 75% of the data.(
) ANNOTATION Span: “highest 75%” Type: Evident Conflict Reason: Original text: “lowest 75% (or highest 25%)” Generative text: “highest 75%” This example showcases a RAGTruth annotation involving specific spans in the answer that conflict with the reference passages, providing both the type of hallucination (e.g., Evident Conflict) and a human-readable rationale. Our method only uses binary annotations and does not use the type of hallucination annotations. Method Image by Author A high-level depiction of LettuceDetect. Here, an example Question, Context, and Answer triplet is processed. First, the text is tokenized, after which LettuceDetect performs token-level classification. Tokens from both the question and context are masked (indicated by the red line in the figure) to exclude them from the loss function. Each token in the answer receives a probability indicating whether it is hallucinated or supported. For span-level detection, we merge consecutive tokens with hallucination probabilities above 0.5 into a single predicted span. We train ModernBERT-base and ModernBERT-large variants as token-classification models on the RAGTruth dataset. The input to the model is a concatenation of Context, Question, and Answer segments, with specialized tokens ([CLS]) (for the context) and ([SEP]) (as separators). We limit the sequence length to 4,096 tokens for computational feasibility, though ModernBERT can theoretically handle up to 8,192 tokens. Tokenization and data processing Tokenizer: We employ AutoTokenizer from the Transformers library to handle subword Tokenization, inserting [CLS] and [SEP] appropriately. Labeling: Context/question tokens are masked (i.e., assigned a label of -100 in PyTorch) so that they do not contribute to the loss. Each answer token receives a label of 0 (supported) or 1 (hallucinated). Model architecture Our models build on Hugging Face’s AutoModelForTokenClassification, using ModernBERT as the encoder and a classification head on top. Unlike some previous encoder-based approaches (e.g., ones pre-trained on NLI tasks), our method uses only ModernBERT with no additional pretraining stage. Training configuration Optimizer: AdamW, with a learning rate of 1 * 10^-5 and weight decay of 0.01. Hardware: Single NVIDIA A100 GPU. Epochs: 6 total training epochs. Batching: Batch size of 8, Data loading with PyTorch DataLoader (shuffling enabled), Dynamic padding via DataCollatorForTokenClassification to handle variable-length sequences efficiently. During training, we monitor token-level F1 scores on a validation split, saving checkpoints using the safetensors format. Once training is complete, we upload the best-performing models to Hugging Face for public access. At inference time, the model outputs a probability of hallucination for each token in the answer. We aggregate consecutive tokens exceeding a 0.5 threshold to produce span-level predictions, indicating exactly which segments of the answer are likely to be hallucinated. The figure above illustrates this workflow. Next, we provide a more detailed evaluation of the model’s performance. Results We evaluate our models on the RAGTruth test set across all task types (Question Answering, Data-to-Text, and Summarization). For each example, RAGTruth includes manually annotated spans indicating hallucinated content. Example-level results We first assess the example-level question: Does the generated answer contain any hallucination at all? Our large model (lettucedetect-large-v1) attains an overall F1 score of 79.22%, surpassing: GPT-4 (63.4%), Luna (65.4%) (the previous state of the art encoder-based model), Fine-tuned Llama-2-13B (78.7%) as presented in the RAGTruth paper. It is second only to the fine-tuned Llama-3-8B from the RAG-HAT paper [15] (83.9%), but LettuceDetect is significantly smaller and faster to run. Meanwhile, our base model (lettucedetect-base-v1) remains highly competitive while using fewer parameters. Image by Author Above is a comparison table illustrating how LettuceDetect aligns against both prompt-based methods (e.g., GPT-4) and alternative encoder-based solutions (e.g., Luna). Overall, lettucedetect-large-v1 and lettucedect-base-v1 are very performant models, while being very effective in inference settings. Span-level results Beyond detecting if an answer contains hallucinations, we also examine LettuceDetect’s ability to identify the exact spans of unsupported content. Here, LettuceDetect achieves state-of-the-art results among models that have reported span-level performance, substantially outperforming the fine-tuned Llama-2-13B model from the RAGTruth paper [1] and other baselines. Image by Author Most methods, like RAG-HAT [15], do not report span-level metrics, so we do not compare to them here. Inference efficiency Both lettucedetect-base-v1 and lettucedetect-large-v1 require fewer parameters than typical LLM-based detectors (e.g., GPT-4 or Llama-3-8B) and can process 30–60 examples per second on a single NVIDIA A100 GPU. This makes them practical for industrial workloads, real-time user-facing systems, and resource-constrained environments. Overall, these results show that LettuceDetect has a good balance: it achieves near state-of-the-art accuracy at a fraction of the size and cost compared to large LLM-based judges, while offering precise, token-level hallucination detection. Get going Install the package: pip install lettucedetect Then, you can use the package as follows: from lettucedetect.models.inference import HallucinationDetector # For a transformer-based approach: detector = HallucinationDetector( method=”transformer”, model_path=”KRLabsOrg/lettucedect-base-modernbert-en-v1″ ) contexts = [“France is a country in Europe. The capital of France is Paris. The population of France is 67 million.”,] question = “What is the capital of France? What is the population of France?” answer = “The capital of France is Paris. The population of France is 69 million.” # Get span-level predictions indicating which parts of the answer are considered hallucinated. predictions = detector.predict(context=contexts, question=question, answer=answer, output_format=”spans”) print(“Predictions:”, predictions) # Predictions: [{‘start’: 31, ‘end’: 71, ‘confidence’: 0.9944414496421814, ‘text’: ‘ The population of France is 69 million.’}] Conclusion We introduced LettuceDetect, a lightweight and efficient framework for hallucination detection in RAG systems. By utilizing ModernBERT’s extended context capabilities, our models achieve strong performance on the RAGTruth benchmark while retaining high inference efficiency. This work lays the groundwork for future research directions, such as expanding to additional datasets, supporting multiple languages, and exploring more advanced architectures. Even at this stage, LettuceDetect demonstrates that effective hallucination detection can be achieved using lean, purpose-built encoder-based models. Citation If you find this work useful, please cite it as follows: @misc{Kovacs:2025,       title={LettuceDetect: A Hallucination Detection Framework for RAG Applications},        author={ÁdĂĄm KovĂĄcs and GĂĄbor Recski},       year={2025},       eprint={2502.17125},       archivePrefix={arXiv},       primaryClass={cs.CL},       url={https://arxiv.org/abs/2502.17125},  } Also, if you use our code, please don’t forget to give us a star ⭐ on our GitHub repository here. References [1] Niu et al., 2024, RAGTruth: A Dataset for Hallucination Detection in Retrieval-Augmented Generation [2] Luna: A Simple and Effective Encoder-Based Model for Hallucination Detection in Retrieval-Augmented Generation [3] ModernBERT: A Modern BERT Model for Long-Context Processing [4] GPT-4 report [5] Llama-3 report [6] Mistral 7B [7] Kaddour et al., 2023, Challenges and Applications of Large Language Models [8] Huang et al., 2025, A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions [9] Gao et al., 2024, Retrieval-Augmented Generation for Large Language Models: A Survey [10] Ji et al., 2023, Survey of Hallucination in Natural Language Generation [11] Sun et al., 2025, ReDeEP: Detecting Hallucination in Retrieval-Augmented Generation via Mechanistic Interpretability [12] Manakul et al., 2023, SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models [13] Cohen et al., 2023, LM vs LM: Detecting Factual Errors via Cross Examination [14] Friel et al., 2023, Chainpoll: A high efficacy method for LLM hallucination detection [15] Song et al., 2024, RAG-HAT: A Hallucination-Aware Tuning Pipeline for {LLM} in Retrieval-Augmented Generation [16] Devlin et al., 2019, BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

Originally published on HuggingFace

TL;DR

We present LettuceDetect, a lightweight hallucination detector for Retrieval-Augmented Generation (RAG) pipelines. It is an encoder-based model built on ModernBERT, released under the MIT license with ready-to-use Python packages and pretrained models.

  • What: LettuceDetect is a token-level detector that flags unsupported segments in LLM answers. đŸ„Ź
  • How: Trained on RAGTruth (18k examples), leveraging ModernBERT for context lengths up to 4k tokens. 🚀
  • Why: It addresses (1) the context-window limits in prior encoder-only models, and (2) the high compute costs of LLM-based detectors. ⚖
  • Highlights:
    • Beats prior encoder-based models (e.g., Luna) on RAGTruth. ✅
    • Surpasses fine-tuned Llama-2-13B [2] at a fraction of the size, and is highly efficient at inference. âšĄïž
    • Entirely open-source with an MIT license. 🔓

LettuceDetect keeps your RAG framework fresh by spotting rotten parts of your LLM’s outputs. 😊

Quick links


Why LettuceDetect?

Large Language Models (LLMs) have made considerable advancements in NLP tasks, like GPT-4 [4], the Llama-3 models [5], or Mistral [6] (and many more). Despite the success of LLMs, Hallucinations remain a key obstacle deploying LLMs in high-stakes scenarios (such as in healthcare or legal) [7,8].

Retrieval-Augmented Generation (RAG) attempts to mitigate hallucinations by grounding an LLM’s responses in retrieved documents, providing external knowledge that the model can reference [9]. But even though RAG is a powerful method to reduce hallucinations, LLMs still suffer from hallucinations in these settings [1]. Hallucinations are information in the output that is nonsensical, factually incorrect, or inconsistent with the retrieved context [8]. Ji et al. [10] categorizes hallucinations into:

  • Intrinsic hallucinations: Stemming from the model’s preexisting internal knowledge.
  • Extrinsic hallucinations: Occurring when the answer conflicts with the context or references provided

While RAG approaches can mitigate intrinsic hallucinations, they are not immune to extrinsic hallucinations. Sun et al. [11] showed that models tend to prioritize their intrinsic knowledge over the external context. As LLMs remain prone to hallucinations, their applications in critical domains e.g. medical or legal, can be still flawed.

Current solutions for hallucination detection

Current solutions for hallucination detection can be categorized into different categories based on the approach they take:

  1. Prompt-based detectors These methods (e.g., RAGAS, Trulens, ARES) typically leverage zero-shot or few-shot prompts to detect hallucinations. They often rely on large LLMs (like GPT-4) and employ strategies such as SelfCheckGPT [12], LM vs. LM [13], or Chainpoll [14]. While often effective, they can be computationally expensive due to repeated LLM calls.
  2. Fine-tuned LLM detectors Large models (e.g., Llama-2, Llama-3) can be fine-tuned for hallucination detection [1,15]. This can yield high accuracy (as shown by the RAGTruth authors using Llama-2-13B or the RAG-HAT work on Llama-3-8B) but is resource-intensive to train and deploy. Inference costs also tend to be high due to their size and slower speeds.
  3. Encoder-based detectors Models like Luna [2] rely on a BERT-style encoder (often limited to 512 tokens) for token-level classification. These methods are generally more efficient than running a full LLM at inference but are constrained by short context windows and attention mechanisms optimized for smaller inputs.

ModernBERT for long context

ModernBERT [3] is a drop-in replacement for BERT and is a state-of-the-art encoder-only transformers architecture that incorporates several modern design improvements over the original BERT model such as it uses Rotary Positional Embeddings (RoPe) to handle sequences of up to 8,192 tokens, unpadding optimization to eliminate wasted computation on padding tokens, and GeGLU activation layers for enhanced expressiveness and alternating attention for more efficient attention computation.

LettuceDetect capitalizes on ModernBERT’s extended context window to build a token-level classifier for hallucination detection. This approach sidesteps many limitations of older BERT-based models (e.g., short context bounds) and avoids the inference overhead of large LLM-based detectors. Our experiments show that LettuceDetect outperforms other encoder-based systems while remaining competitive with fine-tuned LLM detectors at a fraction of their computational cost.

Data

RAGTruth is the first large-scale open-source (MIT License) benchmark specifically designed to evaluate hallucination detection in Retrieval-Augmented Generation (RAG) settings. It contains 18,000 annotated examples spanning multiple tasks:

  • Question answering (QA): Sampled from the MS MARCO dataset, where up to three documents are retrieved for each question, and an LLM is prompted to produce an answer.
  • Data-to-text generation: Based on the Yelp Open Dataset, where LLMs generate reviews for sampled businesses.
  • News summarization: Uses random documents from CNN/DailyMail, prompting LLMs to provide summaries.

Each example in RAGTruth includes:

  1. A question and prompt,
  2. One or more reference passages or context documents,
  3. An LLM-generated response,
  4. Human annotations at the span level indicating whether any part of the answer is factually unsupported or contradictory to the reference passages.

Below is an illustrative sample from RAGTruth, showing an instance of hallucination:


QUESTION

How to explain quartiles?

REFERENCE

Passage 2:
Second quartile (Q2) which is more commonly known as median splits the data in half (50%).
Median divides the data into a lower half and an upper half.
Third quartile (Q3), also known as upper quartile, splits lowest 75% (or highest 25%) of data.
(
)

RESPONSE

The first quartile (Q1) splits the lowest 25% of the data, while the second quartile (Q2) splits the data into two equal halves, with the median being the middle value of the lower half. Finally, the third quartile (Q3) splits the highest 75% of the data.
(
)

ANNOTATION

  • Span: “highest 75%”
  • Type: Evident Conflict
  • Reason:
    • Original text: “lowest 75% (or highest 25%)”
    • Generative text: “highest 75%”

This example showcases a RAGTruth annotation involving specific spans in the answer that conflict with the reference passages, providing both the type of hallucination (e.g., Evident Conflict) and a human-readable rationale. Our method only uses binary annotations and does not use the type of hallucination annotations.

Method

Diagram of LettuceDetect
Image by Author

A high-level depiction of LettuceDetect. Here, an example Question, Context, and Answer triplet is processed. First, the text is tokenized, after which LettuceDetect performs token-level classification. Tokens from both the question and context are masked (indicated by the red line in the figure) to exclude them from the loss function. Each token in the answer receives a probability indicating whether it is hallucinated or supported. For span-level detection, we merge consecutive tokens with hallucination probabilities above 0.5 into a single predicted span.


We train ModernBERT-base and ModernBERT-large variants as token-classification models on the RAGTruth dataset. The input to the model is a concatenation of Context, Question, and Answer segments, with specialized tokens ([CLS]) (for the context) and ([SEP]) (as separators). We limit the sequence length to 4,096 tokens for computational feasibility, though ModernBERT can theoretically handle up to 8,192 tokens.

Tokenization and data processing

  • Tokenizer: We employ AutoTokenizer from the Transformers library to handle subword Tokenization, inserting [CLS] and [SEP] appropriately.
  • Labeling:
    • Context/question tokens are masked (i.e., assigned a label of -100 in PyTorch) so that they do not contribute to the loss.
    • Each answer token receives a label of 0 (supported) or 1 (hallucinated).

Model architecture

Our models build on Hugging Face’s AutoModelForTokenClassification, using ModernBERT as the encoder and a classification head on top. Unlike some previous encoder-based approaches (e.g., ones pre-trained on NLI tasks), our method uses only ModernBERT with no additional pretraining stage.

Training configuration

  • Optimizer: AdamW, with a learning rate of 1 * 10^-5 and weight decay of 0.01.
  • Hardware: Single NVIDIA A100 GPU.
  • Epochs: 6 total training epochs.
  • Batching:
    • Batch size of 8,
    • Data loading with PyTorch DataLoader (shuffling enabled),
    • Dynamic padding via DataCollatorForTokenClassification to handle variable-length sequences efficiently.

During training, we monitor token-level F1 scores on a validation split, saving checkpoints using the safetensors format. Once training is complete, we upload the best-performing models to Hugging Face for public access.

At inference time, the model outputs a probability of hallucination for each token in the answer. We aggregate consecutive tokens exceeding a 0.5 threshold to produce span-level predictions, indicating exactly which segments of the answer are likely to be hallucinated. The figure above illustrates this workflow.

Next, we provide a more detailed evaluation of the model’s performance.

Results

We evaluate our models on the RAGTruth test set across all task types (Question Answering, Data-to-Text, and Summarization). For each example, RAGTruth includes manually annotated spans indicating hallucinated content.

Example-level results

We first assess the example-level question: Does the generated answer contain any hallucination at all? Our large model (lettucedetect-large-v1) attains an overall F1 score of 79.22%, surpassing:

  • GPT-4 (63.4%),
  • Luna (65.4%) (the previous state of the art encoder-based model),
  • Fine-tuned Llama-2-13B (78.7%) as presented in the RAGTruth paper.

It is second only to the fine-tuned Llama-3-8B from the RAG-HAT paper [15] (83.9%), but LettuceDetect is significantly smaller and faster to run. Meanwhile, our base model (lettucedetect-base-v1) remains highly competitive while using fewer parameters.

Comparison table illustrating how LettuceDetect aligns against both prompt-based methods (e.g., GPT-4) and alternative encoder-based solutions (e.g., Luna)
Image by Author

Above is a comparison table illustrating how LettuceDetect aligns against both prompt-based methods (e.g., GPT-4) and alternative encoder-based solutions (e.g., Luna). Overall, lettucedetect-large-v1 and lettucedect-base-v1 are very performant models, while being very effective in inference settings.

Span-level results

Beyond detecting if an answer contains hallucinations, we also examine LettuceDetect’s ability to identify the exact spans of unsupported content. Here, LettuceDetect achieves state-of-the-art results among models that have reported span-level performance, substantially outperforming the fine-tuned Llama-2-13B model from the RAGTruth paper [1] and other baselines.

Image by Author

Most methods, like RAG-HAT [15], do not report span-level metrics, so we do not compare to them here.

Inference efficiency

Both lettucedetect-base-v1 and lettucedetect-large-v1 require fewer parameters than typical LLM-based detectors (e.g., GPT-4 or Llama-3-8B) and can process 30–60 examples per second on a single NVIDIA A100 GPU. This makes them practical for industrial workloads, real-time user-facing systems, and resource-constrained environments.

Overall, these results show that LettuceDetect has a good balance: it achieves near state-of-the-art accuracy at a fraction of the size and cost compared to large LLM-based judges, while offering precise, token-level hallucination detection.

Get going

Install the package:

pip install lettucedetect

Then, you can use the package as follows:

from lettucedetect.models.inference import HallucinationDetector

# For a transformer-based approach:

detector = HallucinationDetector(

    method="transformer", model_path="KRLabsOrg/lettucedect-base-modernbert-en-v1"

)

contexts = ["France is a country in Europe. The capital of France is Paris. The population of France is 67 million.",]

question = "What is the capital of France? What is the population of France?"

answer = "The capital of France is Paris. The population of France is 69 million."

# Get span-level predictions indicating which parts of the answer are considered hallucinated.

predictions = detector.predict(context=contexts, question=question, answer=answer, output_format="spans")

print("Predictions:", predictions)

# Predictions: [{'start': 31, 'end': 71, 'confidence': 0.9944414496421814, 'text': ' The population of France is 69 million.'}]

Conclusion

We introduced LettuceDetect, a lightweight and efficient framework for hallucination detection in RAG systems. By utilizing ModernBERT’s extended context capabilities, our models achieve strong performance on the RAGTruth benchmark while retaining high inference efficiency. This work lays the groundwork for future research directions, such as expanding to additional datasets, supporting multiple languages, and exploring more advanced architectures. Even at this stage, LettuceDetect demonstrates that effective hallucination detection can be achieved using lean, purpose-built encoder-based models.

Citation

If you find this work useful, please cite it as follows:

@misc{Kovacs:2025,
      title={LettuceDetect: A Hallucination Detection Framework for RAG Applications}, 
      author={Ádåm Kovåcs and Gåbor Recski},
      year={2025},
      eprint={2502.17125},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2502.17125}, 
}

Also, if you use our code, please don’t forget to give us a star ⭐ on our GitHub repository here.

References

[1] Niu et al., 2024, RAGTruth: A Dataset for Hallucination Detection in Retrieval-Augmented Generation

[2] Luna: A Simple and Effective Encoder-Based Model for Hallucination Detection in Retrieval-Augmented Generation

[3] ModernBERT: A Modern BERT Model for Long-Context Processing

[4] GPT-4 report

[5] Llama-3 report

[6] Mistral 7B

[7] Kaddour et al., 2023, Challenges and Applications of Large Language Models

[8] Huang et al., 2025, A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions

[9] Gao et al., 2024, Retrieval-Augmented Generation for Large Language Models: A Survey

[10] Ji et al., 2023, Survey of Hallucination in Natural Language Generation

[11] Sun et al., 2025, ReDeEP: Detecting Hallucination in Retrieval-Augmented Generation via Mechanistic Interpretability

[12] Manakul et al., 2023, SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models

[13] Cohen et al., 2023, LM vs LM: Detecting Factual Errors via Cross Examination

[14] Friel et al., 2023, Chainpoll: A high efficacy method for LLM hallucination detection

[15] Song et al., 2024, RAG-HAT: A Hallucination-Aware Tuning Pipeline for {LLM} in Retrieval-Augmented Generation

[16] Devlin et al., 2019, BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

Shape
Shape
Stay Ahead

Explore More Insights

Stay ahead with more perspectives on cutting-edge power, infrastructure, energy,  bitcoin and AI solutions. Explore these articles to uncover strategies and insights shaping the future of industries.

Shape

Why enterprise networks need both reach and resilience

As enterprises expand across regions, so do their cloud platforms and digital ecosystems. But with the rise of AI and its unprecedented appetite for data, networks are now under more pressure. Many businesses are learning the limits of legacy architecture the hard way. In the race to meet today’s standard

Read More »

Can your network handle the demands of today’s connected workplace?

Enterprise innovation is accelerating at a dizzying pace, from AI-powered applications and real-time data platforms to immersive customer experiences. Solutions like Microsoft Copilot are transforming productivity. Platforms-as-a-Service, such as Lattice, are reimagining how teams collaborate and grow. But beneath this digital renaissance lies the hard truth: none of it works

Read More »

Anzana Plans to Acquire Stake in Burundi-DRC-Rwanda Hydropower Project

Anzana Electric Group has signed a tentative deal with Ruzizi III Holding Power Co. Ltd. (RHPCL) to acquire up to 10 percent in a hydroelectric project that would serve Burundi, the Democratic Republic of the Congo (DRC) and Rwanda. The agreement for the 206-megawatt Ruzizi III Regional Hydropower Project was part of over $2.5 billion in deals and commitments between African and American partners signed during the United States-Africa Business Summit in Luanda, Angola, according to the U.S. State Department. The $760-million project is planned to rise on the Ruzizi River between DRC in Central Africa and Rwanda in Eastern Africa. It is a private-public partnership. RHPCL, a special-purpose vehicle registered in Rwanda, is the private partner to the project company Ruzizi III Energy Limited under a build-own-operate-transfer scheme. RHPCL expects to power about 30 million people across the three neighboring countries, “in a region where 54 percent live below the poverty line and electricity access averages just 24 percent”, said a joint statement by RHPCL and Anzana. “The project will nearly double Burundi’s current capacity, boost Rwanda’s by 30 percent, and deliver critical baseload and dispatchable power to eastern DRC, advancing economic growth, regional integration, and energy security in one of Africa’s most underserved regions”. RHPCL and Anzana, which invests in hydropower and grid distribution projects in East, Central and Southern Africa, committed to negotiating for a binding partnership agreement to be penned by September. “The agreement will outline governance rights, investment commitments, and the trajectory for further collaboration”, the companies said. Anzana chief executive Brian Kelly said, “Through this partnership, we are not only powering homes, communities, and industries, we are helping to drive regional integration, strengthen energy security and stability, and pave the way for expanded U.S. investment and trade in Africa’s energy future”. On the same

Read More »

Russia’s Sanctioned Arctic LNG 2 Raises Output to Record Levels

Russia’s sanctioned Arctic LNG 2 project raised production to record levels during the last days of June as the facility appears to have resumed loading cargoes. Natural gas output at the Novatek PJSC-led facility averaged 14 million cubic meters a day on June 28 and June 29, according to a person with knowledge of the matter.  That’s the highest daily level for the plant, historic data shows. Higher natural gas output doesn’t automatically indicate a hike in LNG production, but historically the plant produced more gas when it was able to load cargoes. In December 2023, when it was launched, Arctic LNG 2 pumped an average of 13.7 million cubic meters of gas a day. The facility located above the Arctic Circle is key for Russia’s ambition to triple LNG production by 2030. Those plans were squeezed by international restrictions after the invasion of Ukraine, but a liquefied gas tanker appeared to load a cargo several days ago, suggesting Russia may be finding ways around the penalties. Gas output at Arctic LNG 2 averaged 8.9 million cubic meters a day during most of June, compared with 9.4 million cubic meters a day the month before, the person said, asking not to be identified because the information isn’t public. Novatek, the largest shareholder of Arctic LNG 2, and the plant’s operator didn’t immediately respond to requests for comments. The Iris tanker — previously known as North Sky and blacklisted by the US, the EU and the UK — left the site Sunday. Its draft level, which the crew inputs manually, has increased, potentially indicating the tanker loaded a cargo there, according to ship-tracking data compiled by Bloomberg. The tanker is heading toward the Arctic port of Murmansk, where it’s expected to arrive July 2. Novatek uses waters near Murmansk to transfer LNG cargoes

Read More »

Oil Gains as Mideast Tensions Reignite

Oil edged up from near the lowest levels in a month as tensions once again flared in the Middle East, returning the spotlight to the fragility of a truce between Israel and Iran. West Texas Intermediate rose 0.5% to settle near $65.50 a barrel, while Brent closed above $67. Volumes were trending lower ahead of Friday’s July 4 holiday in the US. Investors are watching closely to see whether Iran’s inventories of near-bomb-grade uranium have been depleted and whether its moves to cut off communication with key United Nations watchdog officials will trigger another wave of US strikes. President Donald Trump has said the US will “be there” unless Iran backs away from its nuclear program. So far, the conflict has not disrupted flows in the region but the mere possibility of supply interruptions now has some traders taking a wait-and-see approach. During the heat of tensions, a quarterly record of combined options contracts for WTI and Brent changed hands as traders bet on the outcome of these fast-evolving conflicts, based on data from the exchanges. Aside from geopolitics, macro factors also lent conflicting signals to oil. The demand outlook for the US darkened slightly after factory activity contracted in June for a fourth consecutive month, although the labor market showed signs of strength. The Middle East developments took away the focus from a meeting between the Organization of the Petroleum Exporting Countries and its allies. The group is expected to agree to a fourth monthly major supply increase during discussions Sunday, according to a Bloomberg survey, as de facto leader Saudi Arabia continues its bid to reclaim market share. Oil lost almost 10% last quarter in a volatile three months that saw prices drop sharply in April on Trump’s tariff plans, and surge in June after Israel attacked Iran,

Read More »

Senate passes megabill that curbs IRA tax credits, drops wind and solar tax

Dive Brief: The Senate voted Tuesday to pass an amended version of the Republican budget megabill that significantly curtails clean energy tax credits. It does not contain a proposed excise tax on wind and solar projects that caught many by surprise when it was added late Friday. The final version carves out an exception to the bill’s new phaseout deadline for wind and solar project tax credits. Previously, the legislation stipulated that wind and solar projects had to be placed in service by the end of 2027 to qualify for the clean energy production credit. This was amended to exempt projects that begin construction within a year after the signing of the legislation. The bill that made it out of the Senate Finance Committee had softened some of the IRA cuts made in the House. That version was supplanted over the weekend by harsher language that included the now-dead excise tax. The Senate bill now heads back to the House, with Republican leadership in both chambers aiming to deliver the bill to President Trump’s desk for him to sign it into law by Friday. Dive Insight: Sen. Rand Paul, R-Ky., and Sen. Thom Tillis, R-N.C., continued to oppose the legislation after voting against it over the weekend. They were joined by Sen. Susan Collins, R-Maine, along with all Democrats. Vice President JD Vance provided the tiebreaking vote. “Under the last-minute carveout, Big Green has 12 months to initiate as many subsidized projects as it wants using the insanely-easy-to-meet ‘construction’ threshold,” tweeted fossil fuel advocate Alex Epstein, who helped congressional Republicans shape the megabill. “Several Senators have already told me they didn’t know about or understand this last-minute paragraph. If that’s the case they should do whatever they can to fix the situation.”  Harry Godfrey, who leads Advanced Energy United’s federal policy team, said

Read More »

USA Diesel Demand in April Stronger Than Expected Despite Tariffs

US diesel demand, a closely watched measure of the country’s economic health, was higher in April than early weekly estimates, the Energy Information Administration said in its monthly report. Distillate fuel oil demand was 3.88 million barrels a day in April, according to the agency’s latest Petroleum Supply Monthly report released Monday. That is 4.7% higher than early estimates published by the agency in its Wednesday weekly report and 2.2% higher than April 2024. April was a volatile month for diesel futures after President Trump announced sweeping tariffs on April 2, causing prices to tank. Demand for jet fuel was revised down by 5% in the monthly EIA report to 1.76 million barrels a day from estimates of 1.86 millions barrels a day. Those same tariffs also clouded the outlook for air travel, with some Americans opting for road trips over flying as they tighten spending.  Demand for gasoline, the most consumed fuel in the US, was in-line with weekly estimates published earlier this year. Total US liquids production eked out a record-high of 20.83 million barrels a day in April, up roughly 50,000 barrels from the previous month, the report said. The number, which includes crude oil and natural gas liquids, came in roughly 340,000 barrels higher than a previous estimate for the month of April. WHAT DO YOU THINK? Generated by readers, the comments included herein do not reflect the views and opinions of Rigzone. All comments are subject to editorial review. Off-topic, inappropriate or insulting comments will be removed.

Read More »

Sapura Energy Restructuring in ‘Final Stages’

Malaysian oil and gas contractor Sapura Energy Bhd.’s restructuring plan to restore financial stability is entering its “final stages,” according to the company’s first-quarter earnings statement. Regulator Bursa Malaysia’s approval of the blueprint to restructure debt puts the company on a path to exit its financially distressed classification set by Malaysia’s stock exchange, the company said. The country’s anti-graft agency said in March it was investigating the cash-strapped company, which reported a net loss in the quarter ended in April, for alleged misappropriation of funds. Prime Minister Anwar Ibrahim said that month he ordered an audit of the firm and change of management. He also approved a 1.1 billion ringgit ($262.5 million) injection into the company, but denied that it was a bailout.  Sapura Energy’s restructuring is “aimed at addressing the group’s unsustainable debt levels and restoring financial stability,” according to its statement. “Restructuring efforts remain on track and have entered the final stages.” The company said the plan will help reduce total borrowings to 5.6 billion ringgit from 10.8 billion ringgit, without giving a time frame. Sapura Energy reported a first-quarter net loss of 478.0 million ringgit compared with a profit of 82.1 million ringgit a year ago. It cited a challenging project in Angola, as well as lower activity across the oil industry’s operations, maintenance and drilling segments, for the loss. What do you think? We’d love to hear from you, join the conversation on the Rigzone Energy Network. The Rigzone Energy Network is a new social experience created for you and all energy professionals to Speak Up about our industry, share knowledge, connect with peers and industry insiders and engage in a professional community that will empower your career in energy.

Read More »

Data center capacity continues to shift to hyperscalers

However, even though colocation and on-premises data centers will continue to lose share, they will still continue to grow. They just won’t be growing as fast as hyperscalers. So, it creates the illusion of shrinkage when it’s actually just slower growth. In fact, after a sustained period of essentially no growth, on-premises data center capacity is receiving a boost thanks to genAI applications and GPU infrastructure. “While most enterprise workloads are gravitating towards cloud providers or to off-premise colo facilities, a substantial subset are staying on-premise, driving a substantial increase in enterprise GPU servers,” said John Dinsdale, a chief analyst at Synergy Research Group.

Read More »

Oracle inks $30 billion cloud deal, continuing its strong push into AI infrastructure.

He pointed out that, in addition to its continued growth, OCI has a remaining performance obligation (RPO) — total future revenue expected from contracts not yet reported as revenue — of $138 billion, a 41% increase, year over year. The company is benefiting from the immense demand for cloud computing largely driven by AI models. While traditionally an enterprise resource planning (ERP) company, Oracle launched OCI in 2016 and has been strategically investing in AI and data center infrastructure that can support gigawatts of capacity. Notably, it is a partner in the $500 billion SoftBank-backed Stargate project, along with OpenAI, Arm, Microsoft, and Nvidia, that will build out data center infrastructure in the US. Along with that, the company is reportedly spending about $40 billion on Nvidia chips for a massive new data center in Abilene, Texas, that will serve as Stargate’s first location in the country. Further, the company has signaled its plans to significantly increase its investment in Abu Dhabi to grow out its cloud and AI offerings in the UAE; has partnered with IBM to advance agentic AI; has launched more than 50 genAI use cases with Cohere; and is a key provider for ByteDance, which has said it plans to invest $20 billion in global cloud infrastructure this year, notably in Johor, Malaysia. Ellison’s plan: dominate the cloud world CTO and co-founder Larry Ellison announced in a recent earnings call Oracle’s intent to become No. 1 in cloud databases, cloud applications, and the construction and operation of cloud data centers. He said Oracle is uniquely positioned because it has so much enterprise data stored in its databases. He also highlighted the company’s flexible multi-cloud strategy and said that the latest version of its database, Oracle 23ai, is specifically tailored to the needs of AI workloads. Oracle

Read More »

Datacenter industry calls for investment after EU issues water consumption warning

CISPE’s response to the European Commission’s report warns that the resulting regulatory uncertainty could hurt the region’s economy. “Imposing new, standalone water regulations could increase costs, create regulatory fragmentation, and deter investment. This risks shifting infrastructure outside the EU, undermining both sustainability and sovereignty goals,” CISPE said in its latest policy recommendation, Advancing water resilience through digital innovation and responsible stewardship. “Such regulatory uncertainty could also reduce Europe’s attractiveness for climate-neutral infrastructure investment at a time when other regions offer clear and stable frameworks for green data growth,” it added. CISPE’s recommendations are a mix of regulatory harmonization, increased investment, and technological improvement. Currently, water reuse regulation is directed towards agriculture. Updated regulation across the bloc would encourage more efficient use of water in industrial settings such as datacenters, the asosciation said. At the same time, countries struggling with limited public sector budgets are not investing enough in water infrastructure. This could only be addressed by tapping new investment by encouraging formal public-private partnerships (PPPs), it suggested: “Such a framework would enable the development of sustainable financing models that harness private sector innovation and capital, while ensuring robust public oversight and accountability.” Nevertheless, better water management would also require real-time data gathered through networks of IoT sensors coupled to AI analytics and prediction systems. To that end, cloud datacenters were less a drain on water resources than part of the answer: “A cloud-based approach would allow water utilities and industrial users to centralize data collection, automate operational processes, and leverage machine learning algorithms for improved decision-making,” argued CISPE.

Read More »

HPE-Juniper deal clears DOJ hurdle, but settlement requires divestitures

In HPE’s press release following the court’s decision, the vendor wrote that “After close, HPE will facilitate limited access to Juniper’s advanced Mist AIOps technology.” In addition, the DOJ stated that the settlement requires HPE to divest its Instant On business and mandates that the merged firm license critical Juniper software to independent competitors. Specifically, HPE must divest its global Instant On campus and branch WLAN business, including all assets, intellectual property, R&D personnel, and customer relationships, to a DOJ-approved buyer within 180 days. Instant On is aimed primarily at the SMB arena and offers a cloud-based package of wired and wireless networking gear that’s designed for so-called out-of-the-box installation and minimal IT involvement, according to HPE. HPE and Juniper focused on the positive in reacting to the settlement. “Our agreement with the DOJ paves the way to close HPE’s acquisition of Juniper Networks and preserves the intended benefits of this deal for our customers and shareholders, while creating greater competition in the global networking market,” HPE CEO Antonio Neri said in a statement. “For the first time, customers will now have a modern network architecture alternative that can best support the demands of AI workloads. The combination of HPE Aruba Networking and Juniper Networks will provide customers with a comprehensive portfolio of secure, AI-native networking solutions, and accelerate HPE’s ability to grow in the AI data center, service provider and cloud segments.” “This marks an exciting step forward in delivering on a critical customer need – a complete portfolio of modern, secure networking solutions to connect their organizations and provide essential foundations for hybrid cloud and AI,” said Juniper Networks CEO Rami Rahim. “We look forward to closing this transaction and turning our shared vision into reality for enterprise, service provider and cloud customers.”

Read More »

Data center costs surge up to 18% as enterprises face two-year capacity drought

“AI workloads, especially training and archival, can absorb 10-20ms latency variance if offset by 30-40% cost savings and assured uptime,” said Gogia. “Des Moines and Richmond offer better interconnection diversity today than some saturated Tier-1 hubs.” Contract flexibility is also crucial. Rather than traditional long-term leases, enterprises are negotiating shorter agreements with renewal options and exploring revenue-sharing arrangements tied to business performance. Maximizing what you have With expansion becoming more costly, enterprises are getting serious about efficiency through aggressive server consolidation, sophisticated virtualization and AI-driven optimization tools that squeeze more performance from existing space. The companies performing best in this constrained market are focusing on optimization rather than expansion. Some embrace hybrid strategies blending existing on-premises infrastructure with strategic cloud partnerships, reducing dependence on traditional colocation while maintaining control over critical workloads. The long wait When might relief arrive? CBRE’s analysis shows primary markets had a record 6,350 MW under construction at year-end 2024, more than double 2023 levels. However, power capacity constraints are forcing aggressive pre-leasing and extending construction timelines to 2027 and beyond. The implications for enterprises are stark: with construction timelines extending years due to power constraints, companies are essentially locked into current infrastructure for at least the next few years. Those adapting their strategies now will be better positioned when capacity eventually returns.

Read More »

Cisco backs quantum networking startup Qunnect

In partnership with Deutsche Telekom’s T-Labs, Qunnect has set up quantum networking testbeds in New York City and Berlin. “Qunnect understands that quantum networking has to work in the real world, not just in pristine lab conditions,” Vijoy Pandey, general manager and senior vice president of Outshift by Cisco, stated in a blog about the investment. “Their room-temperature approach aligns with our quantum data center vision.” Cisco recently announced it is developing a quantum entanglement chip that could ultimately become part of the gear that will populate future quantum data centers. The chip operates at room temperature, uses minimal power, and functions using existing telecom frequencies, according to Pandey.

Read More »

Microsoft will invest $80B in AI data centers in fiscal 2025

And Microsoft isn’t the only one that is ramping up its investments into AI-enabled data centers. Rival cloud service providers are all investing in either upgrading or opening new data centers to capture a larger chunk of business from developers and users of large language models (LLMs).  In a report published in October 2024, Bloomberg Intelligence estimated that demand for generative AI would push Microsoft, AWS, Google, Oracle, Meta, and Apple would between them devote $200 billion to capex in 2025, up from $110 billion in 2023. Microsoft is one of the biggest spenders, followed closely by Google and AWS, Bloomberg Intelligence said. Its estimate of Microsoft’s capital spending on AI, at $62.4 billion for calendar 2025, is lower than Smith’s claim that the company will invest $80 billion in the fiscal year to June 30, 2025. Both figures, though, are way higher than Microsoft’s 2020 capital expenditure of “just” $17.6 billion. The majority of the increased spending is tied to cloud services and the expansion of AI infrastructure needed to provide compute capacity for OpenAI workloads. Separately, last October Amazon CEO Andy Jassy said his company planned total capex spend of $75 billion in 2024 and even more in 2025, with much of it going to AWS, its cloud computing division.

Read More »

John Deere unveils more autonomous farm machines to address skill labor shortage

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Self-driving tractors might be the path to self-driving cars. John Deere has revealed a new line of autonomous machines and tech across agriculture, construction and commercial landscaping. The Moline, Illinois-based John Deere has been in business for 187 years, yet it’s been a regular as a non-tech company showing off technology at the big tech trade show in Las Vegas and is back at CES 2025 with more autonomous tractors and other vehicles. This is not something we usually cover, but John Deere has a lot of data that is interesting in the big picture of tech. The message from the company is that there aren’t enough skilled farm laborers to do the work that its customers need. It’s been a challenge for most of the last two decades, said Jahmy Hindman, CTO at John Deere, in a briefing. Much of the tech will come this fall and after that. He noted that the average farmer in the U.S. is over 58 and works 12 to 18 hours a day to grow food for us. And he said the American Farm Bureau Federation estimates there are roughly 2.4 million farm jobs that need to be filled annually; and the agricultural work force continues to shrink. (This is my hint to the anti-immigration crowd). John Deere’s autonomous 9RX Tractor. Farmers can oversee it using an app. While each of these industries experiences their own set of challenges, a commonality across all is skilled labor availability. In construction, about 80% percent of contractors struggle to find skilled labor. And in commercial landscaping, 86% of landscaping business owners can’t find labor to fill open positions, he said. “They have to figure out how to do

Read More »

2025 playbook for enterprise AI success, from agents to evals

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More 2025 is poised to be a pivotal year for enterprise AI. The past year has seen rapid innovation, and this year will see the same. This has made it more critical than ever to revisit your AI strategy to stay competitive and create value for your customers. From scaling AI agents to optimizing costs, here are the five critical areas enterprises should prioritize for their AI strategy this year. 1. Agents: the next generation of automation AI agents are no longer theoretical. In 2025, they’re indispensable tools for enterprises looking to streamline operations and enhance customer interactions. Unlike traditional software, agents powered by large language models (LLMs) can make nuanced decisions, navigate complex multi-step tasks, and integrate seamlessly with tools and APIs. At the start of 2024, agents were not ready for prime time, making frustrating mistakes like hallucinating URLs. They started getting better as frontier large language models themselves improved. “Let me put it this way,” said Sam Witteveen, cofounder of Red Dragon, a company that develops agents for companies, and that recently reviewed the 48 agents it built last year. “Interestingly, the ones that we built at the start of the year, a lot of those worked way better at the end of the year just because the models got better.” Witteveen shared this in the video podcast we filmed to discuss these five big trends in detail. Models are getting better and hallucinating less, and they’re also being trained to do agentic tasks. Another feature that the model providers are researching is a way to use the LLM as a judge, and as models get cheaper (something we’ll cover below), companies can use three or more models to

Read More »

OpenAI’s red teaming innovations define new essentials for security leaders in the AI era

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More OpenAI has taken a more aggressive approach to red teaming than its AI competitors, demonstrating its security teams’ advanced capabilities in two areas: multi-step reinforcement and external red teaming. OpenAI recently released two papers that set a new competitive standard for improving the quality, reliability and safety of AI models in these two techniques and more. The first paper, “OpenAI’s Approach to External Red Teaming for AI Models and Systems,” reports that specialized teams outside the company have proven effective in uncovering vulnerabilities that might otherwise have made it into a released model because in-house testing techniques may have missed them. In the second paper, “Diverse and Effective Red Teaming with Auto-Generated Rewards and Multi-Step Reinforcement Learning,” OpenAI introduces an automated framework that relies on iterative reinforcement learning to generate a broad spectrum of novel, wide-ranging attacks. Going all-in on red teaming pays practical, competitive dividends It’s encouraging to see competitive intensity in red teaming growing among AI companies. When Anthropic released its AI red team guidelines in June of last year, it joined AI providers including Google, Microsoft, Nvidia, OpenAI, and even the U.S.’s National Institute of Standards and Technology (NIST), which all had released red teaming frameworks. Investing heavily in red teaming yields tangible benefits for security leaders in any organization. OpenAI’s paper on external red teaming provides a detailed analysis of how the company strives to create specialized external teams that include cybersecurity and subject matter experts. The goal is to see if knowledgeable external teams can defeat models’ security perimeters and find gaps in their security, biases and controls that prompt-based testing couldn’t find. What makes OpenAI’s recent papers noteworthy is how well they define using human-in-the-middle

Read More »