BENGAL

Bias Effects and Notable Generative AI Limitations

INTELLIGENCE VALUE

Large language models (LLMs) present massive opportunities to increase the quality and efficiency of intelligence analysis; however, LLMs are known to exhibit vulnerabilities that pose substantial risks. BENGAL aims to understand the landscape of LLM threats and vulnerabilities, with the goal of developing novel technologies to analyze and address these shortcomings, enabling the Intelligence Community (IC) to safely harness LLM technology across a broad range of critical applications.

Summary

The IC is interested in safe uses of LLMs (multi-modal and text-only) for a wide variety of applications including the rapid summarization and contextualization of relevant multilingual information. These applications must preserve attribution to original sources, be free of erroneous outputs, safegaurd sensitive information from unauthorized disclosure, and function in the presence of poisined or corrupted inputs. The US Government is also interested in identifying and mitigating hazardous use of LLMs by potentially nefarious actors.

The BENGAL super seedling program is a two-year effort beginning in 2026 that aims to explore, quantify, and mitigate the threats and vulnerabilities of LLMs (multi-modal and text-only). Performers will focus on one or more of the topic domains below, clearly articulate a taxonomy of threat modes within their domain of interest and develop technologies to detect, and defeat or mitigate these threats.

AI hallucinations and inferences: Techniques to detect and mitigate an ungrounded, incorrect, or misleading output, or hallucination, from an LLM while minimizing the impact on the LLM’s capabilities to produce correct and/or plausible inferences.

Safe information flow in sensitive environments: Methods to identify inputs/outputs from a(n) user/LLM that may be trying to aggregate innocuous facts to derive sensitive information, methods to decouple sensitive information, and methods to “unlearn” information deemed sensitive from a pre-trained or fine-tuned LLM.

Working resiliently to improve poisoned sources: Techniques to improve LLM reliability through evaluating of the quality of sources, explainable techniques to infer source intentions, and extracting reliable information from biased or incomplete sources.

Performers will pursue high-risk, high-payoff research and deliver to the IC turn-key prototype software which is independently validated by testing and evaluation partners.

 

Proposers' Day Information

BENGAL Proposers' Day Registration Site

SAM.gov Reference

BENGAL Teaming Form

BENGAL Draft Technical Description

 

BENGAL Logo

Contact Information

Program Manager

Dr. Steven Rieber

Broad Agency Announcement (BAA)

Link(s) to BAA

BENGAL BAA

Solicitation Status

CLOSED

Proposers' Day Date

October 24, 2023

BAA Release Date

November 17, 2023

Proposal Due Date

February 9, 2024

Program Summary

Testing and Evaluation Partners

  • Los Alamos National Laboratory

Prime Performers

  • Carnegie Mellon University
  • Columbia University
  • IBM Corporation
  • University of Pennsylvania
  • University of Southern California