INTELLIGENCE VALUE
Large language models (LLMs) present massive opportunities to increase the quality and efficiency of intelligence analysis; however, LLMs are known to exhibit vulnerabilities that pose substantial risks. BENGAL aims to understand the landscape of LLM threats and vulnerabilities, with the goal of developing novel technologies to analyze and address these shortcomings, enabling the Intelligence Community (IC) to safely harness LLM technology across a broad range of critical applications.
Summary
The IC is interested in safe uses of LLMs (multi-modal and text-only) for a wide variety of applications including the rapid summarization and contextualization of relevant multilingual information. These applications must preserve attribution to original sources, be free of erroneous outputs, safegaurd sensitive information from unauthorized disclosure, and function in the presence of poisined or corrupted inputs. The US Government is also interested in identifying and mitigating hazardous use of LLMs by potentially nefarious actors.
The BENGAL super seedling program is a two-year effort beginning in 2026 that aims to explore, quantify, and mitigate the threats and vulnerabilities of LLMs (multi-modal and text-only). Performers will focus on one or more of the topic domains below, clearly articulate a taxonomy of threat modes within their domain of interest and develop technologies to detect, and defeat or mitigate these threats.
AI hallucinations and inferences: Techniques to detect and mitigate an ungrounded, incorrect, or misleading output, or hallucination, from an LLM while minimizing the impact on the LLM’s capabilities to produce correct and/or plausible inferences.
Safe information flow in sensitive environments: Methods to identify inputs/outputs from a(n) user/LLM that may be trying to aggregate innocuous facts to derive sensitive information, methods to decouple sensitive information, and methods to “unlearn” information deemed sensitive from a pre-trained or fine-tuned LLM.
Working resiliently to improve poisoned sources: Techniques to improve LLM reliability through evaluating of the quality of sources, explainable techniques to infer source intentions, and extracting reliable information from biased or incomplete sources.
Performers will pursue high-risk, high-payoff research and deliver to the IC turn-key prototype software which is independently validated by testing and evaluation partners.
Proposers' Day Information
BENGAL Proposers' Day Registration Site
BENGAL Draft Technical Description