AI Risk · Privacy

Leaking Proprietary Data

A generative model can reveal proprietary or confidential information from the training data or connected systems (e.g. in a RAG set-up).

📋 Description

Generative AI systems may inadvertently or maliciously disclose proprietary information through their outputs. This includes proprietary or confidential enterprise information. Such information can be integrated into the system from training datasets or integrated retrieval systems. These disclosures can occur unintentionally through regular system interactions or be deliberately elicited by adversarial queries.

Because of the opaque and probabilistic nature of LLMs, it is often difficult to guarantee that private information has not been memorized or can’t be extracted. Even systems that implement fine-tuning or retrieval-based augmentation (RAG) risk exposing source material unless protective mechanisms are enforced throughout the training and inference pipelines.

🔍 Public Examples and Common Patterns

An organization can use proprietary data to train or fine-tune a model to address customer issues in detail. However, without proper protections the model may expose private information in a public application.

📐 External Framework Mapping

- OWASP LLM Top 10: LLM02:2025 - Sensitive Information Disclosure
- MITRE ATLAS: AML.T0057 - LLM Data Leakage
- Databricks AI Security Framework: 10.3 - Sensitive data output from a model
Cite this page
Trustible. "Leaking Proprietary Data." Trustible AI Governance Insights Center, 2026. https://trustible.ai/ai-risks/leaking-proprietary-data/

Manage AI Risk with Trustible

Trustible's AI governance platform helps enterprises identify, assess, and mitigate AI risks like this one at scale.

Explore the Platform