AI Risk · Privacy

Leaking Personal Data

A generative model can reveal personal information (i.e. PII) about individuals from the training data or connected systems (e.g. in a RAG set-up).

📋 Description

Generative AI systems may inadvertently or maliciously disclose personal information through their outputs. This includes individual's personal data (e.g. PII) or information that is considered personally sensitive (e.g., health information). Such information can be integrated into the system from training datasets or integrated retrieval systems. These disclosures can occur unintentionally through regular system interactions or be deliberately elicited by adversarial queries.

Because of the opaque and probabilistic nature of LLMs, it is often difficult to guarantee that private information has not been memorized or can’t be extracted. Even systems that implement fine-tuning or retrieval-based augmentation (RAG) risk exposing source material unless protective mechanisms are enforced throughout the training and inference pipelines.

📐 External Framework Mapping

- OWASP LLM Top 10: LLM02:2025 - Sensitive Information Disclosure
- MITRE ATLAS: AML.T0057 - LLM Data Leakage
- Databricks AI Security Framework: 10.3 - Sensitive data output from a model
Cite this page
Trustible. "Leaking Personal Data." Trustible AI Governance Insights Center, 2026. https://trustible.ai/ai-risks/leaking-sensitive-data/

Manage AI Risk with Trustible

Trustible's AI governance platform helps enterprises identify, assess, and mitigate AI risks like this one at scale.

Explore the Platform