AI Risk · Security

Denial of ML Service

Adversaries can overwhelm an AI system with a large requests resulting in degraded performance or system downtime.

📋 Description

Denial of ML Service occurs when users with bad intentions intentionally overload an AI model with excessive or computationally intensive queries, disrupting functionality for actual users. Unlike traditional Denial-of-Service (DDoS) attacks that flood web servers, these attacks target the model’s inference pipeline, often exploiting high-cost operations like large prompts or complex API chains in LLM applications.

Attacks can be:

- Volume-based, involving a high number of low-cost queries.

- Compute-based, involving fewer, resource-heavy queries designed to exhaust memory, CPU, or the GPU.

🔍 Public Examples and Common Patterns

- OpenAI Rate-Limiting: Developers reported API slowdowns during peak usage windows, sometimes triggered by misuse or high-frequency requests.
- AIID Incident 1001: LLM Scrapers Allegedly Target Multiple Open Source Projects Disrupting the FOSS Ecosystem: In March 2025, KDE’s GitLab infrastructure was disrupted by aggressive AI web scrapers. These bots spoofed browser headers, which overwhelmed the site and caused outages for developers.

📐 External Framework Mapping

- OWASP LLM Top 10: LLM04 – Denial of Service
- MITRE ATLAS: AML.T0029 – Denial of ML Service
- IBM Risk Atlas: System Availability Risk
- Databricks AI Security Framework: 9.3 - Denial of service (DOS)
Cite this page
Trustible. "Denial of ML Service." Trustible AI Governance Insights Center, 2026. https://trustible.ai/ai-risks/denial-of-ml-service/

Manage AI Risk with Trustible

Trustible's AI governance platform helps enterprises identify, assess, and mitigate AI risks like this one at scale.

Explore the Platform