Data Poisoning
Data poisoning is a cyberattack in which an adversary intentionally corrupts or manipulates the data used to train an AI or machine learning model. By injecting malicious, biased, or misleading data into the training pipeline, the attacker causes the resulting model to behave in unintended or harmful ways. The effects typically persist silently within the model after training completes, making them difficult to detect without careful auditing of training data and model outputs.
Data poisoning is an adversarial attack targeting the integrity of machine learning pipelines by compromising datasets used during pre-training, fine-tuning, or embedding stages. An attacker who gains the ability to insert, modify, or remove training samples may introduce backdoors, degrade model accuracy, or embed systematic biases into learned model weights. The attack surface spans data collection, labeling pipelines, third-party dataset sourcing, and model customization workflows. Poisoning may be targeted, designed to cause misclassification of specific inputs, or indiscriminate, aimed at broad performance degradation. Because the manipulation occurs prior to or during training rather than at inference time, static analysis of the trained model artifact alone is typically insufficient to detect the presence of poisoned behavior; detection generally requires dataset provenance controls, training data audits, and runtime behavioral evaluation against known-clean reference outputs.
Why it matters
Data poisoning is significant because the corruption it introduces is embedded in a model's learned weights during training, meaning the resulting behavior persists silently through deployment. Unlike many attack types that target running systems, data poisoning may succeed long before the affected model is ever put into production, and the malicious influence typically remains invisible during standard model evaluation unless specific auditing controls are in place. Organizations that deploy AI models without validating the integrity of their training data may be operating compromised systems without any immediate indication of a problem.
Who it's relevant to
Inside Data Poisoning
Common questions
Answers to the questions practitioners most commonly ask about Data Poisoning.