Data Poisoning
- louisgoh8
- May 29
- 1 min read
Is AI even capable of providing the wrong kind of information? It seems unlikely, but it can happen when data poisoning occurs.
Data Poisoning is a cyberattack that corrupts the training data used to build AI models. These models are very dependent on the quality and integrity of its training data. Sourced from various places like the internet or databases, accurate and false information would overlap, making it vulnerable to manipulation by malicious actors. This would drastically alter the model’s behaviour.
Data poisoning attacks can be classified into two categories: targeted and nontargeted. Targeted attacks manipulate AI in specific ways while non targeted attacks aim to weaken the model’s ability to process data entirely.
How to stop them?
Data Validation and Sanitization
Removes corrupted data before it can compromise the model.
Adversarial Training and Improved Robustness
Strengthens AI models by exposing them to adversarial examples during development.
Continuous Monitoring
Identifying unusual behaviours or discrepancies, enabling quick responses to any threat.
Access Controls
Restricting unauthorized modifications to training data and using encryption to protect data sources.
Comentários