Data Poisoning - what is it and how it is being addressed by the leading Gen AI providers?
Creators
Description
Data poisoning is a serious threat to machine learning models, wherein malicious actors introduce corrupt input into the training data to skew model behavior, potentially leading to biased decision-making and reduced system reliability. Various types of data poisoning attacks exist, including targeted attacks, non-targeted attacks, label poisoning, training data poisoning, model inversion attacks, stealth attacks, and backdoor poisoning. Detecting and mitigating these attacks require close attention to model degradation patterns, securing training data, and employing advanced verification methods. Major AI companies such as OpenAI, Microsoft, Google, and Meta have developed protective mechanisms against data poisoning, providing valuable guidance for organizations leveraging AI technologies. Best practices for reducing data poisoning risks include data validation and sanitization, red teaming, secure data handling, negative testing, and benchmark testing. Collaboration among developers, MLOps communities, and security teams is crucial for robust AI system construction, requiring diligent efforts in data integrity assurance, cross-functional communication, education, and continuous improvement of testing and validation layers. Emphasis on strong defense mechanisms and ongoing innovation will support the growth and safe application of AI across diverse industries. The purpose of this article is to delve into these topics in depth and offer guidance for individuals and organizations working with data for machine learning.
Files
EJAET-11-5-105-109.pdf
Files
(206.3 kB)
Name | Size | Download all |
---|---|---|
md5:982119583cb65b994c911a0f3a76530d
|
206.3 kB | Preview Download |
Additional details
References
- [1]. Amos, Z. (2023, December 15). Cybersecurity Measures to Prevent Data Poisoning. Open Data Science - Your News Source for AI, Machine Learning & More. https://opendatascience.com/cybersecurity-measures-to-prevent-data-poisoning/
- [2]. Brito, R. (2024, May 8). How CSPs and enterprises can safeguard against data poisoning of LLMs. TechRadar; TechRadar pro. https://www.techradar.com/pro/how-csps-and-enterprises-can-safeguard-against-data-poisoning-of-llms
- [3]. D'Alessandro, M. A. (2024, April 25). Data Poisoning attacks on Enterprise LLM applications: AI risks, detection, and prevention. Giskard.ai; Giskard. https://www.giskard.ai/knowledge/data-poisoning-attacks-on-enterprise-llm-applications-ai-risks-detection-and-prevention
- [4]. Dhar, P. (2023, March 24). Protecting AI Models from "Data Poisoning." IEEE Spectrum; IEEE Spectrum. https://spectrum.ieee.org/ai-cybersecurity-data-poisoning
- [5]. Heikkilä, M. (2023, October 23). This new data poisoning tool lets artists fight back against generative AI. MIT Technology Review; MIT Technology Review. https://www.technologyreview.com/2023/10/23/1082189/data-poisoning-artists-fight-generative-ai/
- [6]. Koh, P. W., Steinhardt, J., & Liang, P. (2022). Stronger data poisoning attacks break data sanitization defenses. Machine Learning, 1-47.
- [7]. Obadiaru, A. (2023, July 26). Data Poisoning Attacks: A New Attack Vector within AI | Cobalt. Cobalt.io; Cobalt. https://www.cobalt.io/blog/data-poisoning-attacks-a-new-attack-vector-within-ai
- [8]. Russinovich, M. (2024, April 11). How Microsoft discovers and mitigates evolving attacks against AI guardrails | Microsoft Security Blog. Microsoft Security Blog. https://www.microsoft.com/en-us/security/blog/2024/04/11/how-microsoft-discovers-and-mitigates-evolving-attacks-against-ai-guardrails/
- [9]. Sangavi Senthil. (2024, January 30). Data poisoning: Prevention strategies to keep your data safe - ManageEngine Blog. ManageEngine Blog. https://blogs.manageengine.com/active-directory/log360/2024/01/30/data-poisoning-prevention-strategies-to-keep-your-data-safe.html
- [10]. Zhu, Y., Wen, H., Zhao, R., Jiang, Y., Liu, Q., & Zhang, P. (2023). Research on Data Poisoning Attack against Smart Grid Cyber–Physical System Based on Edge Computing. Sensors, 23(9), 4509.