Definition
Data Poisoning in Machine Learning (ML) refers to attacks carried out manipulating training data to alter the learning process and eventually impacting on ML models’ inference.
Background
Learning from data allows ML models to solve a wide variety of problems with unprecedented performance. However, learning from data also implies that any issue affecting data may potentially influence the learning process and eventually the ML model performance. Attackers may try to exploit the learning process vulnerabilities to alter the ML model training outcome. More specifically, in a Data Poisoning attack, training data manipulation is considered. Data Poisoning highlights the tight relationship between ML security, data governance, and protection: a connection that is not limited to the technical side but that also reverberates, as an example, in the European Union Artificial Intelligence Act (or AI...
References
Barreno M, Nelson B, Sears R, Joseph AD, Tygar JD (2006) Can machine learning be secure? ASIACCS '06: Proceedings of the 2006 ACM Symposium on Information, computer and communications security, March 2006. https://doi.org/10.1145/1128817.1128824
Barreno M, Nelson B, Joseph AD et al (2010) The security of machine learning. Mach Learn 81:121–148. https://doi.org/10.1007/s10994-010-5188-5
Biggio B, Roli F (2018) Wild patterns: ten years after the rise of adversarial machine learning. Pattern Recogn 84:317–331., issn: 0031-3203. https://doi.org/10.1016/j.patcog.2018.07.023
Dalvi N, Domingos P, Mausam, Sanghai S, Verma D (2004) Adversarial classification”, KDD '04: Proceedings of the tenth ACM SIGKDD international conference on Knowledge discovery and data mining, pp 99–108. https://doi.org/10.1145/1014052.1014066
European Commission (2021) Proposal for a regulation (artificial intelligence act) (COM(2021) 206 final). https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:52021PC0206
European Parliament and Council (2016) Regulation (EU) 2016/679. https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX%3A32016R0679
Gu T, Liu K, Dolan-Gavitt B, Garg S (2019) BadNets: evaluating backdooring attacks on deep neural networks. IEEE Access 7:47230–47244. https://doi.org/10.1109/ACCESS.2019.2909068
Yerlikaya FA, Bahtiyar Ş (2022) Data poisoning attacks against machine learning algorithms. Expert Systems with Applications 208:118101., issn: 0957-4174. https://doi.org/10.1016/j.eswa.2022.118101
Zhang H, Cheng N, Zhang Y et al (2021) Label flipping attacks against Naive Bayes on spam filtering systems. Appl Intell 51:4503–4514. https://doi.org/10.1007/s10489-020-02086-4
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Section Editor information
Rights and permissions
Copyright information
© 2024 Springer Science+Business Media LLC
About this entry
Cite this entry
Barezzani, S. (2024). Data Poisoning Attacks in Machine Learning. In: Jajodia, S., Samarati, P., Yung, M. (eds) Encyclopedia of Cryptography, Security and Privacy. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-27739-9_1824-1
Download citation
DOI: https://doi.org/10.1007/978-3-642-27739-9_1824-1
Received:
Accepted:
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-27739-9
Online ISBN: 978-3-642-27739-9
eBook Packages: Springer Reference Computer SciencesReference Module Computer Science and Engineering