

{"id":218,"date":"2017-04-18T09:28:20","date_gmt":"2017-04-18T07:28:20","guid":{"rendered":"https:\/\/project.inria.fr\/wifs2017\/?page_id=218"},"modified":"2017-12-03T10:23:25","modified_gmt":"2017-12-03T09:23:25","slug":"tutorial-on-security-and-privacy-in-machine-learning","status":"publish","type":"page","link":"https:\/\/project.inria.fr\/wifs2017\/program\/tutorials\/tutorial-on-security-and-privacy-in-machine-learning\/","title":{"rendered":"Tutorial on Security and Privacy in Machine Learning"},"content":{"rendered":"<p><strong><a href=\"https:\/\/project.inria.fr\/wifs2017\/files\/2017\/12\/WIFS_T2_Papernot.pdf\" target=\"_blank\" rel=\"noopener\">Copy of the slides<\/a> (draft)\u00a0<\/strong><\/p>\n<p>&nbsp;<\/p>\n<p><strong>Abstract: <\/strong><\/p>\n<p><span style=\"font-weight: 400;\">There is growing recognition that machine learning exposes new security and privacy issues in software systems. In this tutorial, we first articulate a comprehensive threat model for machine learning, then present an attack against model prediction integrity, and finally discuss a framework for learning privately.<\/span><\/p>\n<p>Machine learning models were shown to be vulnerable to adversarial examples&#8211;subtly modified malicious inputs crafted to compromise the integrity of their outputs. Furthermore, adversarial examples that affect one model often affect another model, even if the two models have different architectures, so long as both models were trained to perform the same task. An attacker may therefore conduct an attack with very little information about the victim by training their own substitute model to craft adversarial examples, and then transferring them to a victim model. The attacker need not even collect a training set to mount the attack. Indeed, we demonstrate how adversaries may use the victim model as an oracle to label a synthetic training set for the substitute. We conclude this first part of the tutorial by formally showing that there are (possibly unavoidable) tensions between model complexity, accuracy, and resilience that must be calibrated for the environments in which they will be used.<\/p>\n<p>In addition, some machine learning applications involve training data that is sensitive, such as the medical histories of patients in a clinical trial. A model may inadvertently and implicitly store some of its training data; careful analysis of the model may therefore reveal sensitive information. To address this problem, we demonstrate a generally applicable approach to providing strong privacy guarantees for training data. The approach combines, in a black-box fashion, multiple models trained with disjoint datasets, such as records from different subsets of users. Because they rely directly on sensitive data, these models are not published, but instead used as &#8220;teachers&#8221; for a &#8220;student&#8221; model. The student learns to predict an output chosen by noisy voting among all of the teachers, and cannot directly access an individual teacher or the underlying data or parameters. The student&#8217;s privacy properties can be understood both intuitively (since no single teacher and thus no single dataset dictates the student&#8217;s training) and formally, in terms of differential privacy.<\/p>\n<p><strong>Topics include:<\/strong><\/p>\n<ul>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">An introduction to machine learning<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">A taxonomy of threat models for security and privacy in machine learning<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">Attacks using adversarial examples against vision systems, malware detection, and reinforcement learning agents<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">Black-box attacks against machine learning<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">Adversarial example transferability<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">Defending machine learning with adversarial training and defensive distillation<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">Open problems in defenses such as gradient masking<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">No-free lunch theorem for adversarial machine learning<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">Short tutorial on cleverhans (an open-source library for adversarial machine learning)<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">Differential privacy<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">Privacy-preserving machine learning with the PATE framework<\/span><\/li>\n<\/ul>\n<p><strong>Learning objectives:<\/strong><\/p>\n<ul>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">To explain the fundamentals of security and privacy in machine learning<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">To bring the audience up-to-date with the state-of-the-art attack techniques<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">To make the audience aware of the open problems in defense strategies and as a consequence the risks associated with deploying machine learning in security or privacy sensitive settings<\/span><\/li>\n<li style=\"font-weight: 400;\"><span style=\"font-weight: 400;\">To prepare the audience to make original contributions in this area<\/span><\/li>\n<\/ul>\n<p><strong>Target audience:<\/strong><\/p>\n<p><span style=\"font-weight: 400;\">The target audience is people from the security and privacy community who are interested in (a) deploying machine learning to security problems or (b) making machine learning more secure and private.<\/span><\/p>\n<p><strong>Bio:<\/strong><\/p>\n<p><span style=\"font-weight: 400;\"><a href=\"https:\/\/www.papernot.fr\">Nicolas Papernot<\/a> is a PhD student in Computer Science and Engineering working with Dr. Patrick McDaniel at the Pennsylvania State University. His research interests lie at the intersection of computer security, privacy and machine learning. He is supported by a Google PhD Fellowship in Security. He received a best paper award at ICLR 2017. Nicolas is the co-author of cleverhans, an open-source library for benchmarking the vulnerability of machine learning models. In 2016, he received his M.S. in Computer Science and Engineering from the Pennsylvania State University and his M.S. in Engineering Sciences from the Ecole Centrale de Lyon.<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Copy of the slides (draft)\u00a0 &nbsp; Abstract: There is growing recognition that machine learning exposes new security and privacy issues in software systems. In this tutorial, we first articulate a comprehensive threat model for machine learning, then present an attack against model prediction integrity, and finally discuss a framework for\u2026<\/p>\n<p> <a class=\"continue-reading-link\" href=\"https:\/\/project.inria.fr\/wifs2017\/program\/tutorials\/tutorial-on-security-and-privacy-in-machine-learning\/\"><span>Continue reading<\/span><i class=\"crycon-right-dir\"><\/i><\/a> <\/p>\n","protected":false},"author":1161,"featured_media":0,"parent":319,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"class_list":["post-218","page","type-page","status-publish","hentry"],"_links":{"self":[{"href":"https:\/\/project.inria.fr\/wifs2017\/wp-json\/wp\/v2\/pages\/218","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/project.inria.fr\/wifs2017\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/project.inria.fr\/wifs2017\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/project.inria.fr\/wifs2017\/wp-json\/wp\/v2\/users\/1161"}],"replies":[{"embeddable":true,"href":"https:\/\/project.inria.fr\/wifs2017\/wp-json\/wp\/v2\/comments?post=218"}],"version-history":[{"count":3,"href":"https:\/\/project.inria.fr\/wifs2017\/wp-json\/wp\/v2\/pages\/218\/revisions"}],"predecessor-version":[{"id":620,"href":"https:\/\/project.inria.fr\/wifs2017\/wp-json\/wp\/v2\/pages\/218\/revisions\/620"}],"up":[{"embeddable":true,"href":"https:\/\/project.inria.fr\/wifs2017\/wp-json\/wp\/v2\/pages\/319"}],"wp:attachment":[{"href":"https:\/\/project.inria.fr\/wifs2017\/wp-json\/wp\/v2\/media?parent=218"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}