Presented at
DeepSec 2016 „Ten“,
Unknown date/time
(Unknown duration).
Deep learning and neural networks have gained incredible popularity in recent years. The technology has grown to be the most talked about and least well understood branch of machine learning. Successful applications of deep learning in image and speech recognition have kickstarted movements to integrate it into critical fields like medical imaging, and selfdriving cars. [1] In the security field, deep learning has shown good experimental results in malware/anomaly detection, [2] APT protection, spam/phishing detection and traffic identification. However, most deep learning systems are not designed with security and resiliency in mind, and can be duped by any attacker with a good understanding of the system. [3]
The efficacy of applications using machine learning should not only be measured with precision and recall, but also by their malleability in an adversarial setting. In this talk, we will dive into popular deep learning software and show how it can be tampered with to do what you want it do, while avoiding detection by system administrators. Besides giving a high level overview of deep learning and its inherent shortcomings in an adversarial setting, we will focus on tampering real systems to show real weaknesses in critical systems built with it. In particular, this demodriven session will be focused on manipulating an image recognition, speech recognition and phishing detection system built with deep learning at the core.
By discussing defensive measures that should be put in place to prevent the class of attacks demonstrated, we hope to address the hype behind deep learning from the context of security and look towards a more resilient future of this technology where developers can more safely use it in critical applications.
Citations:
[1] Ashlee Vance. 2015. The First Person to Hack the iPhone Built a SelfDriving Car. In His Garage.
http://www.bloomberg.com/features/2015georgehotzselfdrivingcar/
[2] Zhenlong Yuan, Yongqiang Lu, Zhaoguo Wang, and Yibo Xue. 2014. DroidSec: deep learning in android malware detection.
SIGCOMM Comput. Commun. Rev. 44, 4 (August 2014), 371372. DOI=http://dx.doi.org/10.1145/2740070.2631434
[3] Nicolas Papernot, Patrick McDaniel, Ian Goodfellow, Somesh Jha, Z. Berkay Celik, and Ananthram Swami. 2016. Practical BlackBox Attacks against Deep Learning Systems using Adversarial Examples. (arXiv:1602.02697v2)
Presenters:
-
Clarence Chio
- Shape Security
Clarence Chio graduated with a B.S. and M.S. in Computer Science from Stanford, specializing in data mining and artificial intelligence. He currently works as a Security Research Engineer at Shape Security, building a product that protects high valued web assets from automated attacks. At Shape, he works on the data analysis systems used to tackle this problem. Clarence spoke on Machine Learning and Security at DEFCON 24, PHDays, BSides Las Vegas and NYC, Code Blue, SecTor, and Hack in Paris (2015,2016). He has been a community speaker with Intel, and is also the founder and organizer of the ‘Data Mining for Cyber Security' meetup group, the largest gathering of security data scientists in the San Francisco Bay Area.
Links:
Similar Presentations: