Title:
UnMask: Adversarial Detection and Defense in Deep Learning Through Building-Block Knowledge Extraction

dc.contributor.author Freitas, Scott
dc.contributor.author Chen, Shang-Tse
dc.contributor.author Chau, Duen Horng
dc.contributor.corporatename Georgia Institute of Technology. College of Computing en_US
dc.contributor.corporatename Georgia Institute of Technology. School of Computational Science and Engineering en_US
dc.date.accessioned 2019-02-14T18:14:35Z
dc.date.available 2019-02-14T18:14:35Z
dc.date.issued 2019
dc.description.abstract Deep learning models are being integrated into a wide range of high-impact, security-critical systems, from self-driving cars to biomedical diagnosis. However, recent research has demonstrated that many of these deep learning architectures are highly vulnerable to adversarial attacks—highlighting the vital need for defensive techniques to detect and mitigate these attacks before they occur. To combat these adversarial attacks, we developed UnMask, a knowledge-based adversarial detection and defense framework. The core idea behind UnMask is to protect these models by verifying that an image’s predicted class (“bird”) contains the expected building blocks (e.g., beak, wings, eyes). For example, if an image is classified as “bird”, but the extracted building blocks are wheel, seat and frame, the model may be under attack. UnMask detects such attacks and defends the model by rectifying the misclassification, re-classifying the image based on its extracted building blocks. Our extensive evaluation shows that UnMask (1) detects up to 92.9% of attacks, with a false positive rate of 9.67% and (2) defends the model by correctly classifying up to 92.24% of adversarial images produced by the current strongest attack, Projected Gradient Descent, in the gray-box setting. Our proposed method is architecture agnostic and fast. To enable reproducibility of our research, we have anonymously open-sourced our code and large newly-curated dataset (~5GB) on GitHub (https://github.com/unmaskd/UnMask). en_US
dc.identifier.uri http://hdl.handle.net/1853/60900
dc.language.iso en_US en_US
dc.publisher Georgia Institute of Technology en_US
dc.relation.ispartofseries CSE Technical Reports; GT-CSE- en_US
dc.subject Deep learning en_US
dc.subject Adversarial attack en_US
dc.subject Adversarial detection en_US
dc.subject Building-blocks en_US
dc.subject Knowledge extraction en_US
dc.title UnMask: Adversarial Detection and Defense in Deep Learning Through Building-Block Knowledge Extraction en_US
dc.type Text
dc.type.genre Technical Report
dspace.entity.type Publication
local.contributor.author Chau, Duen Horng
local.contributor.corporatename School of Computational Science and Engineering
local.relation.ispartofseries School of Computational Science and Engineering Technical Report Series
relation.isAuthorOfPublication fb5e00ae-9fb7-475d-8eac-50c48a46ea23
relation.isOrgUnitOfPublication 01ab2ef1-c6da-49c9-be98-fbd1d840d2b1
relation.isSeriesOfPublication 5a01f926-96af-453d-a75b-abc3e0f0abb3
Files
Original bundle
Now showing 1 - 1 of 1
Thumbnail Image
Name:
adversarial_detection_kdd.pdf
Size:
1.73 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
3.13 KB
Format:
Item-specific license agreed upon to submission
Description: