Shang-Tse Chen’s research while affiliated with Georgia Institute of Technology and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (20)


UnMask: Adversarial Detection and Defense Through Robust Feature Alignment
  • Conference Paper

December 2020

·

25 Reads

·

23 Citations

Scott Freitas

·

Shang-Tse Chen

·

Zijie J. Wang

·


UnMask: Adversarial Detection and Defense Through Robust Feature Alignment

February 2020

·

53 Reads

Deep learning models are being integrated into a wide range of high-impact, security-critical systems, from self-driving cars to medical diagnosis. However, recent research has demonstrated that many of these deep learning architectures are vulnerable to adversarial attacks--highlighting the vital need for defensive techniques to detect and mitigate these attacks before they occur. To combat these adversarial attacks, we developed UnMask, an adversarial detection and defense framework based on robust feature alignment. The core idea behind UnMask is to protect these models by verifying that an image's predicted class ("bird") contains the expected robust features (e.g., beak, wings, eyes). For example, if an image is classified as "bird", but the extracted features are wheel, saddle and frame, the model may be under attack. UnMask detects such attacks and defends the model by rectifying the misclassification, re-classifying the image based on its robust features. Our extensive evaluation shows that UnMask (1) detects up to 96.75% of attacks, with a false positive rate of 9.66% and (2) defends the model by correctly classifying up to 93% of adversarial images produced by the current strongest attack, Projected Gradient Descent, in the gray-box setting. UnMask provides significantly better protection than adversarial training across 8 attack vectors, averaging 31.18% higher accuracy. Our proposed method is architecture agnostic and fast. We open source the code repository and data with this paper: https://github.com/unmaskd/unmask.


Talk Proposal: Towards the Realistic Evaluation of Evasion Attacks using CARLA

April 2019

·

9 Reads

In this talk we describe our content-preserving attack on object detectors, ShapeShifter, and demonstrate how to evaluate this threat in realistic scenarios. We describe how we use CARLA, a realistic urban driving simulator, to create these scenarios, and how we use ShapeShifter to generate content-preserving attacks against those scenarios.


ADAGIO: Interactive Experimentation with Adversarial Attack and Defense for Audio: European Conference, ECML PKDD 2018, Dublin, Ireland, September 10–14, 2018, Proceedings, Part III

January 2019

·

23 Reads

·

13 Citations

Lecture Notes in Computer Science

·

Madhuri Shanbhogue

·

Shang-Tse Chen

·

[...]

·

Adversarial machine learning research has recently demonstrated the feasibility to confuse automatic speech recognition (ASR) models by introducing acoustically imperceptible perturbations to audio samples. To help researchers and practitioners gain better understanding of the impact of such attacks, and to provide them with tools to help them more easily evaluate and craft strong defenses for their models, we present Adagio, the first tool designed to allow interactive experimentation with adversarial attacks and defenses on an ASR model in real time, both visually and aurally. Adagio incorporates AMR and MP3 audio compression techniques as defenses, which users can interactively apply to attacked audio samples. We show that these techniques, which are based on psychoacoustic principles, effectively eliminate targeted attacks, reducing the attack success rate from 92.5% to 0%. We will demonstrate Adagio and invite the audience to try it on the Mozilla Common Voice dataset. Code related to this paper is available at: https://github.com/nilakshdas/ADAGIO.


ShapeShifter: Robust Physical Adversarial Attack on Faster R-CNN Object Detector: Recognizing Outstanding Ph.D. Research

January 2019

·

152 Reads

·

167 Citations

Lecture Notes in Computer Science

Given the ability to directly manipulate image pixels in the digital input space, an adversary can easily generate imperceptible perturbations to fool a Deep Neural Network (DNN) image classifier, as demonstrated in prior work. In this work, we propose ShapeShifter, an attack that tackles the more challenging problem of crafting physical adversarial perturbations to fool image-based object detectors like Faster R-CNN. Attacking an object detector is more difficult than attacking an image classifier, as it needs to mislead the classification results in multiple bounding boxes with different scales. Extending the digital attack to the physical world adds another layer of difficulty, because it requires the perturbation to be robust enough to survive real-world distortions due to different viewing distances and angles, lighting conditions, and camera limitations. We show that the Expectation over Transformation technique, which was originally proposed to enhance the robustness of adversarial perturbations in image classification, can be successfully adapted to the object detection setting. ShapeShifter can generate adversarially perturbed stop signs that are consistently mis-detected by Faster R-CNN as other objects, posing a potential threat to autonomous vehicles and other safety-critical computer vision systems. Code related to this paper is available at: https://github.com/shangtse/robust-physical-attack.


Figure 2: Shield uses Stochastic Local Quantization (SLQ) to remove adversarial perturbations from input images. SLQ divides an image into 8 ? 8 blocks and applies a randomly selected JPEG compression quality (20, 40, 60 or 80) to each block to mitigate the attack.
Figure 6: Vaccinating a model by retraining it with compressed images helps recover its accuracy. Each plot shows the model accuracies when preprocessing with different JPEG qualities with the FGSM attack. Each curve in the plot corresponds to a different model. The gray dotted curve corresponds to the original unvaccinated ResNet-v2 50 model. The orange and purple curves correspond to the models retrained on JPEG qualities 80 and 20 respectively. Retraining on JPEG compressed images and applying JPEG preprocessing helps recover accuracy in a gray-box attack.
SHIELD: Fast, Practical Defense and Vaccination for Deep Learning using JPEG Compression
  • Conference Paper
  • Full-text available

July 2018

·

240 Reads

·

195 Citations

The rapidly growing body of research in adversarial machine learning has demonstrated that deep neural networks (DNNs) are highly vulnerable to adversarially generated images. This underscores the urgent need for practical defense techniques that can be readily deployed to combat attacks in real-time. Observing that many attack strategies aim to perturb image pixels in ways that are visually imperceptible, we place JPEG compression at the core of our proposed SHIELD defense framework, utilizing its capability to effectively "compress away" such pixel manipulation. To immunize a DNN model from artifacts introduced by compression, SHIELD "vaccinates" the model by retraining it with compressed images, where different compression levels are applied to generate multiple vaccinated models that are ultimately used together in an ensemble defense. On top of that, SHIELD adds an additional layer of protection by employing randomization at test time that compresses different regions of an image using random compression levels, making it harder for an adversary to estimate the transformation performed. This novel combination of vaccination, ensembling, and randomization makes SHIELD a fortified multi-pronged defense. We conducted extensive, large-scale experiments using the ImageNet dataset, and show that our approaches eliminate up to 98% of gray-box attacks delivered by strong adversarial techniques such as Carlini-Wagner's L2 attack and DeepFool. Our approaches are fast and work without requiring knowledge about the model.

Download

Fig. 1. Adagio usage scenario. (1) Jane uploads an audio file that is transcribed by DeepSpeech; then she performs an adversarial attack on the audio in real time by entering a target transcription after selecting the attack option from the dropdown menu. (2) Jane decides to perturb the audio to change the last word of the sentence from "joanna" to "marissa"; she can listen to the original audio and see the transcription by clicking on the "Original" badge. (3) Jane applies MP3 compression to recover the original, correct transcription from the manipulated audio; clicking on a waveform plays back the audio from the selected position. (4) Jane can experiment with multiple audio samples by adding more cards. For ease of presentation, operations 1, 2 and 3 are shown as separate cards. 
Table 1 . Word Error Rate (WER) and the targeted attack success rate on the Deep- Speech model (lower is better for both). AMR and MP3 eliminate all targeted attacks, and significantly improves WER.
ADAGIO: Interactive Experimentation with Adversarial Attack and Defense for Audio

May 2018

·

122 Reads

Adversarial machine learning research has recently demonstrated the feasibility to confuse automatic speech recognition (ASR) models by introducing acoustically imperceptible perturbations to audio samples. To help researchers and practitioners gain better understanding of the impact of such attacks, and to provide them with tools to help them more easily evaluate and craft strong defenses for their models, we present ADAGIO, the first tool designed to allow interactive experimentation with adversarial attacks and defenses on an ASR model in real time, both visually and aurally. ADAGIO incorporates AMR and MP3 audio compression techniques as defenses, which users can interactively apply to attacked audio samples. We show that these techniques, which are based on psychoacoustic principles, effectively eliminate targeted attacks, reducing the attack success rate from 92.5% to 0%. We will demonstrate ADAGIO and invite the audience to try it on the Mozilla Common Voice dataset.


Fig. 2: Digital perturbations we created use our method. Low confidence perturbations on the top and high confidence perturbations on the bottom. 
Table 2 : As expected, low-confidence perturbations achieve lower success rates.
Fig. 3: Indoor experiment setup. We take photos of the printed adversarial sign, from multiple angles (0 • , 15 • , 30 • , 45 • , 60 • , from the sign's tangent), and distances (5' to 40'). The camera locations are indicated by the red dots, and the camera always points at the sign. 
Table 3 : Sample high-confidence perturbations from indoor experiments. For complete experiment results, please refer to Table 1.
- tances and angles. For each distance-angle combination, we show the detected class and the confidence score. If more than one bounding boxes are detected, we report the highest-scoring one. Confidence values lower than 30% is considered undetected.
Robust Physical Adversarial Attack on Faster R-CNN Object Detector

April 2018

·

659 Reads

·

216 Citations

Given the ability to directly manipulate image pixels in the digital input space, an adversary can easily generate imperceptible perturbations to fool a Deep Neural Network (DNN) image classifier, as demonstrated in prior work. In this work, we tackle the more challenging problem of crafting physical adversarial perturbations to fool image-based object detectors like Faster R-CNN. Attacking an object detector is more difficult than attacking an image classifier, as it needs to mislead the classification results in multiple bounding boxes with different scales. Extending the digital attack to the physical world adds another layer of difficulty, because it requires the perturbation to be robust enough to survive real-world distortions due to different viewing distances and angles, lighting conditions, and camera limitations. We show that the Expectation over Transformation technique, which was originally proposed to enhance the robustness of adversarial perturbations in image classification, can be successfully adapted to the object detection setting. Our approach can generate adversarially perturbed stop signs that are consistently mis-detected by Faster R-CNN as other objects, posing a potential threat to autonomous vehicles and other safety-critical computer vision systems.


Figure 2: Shield uses Stochastic Local Quantization (SLQ) to remove adversarial perturbations from input images. Shield divides images into 8 × 8 blocks and applies a randomly selected JPEG compression quality (20, 40, 60 or 80) to each block to remove adversarial attacks. Note this figure is an illustration; our images are of actual size 299 × 299. 
Figure 5: Runtime comparison for three defenses: (1) total variation denoising (TVD), (2) median filter (MF), and (3) JPEG compression, timed using the full 50k ImageNet validation images, averaged over 3 runs. JPEG is at least 22x faster than TVD, and 14x faster than MF. (Window size of 3 is the smallest possible for median filter.) 
Figure 6: Vaccinating a model by retraining it with compressed images helps recover its accuracy. Each plot shows the model accuracies when preprocessing with different JPEG qualities. Each curve in the plot corresponds to a different model. The gray dotted curve corresponds to the original unvaccinated ResNet-v2 50 model. The orange and purple curves correspond to the models retrained on JPEG qualities 80 and 20 respectively. Retraining on JPEG compressed images and applying JPEG preprocessing helps recover accuracy in a gray-box attack. 
Shield: Fast, Practical Defense and Vaccination for Deep Learning using JPEG Compression

February 2018

·

209 Reads

·

58 Citations

The rapidly growing body of research in adversarial machine learning has demonstrated that deep neural networks (DNNs) are highly vulnerable to adversarially generated images. This underscores the urgent need for practical defense that can be readily deployed to combat attacks in real-time. Observing that many attack strategies aim to perturb image pixels in ways that are visually imperceptible, we place JPEG compression at the core of our proposed Shield defense framework, utilizing its capability to effectively "compress away" such pixel manipulation. To immunize a DNN model from artifacts introduced by compression, Shield "vaccinates" a model by re-training it with compressed images, where different compression levels are applied to generate multiple vaccinated models that are ultimately used together in an ensemble defense. On top of that, Shield adds an additional layer of protection by employing randomization at test time that compresses different regions of an image using random compression levels, making it harder for an adversary to estimate the transformation performed. This novel combination of vaccination, ensembling, and randomization makes Shield a fortified multi-pronged protection. We conducted extensive, large-scale experiments using the ImageNet dataset, and show that our approaches eliminate up to 94% of black-box attacks and 98% of gray-box attacks delivered by the recent, strongest attacks, such as Carlini-Wagner's L2 and DeepFool. Our approaches are fast and work without requiring knowledge about the model.


Predicting Cyber Threats with Virtual Security Products

December 2017

·

122 Reads

·

8 Citations

Cybersecurity analysts are often presented suspicious machine activity that does not conclusively indicate compromise, resulting in undetected incidents or costly investigations into the most appropriate remediation actions. There are many reasons for this: deficiencies in the number and quality of security products that are deployed, poor configuration of those security products, and incomplete reporting of product-security telemetry. Managed Security Service Providers (MSSP's), which are tasked with detecting security incidents on behalf of multiple customers, are confronted with these data quality issues, but also possess a wealth of cross-product security data that enables innovative solutions. We use MSSP data to develop Virtual Product, which addresses the aforementioned data challenges by predicting what security events would have been triggered by a security product if it had been present. This benefits the analysts by providing more context into existing security incidents (albeit probabilistic) and by making questionable security incidents more conclusive. We achieve up to 99% AUC in predicting the incidents that some products would have detected had they been present.


Citations (13)


... If the difference between classifier outputs exceeds a threshold, the input is identified as an adversarial example. Freitas et al [27]. proposed UnMask, which extracts FIGURE 1. Process of the cross-modal semantic embedding-based adversarial example detection method. ...

Reference:

Detecting Adversarial Examples Using Cross-Modal Semantic Embeddings From Images and Text
UnMask: Adversarial Detection and Defense Through Robust Feature Alignment
  • Citing Conference Paper
  • December 2020

... Adversarial examples can be categorized into digital and physical adversarial types. Digital adversarial examples involve introducing pixel-level perturbations to the model's digital input, while physical ones manipulate real-world objects or surroundings, indirectly influencing the model inputs [8]- [10]. ...

ShapeShifter: Robust Physical Adversarial Attack on Faster R-CNN Object Detector: Recognizing Outstanding Ph.D. Research
  • Citing Chapter
  • January 2019

Lecture Notes in Computer Science

... The robustness of attacks is also evaluated against defense methods, including purification methods of NRP [4], DS [52], diffusion-based purification [5] and adversarial training robust models of Inc-V3 [8], Res-50 [9], Swin-B [53], ConvNeXt-B [53]. Two classic image transformation defenses of JPEG compression [54], Bit-depth reduction [55], and another type of defense, random smoothing [11], are also included. Considering the robustness and transferability of attacks are comparable only under close perturbation budget, the unrestricted attacks are not included in this and the next section. ...

SHIELD: Fast, Practical Defense and Vaccination for Deep Learning using JPEG Compression

... We assume the pedestrian is able to obtain footage of the observed scene, and knows the object detector used in the surveillance system. These assumptions are usually considered in adversarial machine learning research [4,6,9,11,[18][19][20][21][22][23]. We also assume that the pedestrian has access to footage recorded by the pedestrian detection system (e.g., access to a live stream on the Internet). ...

Robust Physical Adversarial Attack on Faster R-CNN Object Detector

... Another idea is data recovery [36], where the perturbation patterns are reduced and the resulting recovered image is fed into a regular deep model. Here, by separating clean and perturbed data in the image, one can either directly remove the perturbed data (i.e., data compression) [37] or introduce randomness in the perturbed data to destroy their patterns (i.e., data randomization) [38]. Since discriminatively separating perturbed data and accurately reconstructing clean images are still open problems, this class of methods obviously faces a tricky accuracy-robustness tradeoff [19]. ...

Shield: Fast, Practical Defense and Vaccination for Deep Learning using JPEG Compression

... Some approaches focus on better separating the true positives from the false positives, typically by estimating anomaly or maliciousness scores for each alert [13], [19], [21], [22], [26]. Other approaches support the triage process by extracting and enriching the most relevant information of an alert [8], [39], [40]. ...

Predicting Cyber Threats with Virtual Security Products

... During the procedure of these studies, there is a multitude of irregularities that can occur, such as missing engagement of participants (e.g., participant forgets to fill out a daily questionnaire) or malfunction of measuring instruments (e.g., empty battery or Bluetooth connectivity issues). Yet, the complex nature of the data and the multitude and unpredictability of potential problem sources make an automated detection of irregularities difficult (e.g., from lost or broken hardware that needs to be replaced to software-related connectivity issues that can be resolved by rebooting), especially as each study is tailored to answer their individual research questions [53]. Researchers, therefore, have to stay in contact with their participants and regularly monitor the collected high-dimensional data streams during run time of the study to ensure data quality and prevent data loss commonly associated with mHealth data [72]. ...

Exploratory Visual Analytics of Mobile Health Data: Sensemaking Challenges and Opportunities
  • Citing Chapter
  • July 2017

... Modern manufacturing operations are controlled by various actuators that execute commands from programmable controllers [155], where sensor data is processed and analyzed using different algorithms for decision-making [99]. The signal an actuator perceives can also be changed, causing it to implement wrong decisions [156,157]. For example, in a beverage production line, sensors determine the level of filling in bottles, and a robotic end-effector pushes overfilled and underfilled bottles away from the conveyor belt. ...

Keeping the Bad Guys Out: Protecting and Vaccinating Deep Learning with JPEG Compression

... These sources encompass a diverse range of information, often collected from various origins, to create comprehensive datasets for training and evaluating AI algorithms. Some datasets from recent research that can be cited include: 3. [17] amalgamates data from a diverse array of sources to fuel its analysis. The dataset encompasses 2543 historical fire incidents generously provided by the City of Atlanta Fire Rescue Department (AFRD), 32488 fire inspections, structural information concerning commercial properties procured from the CoStar Group, parcel data from Atlanta's Office of Buildings, and business license records obtained from the City of Atlanta's Office of Revenue, socioeconomic and demographic data from the U.S. Census Bureau, liquor license records, 2014 crime data sourced from the Atlanta Police Department, and Certificate of Occupancy (CO) data from the Atlanta Office of Buildings. ...

Firebird: Predicting Fire Risk and Prioritizing Fire Inspections in Atlanta

... The hope is that, through iterations of the cycle, the performance, believability, and interestingness/complexity of the AI will improve. Within human-in-the-loop systems, visualizations are specifically leveraged to aid readability and interpretability [6,13,23]. Teso et al. [18] demonstrated the benefits of this set up, where explanations led to more interpretable outputs and helped users provide more effective feedback to a model, ultimately improving its performance. Such an approach for behavior-language authored agents may help resolve the issues surrounding interpreting and debugging an agent's behavior, but, to the authors' knowledge, no such system currently exists. ...

TimeStitch: Interactive multi-focus cohort discovery and comparison
  • Citing Conference Paper
  • October 2015