December 2020
·
25 Reads
·
23 Citations
This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.
December 2020
·
25 Reads
·
23 Citations
February 2020
·
53 Reads
Deep learning models are being integrated into a wide range of high-impact, security-critical systems, from self-driving cars to medical diagnosis. However, recent research has demonstrated that many of these deep learning architectures are vulnerable to adversarial attacks--highlighting the vital need for defensive techniques to detect and mitigate these attacks before they occur. To combat these adversarial attacks, we developed UnMask, an adversarial detection and defense framework based on robust feature alignment. The core idea behind UnMask is to protect these models by verifying that an image's predicted class ("bird") contains the expected robust features (e.g., beak, wings, eyes). For example, if an image is classified as "bird", but the extracted features are wheel, saddle and frame, the model may be under attack. UnMask detects such attacks and defends the model by rectifying the misclassification, re-classifying the image based on its robust features. Our extensive evaluation shows that UnMask (1) detects up to 96.75% of attacks, with a false positive rate of 9.66% and (2) defends the model by correctly classifying up to 93% of adversarial images produced by the current strongest attack, Projected Gradient Descent, in the gray-box setting. UnMask provides significantly better protection than adversarial training across 8 attack vectors, averaging 31.18% higher accuracy. Our proposed method is architecture agnostic and fast. We open source the code repository and data with this paper: https://github.com/unmaskd/unmask.
April 2019
·
9 Reads
In this talk we describe our content-preserving attack on object detectors, ShapeShifter, and demonstrate how to evaluate this threat in realistic scenarios. We describe how we use CARLA, a realistic urban driving simulator, to create these scenarios, and how we use ShapeShifter to generate content-preserving attacks against those scenarios.
January 2019
·
23 Reads
·
13 Citations
Lecture Notes in Computer Science
Adversarial machine learning research has recently demonstrated the feasibility to confuse automatic speech recognition (ASR) models by introducing acoustically imperceptible perturbations to audio samples. To help researchers and practitioners gain better understanding of the impact of such attacks, and to provide them with tools to help them more easily evaluate and craft strong defenses for their models, we present Adagio, the first tool designed to allow interactive experimentation with adversarial attacks and defenses on an ASR model in real time, both visually and aurally. Adagio incorporates AMR and MP3 audio compression techniques as defenses, which users can interactively apply to attacked audio samples. We show that these techniques, which are based on psychoacoustic principles, effectively eliminate targeted attacks, reducing the attack success rate from 92.5% to 0%. We will demonstrate Adagio and invite the audience to try it on the Mozilla Common Voice dataset. Code related to this paper is available at: https://github.com/nilakshdas/ADAGIO.
January 2019
·
152 Reads
·
167 Citations
Lecture Notes in Computer Science
Given the ability to directly manipulate image pixels in the digital input space, an adversary can easily generate imperceptible perturbations to fool a Deep Neural Network (DNN) image classifier, as demonstrated in prior work. In this work, we propose ShapeShifter, an attack that tackles the more challenging problem of crafting physical adversarial perturbations to fool image-based object detectors like Faster R-CNN. Attacking an object detector is more difficult than attacking an image classifier, as it needs to mislead the classification results in multiple bounding boxes with different scales. Extending the digital attack to the physical world adds another layer of difficulty, because it requires the perturbation to be robust enough to survive real-world distortions due to different viewing distances and angles, lighting conditions, and camera limitations. We show that the Expectation over Transformation technique, which was originally proposed to enhance the robustness of adversarial perturbations in image classification, can be successfully adapted to the object detection setting. ShapeShifter can generate adversarially perturbed stop signs that are consistently mis-detected by Faster R-CNN as other objects, posing a potential threat to autonomous vehicles and other safety-critical computer vision systems. Code related to this paper is available at: https://github.com/shangtse/robust-physical-attack.
July 2018
·
240 Reads
·
195 Citations
The rapidly growing body of research in adversarial machine learning has demonstrated that deep neural networks (DNNs) are highly vulnerable to adversarially generated images. This underscores the urgent need for practical defense techniques that can be readily deployed to combat attacks in real-time. Observing that many attack strategies aim to perturb image pixels in ways that are visually imperceptible, we place JPEG compression at the core of our proposed SHIELD defense framework, utilizing its capability to effectively "compress away" such pixel manipulation. To immunize a DNN model from artifacts introduced by compression, SHIELD "vaccinates" the model by retraining it with compressed images, where different compression levels are applied to generate multiple vaccinated models that are ultimately used together in an ensemble defense. On top of that, SHIELD adds an additional layer of protection by employing randomization at test time that compresses different regions of an image using random compression levels, making it harder for an adversary to estimate the transformation performed. This novel combination of vaccination, ensembling, and randomization makes SHIELD a fortified multi-pronged defense. We conducted extensive, large-scale experiments using the ImageNet dataset, and show that our approaches eliminate up to 98% of gray-box attacks delivered by strong adversarial techniques such as Carlini-Wagner's L2 attack and DeepFool. Our approaches are fast and work without requiring knowledge about the model.
May 2018
·
122 Reads
Adversarial machine learning research has recently demonstrated the feasibility to confuse automatic speech recognition (ASR) models by introducing acoustically imperceptible perturbations to audio samples. To help researchers and practitioners gain better understanding of the impact of such attacks, and to provide them with tools to help them more easily evaluate and craft strong defenses for their models, we present ADAGIO, the first tool designed to allow interactive experimentation with adversarial attacks and defenses on an ASR model in real time, both visually and aurally. ADAGIO incorporates AMR and MP3 audio compression techniques as defenses, which users can interactively apply to attacked audio samples. We show that these techniques, which are based on psychoacoustic principles, effectively eliminate targeted attacks, reducing the attack success rate from 92.5% to 0%. We will demonstrate ADAGIO and invite the audience to try it on the Mozilla Common Voice dataset.
April 2018
·
659 Reads
·
216 Citations
Given the ability to directly manipulate image pixels in the digital input space, an adversary can easily generate imperceptible perturbations to fool a Deep Neural Network (DNN) image classifier, as demonstrated in prior work. In this work, we tackle the more challenging problem of crafting physical adversarial perturbations to fool image-based object detectors like Faster R-CNN. Attacking an object detector is more difficult than attacking an image classifier, as it needs to mislead the classification results in multiple bounding boxes with different scales. Extending the digital attack to the physical world adds another layer of difficulty, because it requires the perturbation to be robust enough to survive real-world distortions due to different viewing distances and angles, lighting conditions, and camera limitations. We show that the Expectation over Transformation technique, which was originally proposed to enhance the robustness of adversarial perturbations in image classification, can be successfully adapted to the object detection setting. Our approach can generate adversarially perturbed stop signs that are consistently mis-detected by Faster R-CNN as other objects, posing a potential threat to autonomous vehicles and other safety-critical computer vision systems.
February 2018
·
209 Reads
·
58 Citations
The rapidly growing body of research in adversarial machine learning has demonstrated that deep neural networks (DNNs) are highly vulnerable to adversarially generated images. This underscores the urgent need for practical defense that can be readily deployed to combat attacks in real-time. Observing that many attack strategies aim to perturb image pixels in ways that are visually imperceptible, we place JPEG compression at the core of our proposed Shield defense framework, utilizing its capability to effectively "compress away" such pixel manipulation. To immunize a DNN model from artifacts introduced by compression, Shield "vaccinates" a model by re-training it with compressed images, where different compression levels are applied to generate multiple vaccinated models that are ultimately used together in an ensemble defense. On top of that, Shield adds an additional layer of protection by employing randomization at test time that compresses different regions of an image using random compression levels, making it harder for an adversary to estimate the transformation performed. This novel combination of vaccination, ensembling, and randomization makes Shield a fortified multi-pronged protection. We conducted extensive, large-scale experiments using the ImageNet dataset, and show that our approaches eliminate up to 94% of black-box attacks and 98% of gray-box attacks delivered by the recent, strongest attacks, such as Carlini-Wagner's L2 and DeepFool. Our approaches are fast and work without requiring knowledge about the model.
December 2017
·
122 Reads
·
8 Citations
Cybersecurity analysts are often presented suspicious machine activity that does not conclusively indicate compromise, resulting in undetected incidents or costly investigations into the most appropriate remediation actions. There are many reasons for this: deficiencies in the number and quality of security products that are deployed, poor configuration of those security products, and incomplete reporting of product-security telemetry. Managed Security Service Providers (MSSP's), which are tasked with detecting security incidents on behalf of multiple customers, are confronted with these data quality issues, but also possess a wealth of cross-product security data that enables innovative solutions. We use MSSP data to develop Virtual Product, which addresses the aforementioned data challenges by predicting what security events would have been triggered by a security product if it had been present. This benefits the analysts by providing more context into existing security incidents (albeit probabilistic) and by making questionable security incidents more conclusive. We achieve up to 99% AUC in predicting the incidents that some products would have detected had they been present.
... If the difference between classifier outputs exceeds a threshold, the input is identified as an adversarial example. Freitas et al [27]. proposed UnMask, which extracts FIGURE 1. Process of the cross-modal semantic embedding-based adversarial example detection method. ...
December 2020
... Adversarial examples can be categorized into digital and physical adversarial types. Digital adversarial examples involve introducing pixel-level perturbations to the model's digital input, while physical ones manipulate real-world objects or surroundings, indirectly influencing the model inputs [8]- [10]. ...
January 2019
Lecture Notes in Computer Science
... The robustness of attacks is also evaluated against defense methods, including purification methods of NRP [4], DS [52], diffusion-based purification [5] and adversarial training robust models of Inc-V3 [8], Res-50 [9], Swin-B [53], ConvNeXt-B [53]. Two classic image transformation defenses of JPEG compression [54], Bit-depth reduction [55], and another type of defense, random smoothing [11], are also included. Considering the robustness and transferability of attacks are comparable only under close perturbation budget, the unrestricted attacks are not included in this and the next section. ...
July 2018
... We assume the pedestrian is able to obtain footage of the observed scene, and knows the object detector used in the surveillance system. These assumptions are usually considered in adversarial machine learning research [4,6,9,11,[18][19][20][21][22][23]. We also assume that the pedestrian has access to footage recorded by the pedestrian detection system (e.g., access to a live stream on the Internet). ...
April 2018
... Another idea is data recovery [36], where the perturbation patterns are reduced and the resulting recovered image is fed into a regular deep model. Here, by separating clean and perturbed data in the image, one can either directly remove the perturbed data (i.e., data compression) [37] or introduce randomness in the perturbed data to destroy their patterns (i.e., data randomization) [38]. Since discriminatively separating perturbed data and accurately reconstructing clean images are still open problems, this class of methods obviously faces a tricky accuracy-robustness tradeoff [19]. ...
February 2018
... Some approaches focus on better separating the true positives from the false positives, typically by estimating anomaly or maliciousness scores for each alert [13], [19], [21], [22], [26]. Other approaches support the triage process by extracting and enriching the most relevant information of an alert [8], [39], [40]. ...
December 2017
... During the procedure of these studies, there is a multitude of irregularities that can occur, such as missing engagement of participants (e.g., participant forgets to fill out a daily questionnaire) or malfunction of measuring instruments (e.g., empty battery or Bluetooth connectivity issues). Yet, the complex nature of the data and the multitude and unpredictability of potential problem sources make an automated detection of irregularities difficult (e.g., from lost or broken hardware that needs to be replaced to software-related connectivity issues that can be resolved by rebooting), especially as each study is tailored to answer their individual research questions [53]. Researchers, therefore, have to stay in contact with their participants and regularly monitor the collected high-dimensional data streams during run time of the study to ensure data quality and prevent data loss commonly associated with mHealth data [72]. ...
July 2017
... Modern manufacturing operations are controlled by various actuators that execute commands from programmable controllers [155], where sensor data is processed and analyzed using different algorithms for decision-making [99]. The signal an actuator perceives can also be changed, causing it to implement wrong decisions [156,157]. For example, in a beverage production line, sensors determine the level of filling in bottles, and a robotic end-effector pushes overfilled and underfilled bottles away from the conveyor belt. ...
May 2017
... These sources encompass a diverse range of information, often collected from various origins, to create comprehensive datasets for training and evaluating AI algorithms. Some datasets from recent research that can be cited include: 3. [17] amalgamates data from a diverse array of sources to fuel its analysis. The dataset encompasses 2543 historical fire incidents generously provided by the City of Atlanta Fire Rescue Department (AFRD), 32488 fire inspections, structural information concerning commercial properties procured from the CoStar Group, parcel data from Atlanta's Office of Buildings, and business license records obtained from the City of Atlanta's Office of Revenue, socioeconomic and demographic data from the U.S. Census Bureau, liquor license records, 2014 crime data sourced from the Atlanta Police Department, and Certificate of Occupancy (CO) data from the Atlanta Office of Buildings. ...
August 2016
... The hope is that, through iterations of the cycle, the performance, believability, and interestingness/complexity of the AI will improve. Within human-in-the-loop systems, visualizations are specifically leveraged to aid readability and interpretability [6,13,23]. Teso et al. [18] demonstrated the benefits of this set up, where explanations led to more interpretable outputs and helped users provide more effective feedback to a model, ultimately improving its performance. Such an approach for behavior-language authored agents may help resolve the issues surrounding interpreting and debugging an agent's behavior, but, to the authors' knowledge, no such system currently exists. ...
October 2015