ArticlePDF Available

Invisible CMOS Camera Dazzling for Conducting Adversarial Attacks on Deep Neural Networks

MDPI
Sensors
Authors:

Abstract and Figures

Despite the outstanding performance of deep neural networks, they remain vulnerable to adversarial attacks. While digital domain adversarial attacks are well-documented, most physical-world attacks are typically visible to the human eye. Here, we present a novel invisible optical-based physical adversarial attack via dazzling a CMOS camera. This attack involves using a designed light pulse sequence spatially transformed within the acquired image due to the camera’s shutter mechanism. We provide a detailed analysis of the photopic conditions required to keep the attacking light source invisible to human observers while effectively disrupting the image, thereby deceiving the DNN. The results indicate that the light source duty cycle controls the tradeoff between the attack’s success rate and the degree of concealment needed.
This content is subject to copyright.
Academic Editor: Gaochang Wu
Received: 11 February 2025
Revised: 31 March 2025
Accepted: 2 April 2025
Published: 4 April 2025
Citation: Stein, Z.; Hazan, A.; Stern,
A. Invisible CMOS Camera Dazzling
for Conducting Adversarial Attacks on
Deep Neural Networks. Sensors 2025,
25, 2301. https://doi.org/10.3390/
s25072301
Copyright: © 2025 by the authors.
Licensee MDPI, Basel, Switzerland.
This article is an open access article
distributed under the terms and
conditions of the Creative Commons
Attribution (CC BY) license
(https://creativecommons.org/
licenses/by/4.0/).
Article
Invisible CMOS Camera Dazzling for Conducting Adversarial
Attacks on Deep Neural Networks
Zvi Stein , Adir Hazan * and Adrian Stern
School of Electrical and Computer Engineering, Ben-Gurion University of the Negev, Beer-Sheva 8410501, Israel;
tzviste@post.bgu.ac.il (Z.S.); stern@bgu.ac.il (A.S.)
*Correspondence: hazanad@post.bgu.ac.il
Abstract: Despite the outstanding performance of deep neural networks, they remain
vulnerable to adversarial attacks. While digital domain adversarial attacks are well-
documented, most physical-world attacks are typically visible to the human eye. Here, we
present a novel invisible optical-based physical adversarial attack via dazzling a CMOS
camera. This attack involves using a designed light pulse sequence spatially transformed
within the acquired image due to the camera’s shutter mechanism. We provide a detailed
analysis of the photopic conditions required to keep the attacking light source invisible to
human observers while effectively disrupting the image, thereby deceiving the DNN. The
results indicate that the light source duty cycle controls the tradeoff between the attack’s
success rate and the degree of concealment needed.
Keywords: adversarial attack; PSF; rolling shutter; CMOS
1. Introduction
Deep Neural Networks (DNNs) have revolutionized the field of image analysis and
processing, delivering state-of-the-art performance across a range of applications. However,
these systems are inherently vulnerable to adversarial attacks [
1
], which introduce subtle
perturbations to the input signal that cause the DNNs to make incorrect predictions. The
concept of adversarial examples, commonly known as attacked images, was first introduced
a decade ago by Szegedy et al. [
2
], demonstrating that DNNs could be easily misled
by seemingly minor modifications to input images. Since then, numerous approaches
for generating adversarial examples have been explored [
3
], highlighting the significant
security concerns surrounding DNN-based systems.
The underlying mechanism for adversarial susceptibility lies in the way DNNs process
images. Rather than learning the actual semantic content of the image, these networks
often rely on superficial or spurious features for classification, as described by Goodfellow
et al. as a “Potemkin village” of features [
4
]. This explains why two images that are visually
indistinguishable from human vision can be classified differently by a DNN, revealing
a vulnerability that adversarial attacks exploit. These attacks often aim to minimize the
perturbations applied to an image so that the changes are not noticeable to the human eye
while still causing a misclassification.
Adversarial attacks on DNNs can be divided into digital and physical attacks. While
digital attacks manipulate image pixels, they often struggle to transfer to the physical
world due to dynamic conditions and deployment challenges. Physical attacks alter real-
world objects’ visual characteristics and pose a threat but are typically invasive, requiring
visible changes that can be easily dismissed and detected by human vision. However,
Sensors 2025,25, 2301 https://doi.org/10.3390/s25072301
Sensors 2025,25, 2301 2 of 15
optical-based physical adversarial attacks are non-invasive and generate perturbations
that mimic natural effects, making them harder to detect and better suited for real-world
applications [
5
]. Despite advancements in imperceptibility, many of these attacks still have
an obvious trace in the physical domain, limiting their effectiveness and feasibility, with
achieving complete invisibility to the human eye remaining an unresolved challenge.
This paper introduces and demonstrates a novel optical-based physical adversarial
attack that leverages the rolling shutter mechanism of CMOS sensors. The proposed attack
is designed to be invisible in the physical domain, ensuring that the attacking light source
remains undetectable to the scene observer. This involves a designed light pulse sequence
spatially transformed during the image acquisition, effectively disrupting the camera’s
image processing to deceive DNNs with a high attack success rate. Furthermore, our
approach does not require precise alignment of the adversarial spatial pattern with the
target object location, offering greater flexibility in real-world scenarios. A successful
invisible attack is achieved when the beam of Attacking MOdulated Light Source (AMOLS)
covers the camera aperture, such that the following are achieved:
1. The peak irradiance is sufficient to dazzle the sensor temporarily;
2. The average irradiance remains below the sensitivity threshold of the human eye.
The following summarizes the primary contributions of this work:
We propose a physical domain adversarial attack on DNNs that receive images from
a CMOS camera. The attack involves directing a light source toward the camera;
however, the presence of the projected light is completely unnoticed by observers in
the scene.
We introduce an optical attack that is based on dazzling a camera sensor by sending
short pulses. We investigate the effect of the projected pulses on the image captured
by the CMOS camera. We evaluate the irradiance required to attack the image.
We explore the relationship between the human eye’s ability to distinguish the at-
tacking light source directed at the camera and the disruption of DNN performance
caused by the influence of the pulsed laser beam. We analyze the photopic conditions
required to ensure that the attacking light source remains invisible to human observers
while still effectively disrupting the acquired image to mislead the classifier model.
We evaluate the trade-off between the success of DNN attacks caused by dazzling
pulses and their invisibility to the human eye. Our findings indicate that the duty
cycle of the light source can be adjusted to manage the balance between the attack’s
success rate and the level of concealment required.
We present simulated and real experimental results to demonstrate the effectiveness
of our attack.
2. Related Works
While most studies on adversarial attacks have focused on the digital domain, where
perturbations are added to pixel values, growing efforts have expanded into the physical
domain [
6
]. Examples of physical-world attacks typically include using adversarial objects
or imaging system manipulations to fool DNN models. These modifications may include
simple changes, such as adding elements like stickers, eyeglasses, earrings, and others to a
real-world object [
7
], to more complex approaches. The more complex methods typically
involve optical-based techniques [
5
], including temporarily projecting specifically crafted
adversarial perturbations onto target objects [
8
], among others, or strategically illuminating
target objects using infrared light sources [
9
]. Furthermore, synthesizing Three-Dimensional
(3D) adversarial objects has been proposed to confuse classifier models [
10
], and imaging
projection transformation in a 3D physical environment was demonstrated to deceive object
Sensors 2025,25, 2301 3 of 15
detection systems effectively [
11
]. These examples highlight the growing applicability of
adversarial attacks in real-world settings.
Recent studies on physical adversarial examples have increasingly focused on manip-
ulating imaging systems themselves. For instance, Liu et al. [
12
] induced perturbations
in the captured image through an electromagnetic injection attack. They focused on CCD
sensors but noted that CMOS sensors, which have an independent measurement unit for
each pixel, provide greater resilience to electromagnetic interference, making them more
robust against such threats. Additionally, Duan et al. [
13
] employed a laser beam attack to
create spatially tailored perturbations; however, they noted that this approach has a limited
success rate in dynamic conditions. Many physical adversarial attacks require precise
alignment of the adversarial spatial pattern with the target object placement. Moreover,
Liu et al. [
14
] inject their attack after image acquisition, targeting the data lane between the
camera sensor and the endpoint device. This requires physical access to the sensor-enabled
system, which is practically infeasible in certain situations.
In this work, we develop an invisible camera dazzling attack that leverages the rolling
shutter mechanism inherent in CMOS sensors. Unlike the continuous-wave operation
of light sources, where the degree of dazzle on CMOS sensors can be depicted by the
dazzling area or the number of saturated pixels [
15
], temporally modulated light can pro-
duce adjustable stripes in a captured image—introducing a unique approach to injecting
adversarial spatial patterns. The rolling shutter effect is primarily studied in the context of
mitigating distortions caused by fast-moving objects that approach the camera’s scanning
frequency [
16
]. Accordingly, models have been developed to correct these distortions.
Moreover, it was proposed that a smartphone camera can be used for visible light commu-
nications to detect and convert a temporal signal into spatial patterns by exploiting the
rolling shutter effect of CMOS sensors [17].
Adversarial attacks leveraging the rolling shutter mechanism have been introduced in
references [
18
22
], where temporally modulated LEDs are used to illuminate a target object,
as shown in Figure 1a. This results in distortions in the acquired image due to the camera’s
row-wise scanning process. The first configuration [
18
] was introduced as a black-box
backdoor attack on face recognition systems, where illuminating the entire scene induces
perturbations employing the rolling shutter effect. While the first two studies [
18
,
19
]
utilize programmable RGB LEDs, resulting in an adversarial signal with three adjustable
components of Red, Green, and Blue, later work [
20
] demonstrated the use of a common
commercial LED with a modulator to control the frequency of the emitted white light. In
addition, further schemes [
21
,
22
] expanded the application of the white light attack method,
showcasing the generalization and transferability of adversarial samples across different
models and tasks, including traffic sign recognition systems and lane detection models.
However, these approaches require comprehensive illumination of the whole scene and
usually fail to remain invisible to the human eye. Despite the light pulse sequence being
designed with a modulation frequency that prevents flickering perceived by the human
eye, the illumination source still appears steady and is not stealthy to the human observer
in the scene.
Here, we propose to employ an AMOLS beam that directly illuminates the camera’s
aperture as shown in Figure 1b, taking advantage of the rolling shutter’s scanning process
to induce real-world adversarial perturbations on the acquired image. Since the pulsed light
beam is directed toward the camera rather than reflecting off a target object (see Figure 1),
the average power requirements are significantly reduced compared to previous methods.
While Kohler et al. [
23
] and Yan et al. [
24
] introduced such a camera attack utilizing a laser
and exploiting the rolling shutter mechanism, their approaches still leave an obvious trace
of the attack in the physical domain and remain visible to the human eye.
Sensors 2025,25, 2301 4 of 15
Sensors2025,25,xFORPEERREVIEW4of16
Figure1.Practicalphysical-worldadversarialaack.Theaackcanbecarriedouteither(a)bytem-
porallymodulatingalightsourcetoilluminatetheentirescene,whichreectslightpulsesontothe
CMOSsensor,or(b)bydirectingapulsedlaserbeamspecicallyataCMOSsensor.Theredarrows
indicatethepropagationdirectionofthelight.
Here,weproposetoemployanAMOLSbeamthatdirectlyilluminatesthecameras
apertureasshowninFigure1b,takingadvantageoftherollingshuer’sscanningprocess
toinducereal-worldadversarialperturbationsontheacquiredimage.Sincethepulsed
lightbeamisdirectedtowardthecameraratherthanreectingoffatargetobject(seeFig-
ure1),theaveragepowerrequirementsaresignicantlyreducedcomparedtoprevious
methods.WhileKohleretal.[23]andYanetal.[24]introducedsuchacameraaackuti-
lizingalaserandexploitingtherollingshuermechanism,theirapproachesstillleavean
obvioustraceoftheaackinthephysicaldomainandremainvisibletothehumaneye.
Sincetheintegrationtimeofthehumaneyeissignicantlylongerthantheacquisi-
tiontimeofeachrowinarollingshuerscanningprocess,ahigh-frequencymodulated
signalisseenascontinuousbythehumaneye.IfdenotingthedutycycleoftheAMOLS
as𝐷,theintensityperceivedbythehumaneyecanbeexpressedasfollows:
𝐼

𝐷𝐼

.(1)
Thatis,thehumaneyeonlyperceivesthesignalsaveragepower.Consequently,itispos-
sibletocontrolthisintensitybyappropriatelyreducingthedutycycleoftheAMOLS.In
thispaper,weexploretherelationshipbetweentheeectivenessofadutycycleduringa
directcameraaackandtheabilitytodistinguishtheAMOLSimplementation.First,we
reviewtheeectofAMOLSonthecameraanddeterminetheirradianceneededtopro-
ducethedesireddisruptiveeectonthecamera.Next,weevaluatethedazzlingirradi-
anceonthehumaneyeanddeterminetheconditionsthatinuencetheeyesabilityto
perceiveandrecognizethelightsource.Finally,afterestablishingtheirradiancerequire-
ments,weexaminetheeciencyofimagedistortioncausedbythedesignedpulsese-
quenceonawell-knownclassier,theResidualNeuralNetwork(ResNet50)architecture,
throughsimulationsandexperiments.
3.MaterialsandMethods
3.1.DazzleEectwithRollingShuer
Camera
Thespatialspreadofapointsourceintheimageplaneisconventionallydescribed
bythediractionofthePointSparedFunction(PSF),generallygivenbytheFouriertrans-
formationoftheentrancepupil.However,particularlyforbrightpowersources(e.g.,a
Figure 1. Practical physical-world adversarial attack. The attack can be carried out either (a) by
temporally modulating a light source to illuminate the entire scene, which reflects light pulses onto
the CMOS sensor, or (b) by directing a pulsed laser beam specifically at a CMOS sensor. The red
arrows indicate the propagation direction of the light.
Since the integration time of the human eye is significantly longer than the acquisition
time of each row in a rolling shutter scanning process, a high-frequency modulated signal
is seen as continuous by the human eye. If denoting the duty cycle of the AMOLS as
D
, the
intensity perceived by the human eye can be expressed as follows:
Ieye =D·Isource. (1)
That is, the human eye only perceives the signal’s average power. Consequently, it is
possible to control this intensity by appropriately reducing the duty cycle of the AMOLS.
In this paper, we explore the relationship between the effectiveness of a duty cycle during a
direct camera attack and the ability to distinguish the AMOLS implementation. First, we
review the effect of AMOLS on the camera and determine the irradiance needed to produce
the desired disruptive effect on the camera. Next, we evaluate the dazzling irradiance on
the human eye and determine the conditions that influence the eye’s ability to perceive
and recognize the light source. Finally, after establishing the irradiance requirements,
we examine the efficiency of image distortion caused by the designed pulse sequence on
a well-known classifier, the Residual Neural Network (ResNet50) architecture, through
simulations and experiments.
3. Materials and Methods
3.1. Dazzle Effect with Rolling Shutter Camera
The spatial spread of a point source in the image plane is conventionally described
by the diffraction of the Point Spared Function (PSF), generally given by the Fourier
transformation of the entrance pupil. However, particularly for bright power sources
(e.g., a laser source), other effects such as stray light scattering and halo [
25
] may occur in
addition to the PSF diffraction, which may be considerably more significant than the PSF.
The dazzling effect is demonstrated in Figure 2, where the measurement is acquired from a
laptop camera (installed on a DELL-INSPIRON laptop with 0.92 Megapixel, 88
diagonal
viewing angle). The AMOLS average power was 5
mW
with
3.5
mm
spot diameter. As
shown in Figure 2, a notable dazzling effect is observed when utilizing such a power level.
Sensors 2025,25, 2301 5 of 15
Sensors2025,25,xFORPEERREVIEW5of16
lasersource),othereectssuchasstraylightscaeringandhalo[25]
mayoccurinaddi-
tiontothePSFdiraction,whichmaybeconsiderablymoresignicantthanthePSF.The
dazzlingeectisdemonstratedinFigure2,wherethemeasurementisacquiredfroma
laptopcamera(installedonaDELL-INSPIRONlaptopwith0.92Megapixel,88°diagonal
viewingangle).TheAMOLSaveragepowerwas5 mWwith~3.5 mm
spotdiameter.As
showninFigure2,anotabledazzlingeectisobservedwhenutilizingsuchapowerlevel.
Figure2.ExperimentalPSFmeasurement.Thecamera’sresponsetoplacedpointsourcewithinthe
eldofview.Theradiantuxmeasuredintheobjectplaneis
~50 mW/cm
.
Previousstudiesoninfraredimagers[26,27]haveempiricallyshownthatthediame-
terofthesaturatedareaintheimageplane,denotedas𝑥

,canbeapproximatedasfol-
lows:
𝑥

󰇡

󰇢
 (2)
where𝐼
and𝐼

arethelaserirradianceandthesaturationlevel,respectively.Basedon
resultsforvisiblelightusingaCMOScamera[28,29]aminimumaverageirradianceof
50 mW/cm
duringeachrowexposureisrequired,andatleast0.1 mW/cm
pickirradi-
ancetoachievedazzlingwithshorterpulses.Weexperimentallyfoundthatsimilarcon-
ditionsholdforthecamerausedinthiswork,asobservedinFigure2.
Next,thedazzlingeectformedintheaackedimageisexamined.Witharolling
shuercamera,everyrowintheframecollectsambientlightduringdierentperiods.As
showninFigure3,the𝑖-throwofthesensorrecordsthelightintegratedduringtheperiod
from𝑡
𝑡

till𝑡
,whileforthefollowingrow𝑖1,theintegrationtimewillbeuntil
𝑡
𝑡

,where𝑡

denotesthereadingtimeofasinglerowand𝑡

denotestheex-
posuretimeofasinglerow.Thedurationofscanningeachframe,denotedby𝑡

,can
beexpressedasfollows[16]:
𝑡

𝑡

󰇛𝑁
𝑁

󰇜𝑡

.(3)
where𝑁
and𝑁

arethenumberofpixelrowsandthenumberofhiddenpixelrowsin
eachframe,respectively.
Figure 2. Experimental PSF measurement. The camera’s response to placed point source within the
field of view. The radiant flux measured in the object plane is ~50 mW/cm2.
Previous studies on infrared imagers [
26
,
27
] have empirically shown that the diameter
of the saturated area in the image plane, denoted as xsat , can be approximated as follows:
xsat I0
Isat 1
3(2)
where
I0
and
Isat
are the laser irradiance and the saturation level, respectively. Based
on results for visible light using a CMOS camera [
28
,
29
] a minimum average irradiance
of 50
mW/cm2
during each row exposure is required, and at least 0.1
mW/cm2
pick
irradiance to achieve dazzling with shorter pulses. We experimentally found that similar
conditions hold for the camera used in this work, as observed in Figure 2.
Next, the dazzling effect formed in the attacked image is examined. With a rolling
shutter camera, every row in the frame collects ambient light during different periods.
As shown in Figure 3, the
i
-th row of the sensor records the light integrated during the
period from
titex p
till
ti
, while for the following row
i+
1, the integration time will be
until
ti+tread
, where
tread
denotes the reading time of a single row and
tex p
denotes the
exposure time of a single row. The duration of scanning each frame, denoted by
tf ram e
, can
be expressed as follows [16]:
tf ram e =tread(Nr+NrH)+tex p. (3)
where
Nr
and
NrH
are the number of pixel rows and the number of hidden pixel rows in
each frame, respectively.
Sensors2025,25,xFORPEERREVIEW6of16
Figure3.AschematicillustrationoftherollingshuereectcausedbydazzlingAMOLS.Theroll-
ingshuermechanismtransformsthetemporalsignalwithadesignedsequenceoflaserpulses
(markedinblueatthetop)intospatialdistortion.Thisdistortionoccursduringdierentperiodsof
readingandexposureforthepixelrowsintheframe(indicatedbywhiteandgrayblocks).Asa
result,astripe-likepaernemergesintheacquiredimage(right).
Theratio𝑅
𝑡

/𝑡

determinesthenumberofexposedrowsatanygiventime
(seeFigure4).Thus,𝑅
isreferredtoastherow’sexposureconstant.Itisworthhigh-
lightingthatifthepulsedurationgeneratedbytheAMOLSisshorterthan𝑡

,exactly
𝑅
rowswillbedazzled,regardlessofthepulsewidth.Forinstance,bothpulseswitha
duration1 μsand2 μswillproducethesamepaernwhenusingatypicalcamerawith
areadingtimeof𝑡

30 μs.Theexperimentallyobtaineddazzlepaernfortherolling
shuersensorwhentheAMOLSisappliedisshowninFigure4,alongwiththesimulated
stripe-linepaernutilizing𝑅
37obtainedwithacalibrationprocess.Thesimulation
resultcorrespondswellwiththeexperimentalmeasurement,withastructuralsimilarity
of93%.
Figure4.Dazzleeectofrollingshuersensorbyamodulatedlightsource.(a,b)Theresulting
dazzlepaernforAMOLSvia(a)experimentand(b)simulationwith𝑅
37.
3.2.PhotopicConditionsforInvisibility
Thissectionfocusesondeterminingthephotometricconditionsrequiredtokeepthe
AMOLSeectivelyinvisible.TheaackscenarioisdepictedinFigure5,whereatarget
object(car)andtheAMOLSareplacedinfrontofacamerawhileanobserverisnearthe
cameraatanangle𝜃relativetotheopticalaxis.TheacquiredimageisthenfedtoDNN
toclassifythetarget.ConsiderthattheAMOLSpowerissettoproduceanirradianceof
𝑒50 mW/cm
atthesensorplanewhenactive.Theaveragepower𝐸receivedbythe
observer’seyesfromtheAMOLSisinuencedbythedutycycleoftheAMOLSduring
Figure 3. A schematic illustration of the rolling shutter effect caused by dazzling AMOLS. The rolling
shutter mechanism transforms the temporal signal with a designed sequence of laser pulses (marked
in blue at the top) into spatial distortion. This distortion occurs during different periods of reading
and exposure for the pixel rows in the frame (indicated by white and gray blocks). As a result, a
stripe-like pattern emerges in the acquired image (right).
Sensors 2025,25, 2301 6 of 15
The ratio
Rn=tex p/tread
determines the number of exposed rows at any given time
(see Figure 4). Thus,
Rn
is referred to as the row’s exposure constant. It is worth highlighting
that if the pulse duration generated by the AMOLS is shorter than
tread
, exactly
Rn
rows
will be dazzled, regardless of the pulse width. For instance, both pulses with a duration
1
µs
and 2
µs
will produce the same pattern when using a typical camera with a reading
time of
tread
30
µs
. The experimentally obtained dazzle pattern for the rolling shutter
sensor when the AMOLS is applied is shown in Figure 4, along with the simulated stripe-
line pattern utilizing
Rn=
37 obtained with a calibration process. The simulation result
corresponds well with the experimental measurement, with a structural similarity of 93%.
Sensors2025,25,xFORPEERREVIEW6of16
Figure3.AschematicillustrationoftherollingshuereectcausedbydazzlingAMOLS.Theroll-
ingshuermechanismtransformsthetemporalsignalwithadesignedsequenceoflaserpulses
(markedinblueatthetop)intospatialdistortion.Thisdistortionoccursduringdierentperiodsof
readingandexposureforthepixelrowsintheframe(indicatedbywhiteandgrayblocks).Asa
result,astripe-likepaernemergesintheacquiredimage(right).
Theratio𝑅
𝑡

/𝑡

determinesthenumberofexposedrowsatanygiventime
(seeFigure4).Thus,𝑅
isreferredtoastherow’sexposureconstant.Itisworthhigh-
lightingthatifthepulsedurationgeneratedbytheAMOLSisshorterthan𝑡

,exactly
𝑅
rowswillbedazzled,regardlessofthepulsewidth.Forinstance,bothpulseswitha
duration1 μsand2 μswillproducethesamepaernwhenusingatypicalcamerawith
areadingtimeof𝑡

30 μs.Theexperimentallyobtaineddazzlepaernfortherolling
shuersensorwhentheAMOLSisappliedisshowninFigure4,alongwiththesimulated
stripe-linepaernutilizing𝑅
37obtainedwithacalibrationprocess.Thesimulation
resultcorrespondswellwiththeexperimentalmeasurement,withastructuralsimilarity
of93%.
Figure4.Dazzleeectofrollingshuersensorbyamodulatedlightsource.(a,b)Theresulting
dazzlepaernforAMOLSvia(a)experimentand(b)simulationwith𝑅
37.
3.2.PhotopicConditionsforInvisibility
Thissectionfocusesondeterminingthephotometricconditionsrequiredtokeepthe
AMOLSeectivelyinvisible.TheaackscenarioisdepictedinFigure5,whereatarget
object(car)andtheAMOLSareplacedinfrontofacamerawhileanobserverisnearthe
cameraatanangle𝜃relativetotheopticalaxis.TheacquiredimageisthenfedtoDNN
toclassifythetarget.ConsiderthattheAMOLSpowerissettoproduceanirradianceof
𝑒50 mW/cm
atthesensorplanewhenactive.Theaveragepower𝐸receivedbythe
observer’seyesfromtheAMOLSisinuencedbythedutycycleoftheAMOLSduring
Figure 4. Dazzle effect of rolling shutter sensor by a modulated light source. (a,b) The resulting
dazzle pattern for AMOLS via (a) experiment and (b) simulation with Rn=37.
3.2. Photopic Conditions for Invisibility
This section focuses on determining the photometric conditions required to keep the
AMOLS effectively invisible. The attack scenario is depicted in Figure 5, where a target
object (car) and the AMOLS are placed in front of a camera while an observer is near the
camera at an angle
θ
relative to the optical axis. The acquired image is then fed to DNN
to classify the target. Consider that the AMOLS power is set to produce an irradiance of
e=
50
mW/cm2
at the sensor plane when active. The average power
E
received by the
observer’s eyes from the AMOLS is influenced by the duty cycle of the AMOLS during the
frame exposure period. The light source duty cycle denoted by
D
determines the average
power of the light source, which can be expressed as
E=e·D
. In addition, assuming the
AMOLS is smaller than the human eye’s angular resolution, the strictest condition would
be the concentration of the seen power from any given source. Thus, a larger angular extent
covered by the AMOLS would yield a lower peak power.
Sensors2025,25,xFORPEERREVIEW7of16
theframeexposureperiod.Thelightsourcedutycycledenotedby𝐷determinestheav-
eragepowerofthelightsource,whichcanbeexpressedas𝐸𝑒𝐷.Inaddition,assum-
ingtheAMOLSissmallerthanthehumaneyesangularresolution,thestrictestcondition
wouldbetheconcentrationoftheseenpowerfromanygivensource.Thus,alargeran-
gularextentcoveredbytheAMOLSwouldyieldalowerpeakpower.
Figure5.InvisibleAMOLSimplementationfordirectcameraaack.Atargetobject(e.g.,acar)is
placedinthecamera’seldofview,andalightsourcedirectlyilluminatesthecamera(bysending
abeambetweentheredarrows).ThetaskoftheDNNistoclassifytheacquiredimage.Whenap-
plyingtheAMOLS,itmustremaininvisibletoanobserveratanangle𝜃relativetotheopticalaxis.
Bydenotingthebackgroundbrightnessby𝐿
andtheAMOLSbrightnessby𝐿

,the
contrastcanbegivenbythefollowing:
𝐶


, (4)
PreviousstudiesbyH.R.Blackwell[30]andW.Adrian[31]investigatedthethresh-
oldcontrast𝐶

requiredtodetectanobject.AccordingtoW.Adrian,atargetcontrastof
1atasmallangleissucienttorecognizethetarget.Sinceradianceisaphysicalquantity
conservedthroughoutanopticalsystem,itdictatesthebrightness.Whenthesolidangle
coveredbythetargetissmallerthanthesystemsresolvingpower,theAMOLSbrightness
hasthefollowingform[32]:
𝐿

𝐸683 ∙𝑉
∙Ω


󰇟cd m

󰇠,(5)
where𝑉
denotesthephotopicecacyandΩ

istheresolvingpowerofthehuman
eye(representingthestrictestconditionregardingthereceivedpower).Employingacam-
eramodeltorepresenttheeyemodel,C.A.WilliamsonandL.N.McLin[33,34]proposed
ascaeringfunctionbasedonempiricalndingsbyJ.Vosetal.[35],withaneective
solidanglecollectedbytheeye:
𝑓

󰇛𝜃,𝐴,𝑝,𝐿
󰇜𝑆𝐿
∙𝑔

󰇛𝜃,𝐴,𝑝󰇜 󰇟sr

󰇠,(6)
where𝑔

canbedeterminedbytheo-axisangle𝜃(seeFigure5),theageA(inyears),
andtheeyepigment𝑝,whichisgivenbythefollowing:
𝑔

󰇛𝜃,𝐴,𝑝󰇜

󰇣
.
󰇤1󰇡
.
󰇢
0.0025𝑝 󰇟sr

󰇠,(7)
Substitutingthetermoftheaveragepower𝐸,andtheangularresolutionby𝑓

,the
AMOLSbrightnessexpressedinEquation(5)takesthefollowingform:
𝐿

𝑒∙𝐷 683 ∙𝑉
𝑓

󰇟cd m

󰇠,(8)
Figure 5. Invisible AMOLS implementation for direct camera attack. A target object (e.g., a car) is
placed in the camera’s field of view, and a light source directly illuminates the camera (by sending a
beam between the red arrows). The task of the DNN is to classify the acquired image. When applying
the AMOLS, it must remain invisible to an observer at an angle θrelative to the optical axis.
Sensors 2025,25, 2301 7 of 15
By denoting the background brightness by
Lb
and the AMOLS brightness by
LAS
, the
contrast can be given by the following:
C=LAS Lb
Lb
, (4)
Previous studies by H.R. Blackwell [
30
] and W. Adrian [
31
] investigated the threshold
contrast
Cthr
required to detect an object. According to W. Adrian, a target contrast of 1
at a small angle is sufficient to recognize the target. Since radiance is a physical quantity
conserved throughout an optical system, it dictates the brightness. When the solid angle
covered by the target is smaller than the system’s resolving power, the AMOLS brightness
has the following form [32]:
LAS =E·683·Vλ·2
eye hcd·m2i, (5)
where
Vλ
denotes the photopic efficacy and
eye
is the resolving power of the human eye
(representing the strictest condition regarding the received power). Employing a camera
model to represent the eye model, C.A. Williamson and L.N. McLin [
33
,
34
] proposed a
scattering function based on empirical findings by J. Vos et al. [
35
], with an effective solid
angle collected by the eye:
feye(θ,A,p,Lb)=S·LT
b·geye(θ,A,p)hsr1i, (6)
where
geye
can be determined by the off-axis angle
θ
(see Figure 5), the age A(in years), and
the eye pigment p, which is given by the following:
geye(θ,A,p)=10
θ3+5
θ2+0.1p
θ"1+A
62.52#+0.0025p[sr1], (7)
Substituting the term of the average power
E
, and the angular resolution by
feye
, the
AMOLS brightness expressed in Equation (5) takes the following form:
LAS =e·D·683·Vλ·feye hcd·m2i, (8)
Finally, by substituting Equation (8) into Equation (4), the light source duty cycle can
be expressed by the following:
D=L1T
b
Cthr (Lb)+1
e·683·Vλ·S·geye(θ,A,p). (9)
Figure 6shows the light source duty cycle
D
required for dazzling as a function of the
background illumination for various viewing aspect angles. As the aspect angle increases,
the effective radiance on the retina decreases. Consequently, the contrast decreases with
the increasing background brightness, requiring more power to exceed the threshold. It is
observed from the results shown in Figure 6that for observers placed at an angle greater
than 10 degrees, a duty cycle of 0.5% is sufficient to keep the source invisible, regardless of
the background illumination level. The following sections will present a technique that can
be operated even at lower duty cycle percentages.
Sensors 2025,25, 2301 8 of 15
Sensors2025,25,xFORPEERREVIEW8of16
Finally,bysubstitutingEquation(8)intoEquation(4),thelightsourcedutycyclecan
beexpressedbythefollowing:
𝐷𝐿


󰇛
󰇜
∙
∙∙

󰇛,,󰇜
. (9)
Figure6showsthelightsourcedutycycle𝐷requiredfordazzlingasafunctionof
thebackgroundilluminationforvariousviewingaspectangles.Astheaspectanglein-
creases,theeectiveradianceontheretinadecreases.Consequently,thecontrastde-
creaseswiththeincreasingbackgroundbrightness,requiringmorepowertoexceedthe
threshold.ItisobservedfromtheresultsshowninFigure6thatforobserversplacedatan
anglegreaterthan10degrees,adutycycleof0.5%issucienttokeepthesourceinvisi-
ble,regardlessofthebackgroundilluminationlevel.Thefollowingsectionswillpresenta
techniquethatcanbeoperatedevenatlowerdutycyclepercentages.
Figure6.Thedutycycle𝐷oftheAMOLSatthethresholdofhumandiscriminationasafunction
ofthebackgroundluminanceforvariousviewingangles𝜃.
3.3.GeneratingthePhysicalAdversarialAack
Followingtheformalismin[2],theproblemofndinganadversarialexamplecanbe
formallydenedasfollows:
minimize |𝑥
󰆒
𝑥|
s. t. 𝐶󰇛𝑥
󰆒
󰇜𝑙
𝑥
󰆒
󰇟0, 1󰇠
,
(10)
where𝑥 istheundistributedimage,𝑥′ istheperturbatedimage,𝑙 representsthe
groundtruthlabeloftheimage𝑥,and𝐶󰇛𝑥󰇜denotestheDNNusedasaclassier.Peri-
odically,solvingsuchaproblemcanbeincrediblycomplex,whichleadstosolvingamore
straightforwardprobleminstead,assuggestedin[36].Inbrief,thegoalistondasmall
perturbation𝛿 𝑥
󰆒
𝑥,whichcanbeappliedtoanimage𝑥 toalteritsclassication
whileensuringthattheresultingimageremainsvalid.ConsideringthattheSoftmax𝑉
isappliedontopoftheDNNlogits,thelossfunctionmappinganimage𝑥toapositive
realnumbercanbedescribedasfollows
𝑓
󰇛𝑥󰇜 𝐿𝑜𝑠𝑠
,
󰇛𝑉
󰇛𝑥󰇜󰇜,(11)
Accordingly,insteadofformulatingtheconstraintminimizationproblemasinEquation
(10),onecanuseanalternativeformulationandsolvethefollowingproblem:
Figure 6. The duty cycle
D
of the AMOLS at the threshold of human discrimination as a function of
the background luminance for various viewing angles θ.
3.3. Generating the Physical Adversarial Attack
Following the formalism in [
2
], the problem of finding an adversarial example can be
formally defined as follows:
minimize ||xx||22
s.t. C(x)=l
x[0, 1]n,
(10)
where
x
is the undistributed image,
x
is the perturbated image,
l
represents the ground
truth label of the image
x
, and
C(x)
denotes the DNN used as a classifier. Periodically,
solving such a problem can be incredibly complex, which leads to solving a more straightfor-
ward problem instead, as suggested in [
36
]. In brief, the goal is to find a small perturbation
δ=xx
, which can be applied to an image
x
to alter its classification while ensuring that
the resulting image remains valid. Considering that the Softmax
Vn
is applied on top of
the DNN logits, the loss function mapping an image
x
to a positive real number can be
described as follows
f(x)=LossC,l(Vn(x)), (11)
Accordingly, instead of formulating the constraint minimization problem as in Equation
(10), one can use an alternative formulation and solve the following problem:
minimize ||δ||0α·f(x+δ)
s.t. x+δ[0, 1]n.(12)
where
α
represents the ratio between the magnitude of the disturbance and its effect’s
intensity on the output, and · 0denotes the zero norm.
In our case, we aim to establish a relation between the pulsed laser activity and the
resulting adversarial perturbation caused by the rolling shutter mechanism of the CMOS
camera. This mechanism converts the temporal signal of the designed laser pulse se-
quence into a spatial distortion within the acquired image.
Ee f f
is an
N
-dimensional
binary row vector representing the pulsed laser activity, which can be expressed by
N=(Nr+NrH)/Rn
, where
Rn
denotes the number of dazzled pixel rows by each pulse
and
(Nr+NrH)
indicates the sensor’s total number of pixel rows (see Section 3.1). Specifi-
cally, a unit value at the
i
-th component of this vector
Ee f f [i]=
1, indicates a pulse occurring
Sensors 2025,25, 2301 9 of 15
at the time
t=i·tf ram e/N
, and dazzles the sensor’s pixel rows from
i·Rn
to
(i+1)·Rn
.
Thus, the indices of the dazzled pixel rows in the acquired image can be obtained by
substituting each unit entry of the pulsed laser activity vector
ET
e f f
with a size
Rn
vector of
ones, which is given by the following:
ET
r=ET
e f f O1T
Rn(13)
where
is the Kronecker product and
1T
Rn
is an
Rn
-dimensional column vector of ones.
Consequently,
ET
r
is an
N·Rn
-dimensional binary column vector in which unit entries
indicate the dazzled sensor’s pixel rows. Next, the resulting dazzle pattern in the acquired
N×Mimage (e.g., Figure 4) can be obtained by the following:
δ=ET
rO1M=ET
e f f O1T
RnO1M, (14)
where
1M
is a size
M
vector of ones corresponding to the number of pixel columns in the
acquired image. Instead of formulating the minimization problem following Equation (12),
we now use an alternative formulation expressed in terms of the pulsed laser activity vector
Ee f f —the problem then becomes as follows: given x, find δthat satisfies the following:
minimize ET
e f f 0α·f(x+δ)
s.t. δ[0, 1]n,(15)
In practice, to implement a typical gradient-based optimization algorithm (such as
SGD or ADAM) for solving Equation (15), we replace the binary vector derived from
Equation (14). Rather than optimizing over the variable
δ
defined above, we change the
variables and optimize over ωT, which has the following form:
δ=1
2tanhωT+1O1T
RnO1M, (16)
where δ[0, 1]n, and ωThas the same dimensions as ET
e f f .
Since the exact moment of camera exposure is unknown to the attacker in a real-world
setting, applying the AMOLS, consisting of a designed sequence of laser pulses, yields a
dazzling pattern with a random horizontal shift. Considering the asynchrony between
the attacking light pulse sequence and the camera’s exposure moment, we utilize the
Expectation over Transformation (EoT) method [10] as follows:
minimize Et0TnET
e f f 0α·f(x+δ)o. (17)
where Tis the space of all possible instances of frame exposure, denoted as t0.
4. Results and Discussion
This section presents the feasibility of conducting invisible adversarial attacks on
DNNs in the physical domain by dazzling the camera. In addition, we evaluate the
AMOLS performance using optimal dazzle patterns following the method described in
Section 3.3, considering the pulsed laser activity depicted in Section 3.1. In the following
sections, we employ both simulations and real experiments. First, we conduct simulations
to investigate the effect of the AMOLS duty cycle while maintaining a constant pulse width.
Next, we optically demonstrate the attack and examine its sensitivity to the pulse width.
Sensors 2025,25, 2301 10 of 15
4.1. Effectiveness of the AMOLS
We evaluate the effectiveness of the AMOLS based on the duty cycle of a pulsed
laser (as discussed in Section 3.2) while keeping a constant pulse width. The ResNet50
classifier [
37
] and the standard cross-entropy loss function are utilized to simulate the
adversarial attacks on the image classification model. Figure 7shows simulation results
of the loss function depending on the AMOLS duty cycle for two cases: where an object
covers (1) approximately 40% of the field of view (FOV), and (2) approximately 85% of the
FOV. These results focus on the “Coffee mug” as the target object, with the highest obtained
values for each examined duty cycle as a result of optimizing the attack (as discussed in
Section 3.3). It can be empirically determined that loss function values exceeding 2 exhibit
poor classifier performances, resulting in misclassification across a significant number
of input images—specifically, this enhances the effectiveness of the AMOLS. The results
presented in Figure 7indicate that when the duty cycle is set lower than 0.2%, the attack
remains feasible—yet the classification model tends to yield better results when the target
object covers
40% of the FOV. Conversely, increasing the AMOLS duty cycle substantially
raises the loss, thereby enhancing the effectiveness of the attack in the case of an object
occupying
40% of the FOV. Additionally, for a target object that covers
85% of the
FOV, the attack proves effective across the entire duty cycle range examined, with a milder
dependence on changes in the AMOLS duty cycle.
Figure 7. The effectiveness of the proposed attack on the loss function and its dependency on the
duty cycle Dof the pulsed laser beam.
In addition, we examined a range of target objects during the attack, imaged from
various angles of view corresponding to different classes—several samples are shown in
Figure 8a. An analysis of the effect of the AMOLS duty cycle, while maintaining constant
pulse width, on the classifier’s loss function across diverse input images is shown in
Figure 8b. We empirically found the critical values of the cross-entropy loss function at
which the DNN begins to misclassify objects across different classes, considering an offset
in the obtained loss curves above these critical values. It is observed from the results shown
in Figure 8b that an AMOLS duty cycle of 0.4%, which corresponds to a designed sequence
of 4 laser pulses, successfully fools the classifier in all cases.
Sensors 2025,25, 2301 11 of 15
Sensors2025,25,xFORPEERREVIEW11of16
Figure8b.Weempiricallyfoundthecriticalvaluesofthecross-entropylossfunctionat
whichtheDNNbeginstomisclassifyobjectsacrossdierentclasses,consideringanoset
intheobtainedlosscurvesabovethesecriticalvalues.Itisobservedfromtheresults
showninFigure8bthatanAMOLSdutycycleof0.4%,whichcorrespondstoadesigned
sequenceof4laserpulses,successfullyfoolstheclassierinallcases.
(a)(b)
Figure8.TheAMOLSisappliedtodierentobjects.(a)Examplesofaackedimages.(b)Thede-
pendenceofthelossfunctionontheaackinglightsourcedutycycleforvariousobjects.
4.2.RealExperimentsonPhysical-WorldAdversarialAack
Wecarriedoutrealexperimentstoevaluatethefeasibilityoftheproposedoptical-
basedphysicaladversarialaacksinreal-worldscenariosformedbyconvertingthelight
temporalsignaltoaspatialdistortionwithintheacquiredimage.Acoeemugisusedas
thetargetobjectandplacedinsidetheFOVofalaptopcamera(installedonaDELL-IN-
SPIRONlaptopwith0.92Megapixel,88° diagonalviewingangle).Fortheaack,a
pulsedlaserbeamisdirectedatthecamerafromapositionadjacenttotheobject,pro-
ducedfroma650nmdotdiodelaser,withanaveragepowerof5mWandaspotsizeof
3.5mm.Asequenceofpulsesisdesignedtogeneratetheadversarialdazzlepaernfol-
lowingtheoptimizationmethoddescribedinSection3.2,wherethetemporalmodulating
signalisproducedutilizingtheArduino-Unomicrocontrollerboard.Thecameracaptures
boththelightreectedfromtheobjectandthelightemiedbytheAMOLS.Theacquired
imagesarethenfedtotheDNNforclassication.Weconductedourexperimentswithno
ambientlight,asthisrepresentsthemostchallengingconditionforourproblemseing,
whichrequiresthelightsourcetoremaininvisibletoahumanobserver.Asillustratedin
Figure6,asthebackgroundilluminationdecreases,theallowableAMOLSillumination
budgetthatcanremaininvisiblealsodecreases.Conversely,alowerAMOLSillumination
budgetchallengesthesuccessofaacks,asindicatedbythereducedclassicationloss
showninFigure8b.
Figure9a,bshowstwooptical-basedphysicaladversarialexamplesandtheircorre-
spondingpredictionsfromtheimageclassicationmodel.Theseexamplesweregener-
atedfromtwoseparateexposureshots,wheretheAMOLSuseddierentpulsewidths.It
isworthmentioningthattheaackinglightpulsesequenceisnotsynchronizedwiththe
camerasexposuremoment(seeSection3.3),leadingtovariationsinthedazzlepaern
acrosseachframe,specicallyintroducingahorizontalshift.Videosshowingthefootage
fromtheaackedcamerasequenceareprovidedintheSupplementaryMaterials.Addi-
tionalexamplescanbefoundontheGitHubrepositoryassociatedwiththispaperat
hps://github.com/ZviSteinOpt/RollingShuerAack/tree/main(accessedon1April
2025).TheinvisibleCMOScameradazzlingaackinducesmisclassicationacrossthe
Figure 8. The AMOLS is applied to different objects. (a) Examples of attacked images. (b) The
dependence of the loss function on the attacking light source duty cycle for various objects.
4.2. Real Experiments on Physical-World Adversarial Attack
We carried out real experiments to evaluate the feasibility of the proposed optical-
based physical adversarial attacks in real-world scenarios formed by converting the light
temporal signal to a spatial distortion within the acquired image. A coffee mug is used
as the target object and placed inside the FOV of a laptop camera (installed on a DELL-
INSPIRON laptop with 0.92 Megapixel, 88
diagonal viewing angle). For the attack, a
pulsed laser beam is directed at the camera from a position adjacent to the object, produced
from a 650 nm dot diode laser, with an average power of 5 mW and a spot size of 3.5 mm.
A sequence of pulses is designed to generate the adversarial dazzle pattern following the
optimization method described in Section 3.2, where the temporal modulating signal is
produced utilizing the Arduino-Uno microcontroller board. The camera captures both the
light reflected from the object and the light emitted by the AMOLS. The acquired images
are then fed to the DNN for classification. We conducted our experiments with no ambient
light, as this represents the most challenging condition for our problem setting, which
requires the light source to remain invisible to a human observer. As illustrated in Figure 6,
as the background illumination decreases, the allowable AMOLS illumination budget that
can remain invisible also decreases. Conversely, a lower AMOLS illumination budget
challenges the success of attacks, as indicated by the reduced classification loss shown in
Figure 8b.
Figure 9a,b shows two optical-based physical adversarial examples and their corre-
sponding predictions from the image classification model. These examples were generated
from two separate exposure shots, where the AMOLS used different pulse widths. It
is worth mentioning that the attacking light pulse sequence is not synchronized with
the camera’s exposure moment (see Section 3.3), leading to variations in the dazzle pat-
tern across each frame, specifically introducing a horizontal shift. Videos showing the
footage from the attacked camera sequence are provided in the Supplementary Materi-
als. Additional examples can be found on the GitHub repository associated with this
paper at https://github.com/ZviSteinOpt/RollingShutterAttack/tree/main (accessed on
1 April 2025).
The invisible CMOS camera dazzling attack induces misclassification across
the input images, significantly reducing the classifier’s confidence in the correct 500th class,
as shown in Figure 9c.
Sensors 2025,25, 2301 12 of 15
Sensors2025,25,xFORPEERREVIEW12of16
inputimages,signicantlyreducingtheclassier’scondenceinthecorrect500thclass,
asshowninFigure9c.
Figure9.ResultsofAMOLSrealizationonanimageclassicationmodel.Physical-worldadversar-
ialexamplesgeneratedviatwoshotsrecordingwhenseingdierentAMOLSactivities:(a)pulse
widthof1 μswith𝐷0.01%,and(b)pulsewidthof70 μswith𝐷0.85%.(c)TheDNNscon-
denceinthepredictedresultsacrossthe1000classesitwastrainedon,withtheindexforthecor-
rect“coeemug”labelbeing#500.
ThedistributionofpredictionsmadebyatargetedDNNmodelacrossvariousclasses
duringtheoptical-basedphysicaladversarialaacksisdepictedinFigure10.Itisbased
on254repeatedtrials,wheretheAMOLSoperatesfourpulses,havingapulsedurationof
1 μs.Theresultsindicatethatthedesignedaackachievedan85%successrateunder
theseconditions.TheresultsshowninFigure11indicatethatahigheraacksuccessrate
canbeachievedbyincreasingthepulsewidth.WhentheAMOLSpulsewidthexceeds
approximately70 μs,thephysical-worldaacksuccessrateapproaches98%.However,
followingSection3.2,increasingthepulsewidthreducestherangeofconcealedviewing
angles(seeFigure6).Theseexhibitatradeoffbetweentheangularrealmachievinginvis-
ibilityandthesuccessrateofthephysical-worldaackastheAMOLSdutycyclevaries.
Consideringthatthecameracaptures30framespersecond,apulseof1 μscorresponds
toalowdutycycleof0.012%(𝐷100 41 μs30 s

0.012%),whereas
apulsedura-
tionof 70 μsresultsinahigherdutycycleof0.84%.Itcanbeobservedfromtheresults
showninFigure6thatseingadutycycleof0.01%ensurestheAMOLSactivityremains
invisibletotheobserverlocatedatanglesgreaterthanapproximatelyfromtheoptical
axis.Incomparison,adutycycleof0.85%couldbesucienttomaintaintheinvisibility
ofoptical-basedphysicaladversarialaacksataviewingangleof15°.
TheperformanceandpropertiesofouraackaresummarizedinTableA1inAppen-
dixA,togetherwithacomparisontothatofotherphysicaladversarialaacksinvolving
imagesensors.
Figure 9. Results of AMOLS realization on an image classification model. Physical-world adversarial
examples generated via two shots recording when setting different AMOLS activities: (a) pulse width
of 1
µs
with
D=
0.01%, and (b) pulse width of 70
µs
with
D=
0.85%. (c) The DNN’s confidence in
the predicted results across the 1000 classes it was trained on, with the index for the correct “coffee
mug” label being #500.
The distribution of predictions made by a targeted DNN model across various classes
during the optical-based physical adversarial attacks is depicted in Figure 10. It is based
on 254 repeated trials, where the AMOLS operates four pulses, having a pulse duration
of 1
µs
. The results indicate that the designed attack achieved an 85% success rate under
these conditions. The results shown in Figure 11 indicate that a higher attack success
rate can be achieved by increasing the pulse width. When the AMOLS pulse width
exceeds approximately 70
µs
, the physical-world attack success rate approaches 98%.
However, following Section 3.2, increasing the pulse width reduces the range of concealed
viewing angles (see Figure 6). These exhibit a tradeoff between the angular realm achieving
invisibility and the success rate of the physical-world attack as the AMOLS duty cycle varies.
Considering that the camera captures 30 frames per second, a pulse of 1
µs
corresponds to
a low duty cycle of 0.012% (
D=
100
·
4
·
1
µs·
30
s1=
0.012
%
), whereas a pulse duration of
70
µs
results in a higher duty cycle of 0.84
%
. It can be observed from the results shown in
Figure 6that setting a duty cycle of 0.01% ensures the AMOLS activity remains invisible
to the observer located at angles greater than approximately 5
from the optical axis.
In comparison, a duty cycle of 0.85% could be sufficient to maintain the invisibility of
optical-based physical adversarial attacks at a viewing angle of 15.
Sensors2025,25,xFORPEERREVIEW13of16
Figure10.ThefrequencydistributionoftheDNNpredictionsduringtheaack.Whiletheobject’s
correctlabelisa“coeemug”,theaackexhibitsanaacksuccessrateof85%.
Figure11.TheaverageaacksuccessrateasafunctionoftheAMOLSpulsewidth.
5.Conclusions
Insummary,weintroducedanovelmethodforconductingoptical-basedphysical
adversarialaacksonDNN.Theaackisdemonstratedbydirectingapulsedlightata
CMOScamera.Therollingshuermechanismofthecameraconvertsthetemporalsignal,
whichconsistsofthedesignedsequenceoflightpulses,intoaspatialdistortionwithinthe
physical-worldadversarialimage.Thephotometricconditionsandlightpulsecharacter-
isticsareanalyzedtodazzletheCMOScamerasuciently,therebyfoolingtheDNN
modelwhilekeepingtheAMOLSactivityinvisibletoobserversintheenvironment.
Wedemonstratedthatthelightsourcedutycycleenablesthecontrolofthetradeoff
betweentheaack’ssuccessrateandtherequiredangulardegreeofconcealment.For
instance,withtheproposedmethod,an85%successrateforthephysical-worldaackcan
beachievedwhileensuringtheinvisibilityoflightsourceactivitytotheobserverexcept
foranarrowangularrangeoffromtheopticalaxis.However,theaacksuccessrate
couldbeincreasedto98%byallowingaslightreductionof10°intheangularconceal-
mentrange.
SupplementaryMaterials:Thefollowingsupportinginformationcanbedownloadedat:
www.mdpi.com/xxx/s1.Thefollowingsupportingvideosshowthefootagefromtheaackedcam-
erasequencewhenseingdierentAMOLSactivities.VideoS1:pulsewidthof1 μs with𝐷
0.01%,andVideoS2:pulsewidthof70 μswith𝐷0.85%.
Chain
Coffe mug
Cup
Digital cloack
Hodometer
Knot
Ruler
Website
Class
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
0.45
0.5
Frequency [%]
Figure 10. The frequency distribution of the DNN predictions during the attack. While the object’s
correct label is a “coffee mug”, the attack exhibits an attack success rate of 85%.
Sensors 2025,25, 2301 13 of 15
Sensors2025,25,xFORPEERREVIEW13of16
Figure10.ThefrequencydistributionoftheDNNpredictionsduringtheaack.Whiletheobject’s
correctlabelisa“coeemug”,theaackexhibitsanaacksuccessrateof85%.
Figure11.TheaverageaacksuccessrateasafunctionoftheAMOLSpulsewidth.
5.Conclusions
Insummary,weintroducedanovelmethodforconductingoptical-basedphysical
adversarialaacksonDNN.Theaackisdemonstratedbydirectingapulsedlightata
CMOScamera.Therollingshuermechanismofthecameraconvertsthetemporalsignal,
whichconsistsofthedesignedsequenceoflightpulses,intoaspatialdistortionwithinthe
physical-worldadversarialimage.Thephotometricconditionsandlightpulsecharacter-
isticsareanalyzedtodazzletheCMOScamerasuciently,therebyfoolingtheDNN
modelwhilekeepingtheAMOLSactivityinvisibletoobserversintheenvironment.
Wedemonstratedthatthelightsourcedutycycleenablesthecontrolofthetradeoff
betweentheaack’ssuccessrateandtherequiredangulardegreeofconcealment.For
instance,withtheproposedmethod,an85%successrateforthephysical-worldaackcan
beachievedwhileensuringtheinvisibilityoflightsourceactivitytotheobserverexcept
foranarrowangularrangeoffromtheopticalaxis.However,theaacksuccessrate
couldbeincreasedto98%byallowingaslightreductionof10°intheangularconceal-
mentrange.
SupplementaryMaterials:Thefollowingsupportinginformationcanbedownloadedat:
www.mdpi.com/xxx/s1.Thefollowingsupportingvideosshowthefootagefromtheaackedcam-
erasequencewhenseingdierentAMOLSactivities.VideoS1:pulsewidthof1 μs with𝐷
0.01%,andVideoS2:pulsewidthof70 μswith𝐷0.85%.
Chain
Coffe mug
Cup
Digital cloack
Hodometer
Knot
Ruler
Website
Class
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
0.45
0.5
Frequency [%]
Figure 11. The average attack success rate as a function of the AMOLS pulse width.
The performance and properties of our attack are summarized in Table A1 in
Appendix A
, together with a comparison to that of other physical adversarial attacks
involving image sensors.
5. Conclusions
In summary, we introduced a novel method for conducting optical-based physical
adversarial attacks on DNN. The attack is demonstrated by directing a pulsed light at
a CMOS camera. The rolling shutter mechanism of the camera converts the temporal
signal, which consists of the designed sequence of light pulses, into a spatial distortion
within the physical-world adversarial image. The photometric conditions and light pulse
characteristics are analyzed to dazzle the CMOS camera sufficiently, thereby fooling the
DNN model while keeping the AMOLS activity invisible to observers in the environment.
We demonstrated that the light source duty cycle enables the control of the tradeoff
between the attack’s success rate and the required angular degree of concealment. For
instance, with the proposed method, an 85% success rate for the physical-world attack can
be achieved while ensuring the invisibility of light source activity to the observer except for
a narrow angular range of 5
from the optical axis. However, the attack success rate could
be increased to 98% by allowing a slight reduction of 10
in the angular concealment range.
Supplementary Materials: The following supporting information can be downloaded at: https:
//www.mdpi.com/article/10.3390/s25072301/s1. The following supporting videos show the footage
from the attacked camera sequence when setting different AMOLS activities. Video S1: pulse width
of 1 µs with D=0.01%, and Video S2: pulse width of 70 µs with D=0.85%.
Author Contributions: Conceptualization, A.S. and A.H.; methodology, Z.S.; software, Z.S.; val-
idation, Z.S., A.H. and A.S.; investigation, Z.S.; data curation, Z.S.; writing—original draft, Z.S.;
writing—review and editing, A.H. and A.S.; supervision, A.H. and A.S.; project administration, A.S.
All authors have read and agreed to the published version of the manuscript.
Funding: This research received no external funding.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The original data presented in the study are openly available in the
GitHub repository at https://github.com/ZviSteinOpt/RollingShutterAttack/tree/main (accessed
on 1 April 2025).
Conflicts of Interest: The authors declare no conflicts of interest.
Sensors 2025,25, 2301 14 of 15
Appendix A
Table A1. Comparison of physical adversarial attacks involving the image sensors.
Physical World Attack Attack Mechanism Targeting Camera
Sensors
Adversary Physical
Access
Achievable Attack
Success Rate
Invisibility
Criterion
EM Injection [12] CCD interface X Near distances 50% 94% a
AdvLB [13] Spatial laser beam X X 77.43% 100% bX
CamData Lane [14] Camera data lane X Camera interface 89.2% 96% c
RS Backdoor Attack [18] CMOS dazzling X40% 88% dXf
Adversarial RS [19] CMOS dazzling X84% Xf
Our Attack Invisible AMOLS X 85% 98% e
a
Average performance from various viewpoints depending on the threat model.
b
Depending on indoor or
outdoor attacks.
c
Depending on the DNN model.
d
Based on simulation study or physical-domain study.
e
Depending on the observer zone location restriction (Figure 5).
f
Designed to prevent visible flickering, although
the illumination source may be seen shining. RS—Rolling Shutter. X and
represent whether the attacks target
the camera sensors to inject their perturbations, require physical access by the adversary, or satisfy the invisibility
criterion in the physical domain. represents a designed attack assuming adversary physical access.
References
1. Heaven, D. Why deep-learning AIs are so easy to fool. Nature 2019,574, 163–166. [CrossRef] [PubMed]
2.
Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks.
In Proceedings of the International Conference on Learning Representations (ICLR), Banff, AB, Canada, 14–16 April 2014.
3. Yuan, X.; He, P.; Zhu, Q.; Li, X. Adversarial Examples: Attacks and Defenses for Deep Learning. IEEE Trans. Neural Netw. Learn.
Syst. 2019,30, 2805–2824. [CrossRef] [PubMed]
4.
Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. In Proceedings of the International
Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015.
5.
Fang, J.; Jiang, Y.; Jiang, C.; Jiang, Z.L.; Liu, C.; Yiu, S.M. State-of-the-art optical-based physical adversarial attacks for deep
learning computer vision systems. Expert Syst. Appl. 2024,250, 123761. [CrossRef]
6.
Wang, J.; Wang, C.; Lin, Q.; Luo, C.; Wu, C.; Li, J. Adversarial attacks and defenses in deep learning for image recognition: A
survey. Neurocomputing 2022,514, 162–181.
7.
Wei, H.; Tang, H.; Jia, X.; Wang, Z.; Yu, H.; Li, Z.; Satoh, S.I.; Van Gool, L.; Wang, Z. Physical adversarial attack meets computer
vision: A decade survey. IEEE Trans. Pattern Anal. Mach. Intell. 2024,46, 9797–9817. [CrossRef] [PubMed]
8.
Lovisotto, G.; Turner, H.; Sluganovic, I.; Strohmeier, M.; Martinovic, I. SLAP: Improving Physical Adversarial Examples with
Short-Lived Adversarial Perturbations. arXiv 2020, arXiv:2007.04137.
9.
Zhou, Z.; Tang, D.; Wang, X.; Han, W.; Lu, X.; Zhang, K. Invisible Mask: Practical attacks on face recognition with infrared. arXiv
2018, arXiv:1803.04683v1.
10.
Athalye, A.; Engstrom, L.; Ilyas, A.; Kwok, K. Synthesizing Robust Adversarial Examples. In Proceedings of the 35th International
Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018.
11.
Cui, J.; Guo, W.; Huang, H.; Lv, X.; Cao, H.; Li, H. Adversarial examples for vehicle detection with projection transformation.
IEEE Trans. Geosci. Remote Sens. 2024,62, 5632418. [CrossRef]
12.
Liu, Z.; Lin, F.; Ba, Z.; Lu, L.; Ren, K. MagShadow: Physical Adversarial Example Attacks via Electromagnetic Injection. IEEE
Trans. Dependable Secur. Comput. 2025, 1–17. [CrossRef]
13.
Duan, R.; Mao, X.; Qin, A.K.; Chen, Y.; Ye, S.; He, Y.; Yang, Y. Adversarial laser beam: Effective physical-world attack to dnns in a
blink. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June
2021; pp. 16062–16071.
14.
Liu, W.; He, W.; Hu, B.; Chang, C.H. A practical man-in-the-middle attack on deep learning edge device by sparse light strip
injection into camera data lane. In Proceedings of the 2022 IEEE 35th International System-on-Chip Conference (SOCC), Belfast,
UK, 5–8 September 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6.
15.
Wang, X.; Xu, Z.; Zhong, H.; Cheng, X.A.; Xing, Z.; Zhang, J. Fresnel Diffraction Model for Laser Dazzling Spots of Complementary
Metal Oxide Semiconductor Cameras. Sensors 2024,24, 5781. [CrossRef] [PubMed]
16.
Chia-Kai, L. Analysis and Compensation of Rolling Shutter Effect. IEEE Trans. Image Process. 2008,17, 1323–1330. [CrossRef]
[PubMed]
17.
Danakis, C.; Afgani, M.; Povey, G.; Underwood, I.; Haas, H. Using a CMOS Camera Sensor for Visible Light Communication. In
Proceedings of the IEEE Globecom Workshops, Anaheim, CA, USA, 3–7 December 2012.
18.
Li, H.; Wang, Y.; Xie, X.; Liu, Y.; Wang, S.; Wan, R.; Chau, L.P.; Kot, A.C. Light Can Hack Your Face! Black-box Backdoor Attack on
Face Recognition Systems. arXiv 2020, arXiv:2009.06996.
Sensors 2025,25, 2301 15 of 15
19.
Sayles, A.; Hooda, A.; Gupta, M.; Chatterjee, R.; Fernandes, E. Invisible Perturbations: Physical Adversarial Examples Exploiting
the Rolling Shutter Effect. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville,
TN, USA, 20–25 June 2021; pp. 14666–14675.
20.
Chen, Z.; Lin, P.; Jiang, Z.L.; Wei, Z.; Yuan, S.; Fang, J. An illumination modulation-based adversarial attack against automated
face recognition system. In Proceedings of the Information Security and Cryptology: 16th International Conference, Inscrypt
2020, Guangzhou, China, 11–14 December 2020; Revised Selected Papers. Springer International Publishing: Cham, Switzerland,
2021; pp. 53–69.
21.
Shen, Y.; Cheng, Y.; Lin, Y.; Long, S.; Jiang, C.; Li, D.; Dai, S.; Jiang, Y.; Fang, J.; Jiang, Z.L.; et al. MLIA: Modulated LED
illumination-based adversarial attack on traffic sign recognition system for autonomous vehicle. In Proceedings of the 2022 IEEE
International Conference on Trust, Security and Privacy in Computing and Communications (TrustCom), Wuhan, China, 9–11
December 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1020–1027.
22.
Fang, J.; Yang, Z.; Dai, S.; Jiang, Y.; Jiang, C.; Jiang, Z.L.; Liu, C.; Yiu, S.M. Cross-task physical adversarial attack against lane
detection system based on LED illumination modulation. In Proceedings of the Chinese Conference on Pattern Recognition and
Computer Vision (PRCV), Xiamen, China, 13–15 October 2023; Springer Nature: Singapore, 2023; pp. 478–491.
23.
Köhler, S.; Lovisotto, G.; Birnbach, S.; Baker, R.; Martinovic, I. They See Me Rollin’: Inherent Vulnerability of the Rolling Shutter
in CMOS Image Sensors. In Proceedings of the 37th Annual Computer Security Applications Conference, Virtual, 6–10 December
2021; pp. 399–413.
24.
Yan, C.; Xu, Z.; Yin, Z.; Mangard, S.; Ji, X.; Xu, W.; Zhao, K.; Zhou, Y.; Wang, T.; Gu, G.; et al. Rolling colors: Adversarial laser
exploits against traffic light recognition. In Proceedings of the 31st USENIX Security Symposium (USENIX Security 22), Boston,
MA, USA, 10–12 August 2022; pp. 1957–1974.
25.
Nilson, D.G.; Hill, D.N.; Evans, J.C. Thomson Scattering Stray Light Reduction Techniques Using a CCD Camera; Lawrence Livermore
National Laboratory: Livermore, CA, USA, 1997.
26.
Schleijpen, R.H.M.A.; Dimmeler, A.; Eberle, B.; van den Heuvel, J.C.; Mieremet, A.L.; Beckman, H.; Mellier, B. Laser Dazzling of
Focal Plane Array Cameras. In Proceedings of the Defense and Security Symposium, Orlando, FL, USA, 10 October 2007.
27.
Schleijpen, H.M.A.; Carpenter, S.R.; Mellier, B.; Dimmeler, A. Imaging Seeker Surrogate for IRCM evaluation. In Proceedings of
the Optics/Photonics in Security and Defence, Stockholm, Sweden, 5 October 2006.
28.
Santos, C.N.; Chrétien, S.; Merella, L.; Vandewal, M. Visible and near-infrared laser dazzling of CCD and CMOS cameras. In
Proceedings of the Technologies for Optical Countermeasures XV, Berlin, Germany, 9 October 2018.
29.
Eberle, B.; Kinerk, W.T.; Koerber, M.; Öhgren, J.; Ritt, G.; Santos, C.N.; Schwarz, B.; Steinvall, O.; Tipper, S.M.; Vandewal, M.;
et al. NATO SET-249 joint measurement campaign on laser dazzle effects in airborne scenarios. In Proceedings Volume 11161,
Technologies for Optical Countermeasures XVI; SPIE: Bellingham, WA, USA, 2019; pp. 119–138.
30. Blackwell, H.R. Contrast Thresholds of the Human Eye. J. Opt. Soc. Am. 1946,36, 624–643. [CrossRef] [PubMed]
31. Adrian, W. Visibility of targets: Model for calculation. Light Res. Technol. 1989,21, 181–188.
32.
Paschotta, R. RP Photonics Encyclopedia. Available online: https://www.rp-photonics.com/radiance.html (accessed on 1 April 2025).
33. Williamson, C.A.; McLin, L.N. Nominal ocular dazzle distance (NODD). Appl. Opt. 2015,54, 1564–1572. [CrossRef]
34.
McLin, L.N.; Smith, P.A.; Barnes, L.E.; Dykes, J.R. Scaling laser disability glare functions with “K” factors to predict dazzle. In
International Laser Safety; AIP Publishing: Albuquerque, NM, USA, 2015.
35.
Vos, J.; Cole, B.; Bodmann, H.-W.; Colombo, E.; Takeuchi, T.; van den Berg, T.J.T.P. CIE Equations for Disability Glare; CIE TC:
Vienna, Austria, 2002.
36.
Carlini, N.; Wagner, D. Towards Evaluating the Robustness of Neural Networks. In Proceedings of the 2017 IEEE Symposium on
Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017.
37.
He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on
Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.
ResearchGate has not been able to resolve any citations for this publication.
Article
Full-text available
Laser dazzling on complementary metal oxide semiconductor (CMOS) image sensors is an effective method in optoelectronic countermeasures. However, previous research mainly focused on the laser dazzling under far fields, with limited studies on situations that the far-field conditions were not satisfied. In this paper, we established a Fresnel diffraction model of laser dazzling on a CMOS by combining experiments and simulations. We calculated that the laser power density and the area of saturated pixels on the detector exhibit a linear relationship with a slope of 0.64 in a log-log plot. In the experiment, we found that the back side illumination (BSI-CMOS) matched the simulations, with an error margin of 3%, while the front side illumination (FSI-CMOS) slightly mismatched the simulations, with an error margin of 14%. We also found that the full-screen saturation threshold for the BSI-CMOS was 25% higher than the FSI-CMOS. Our work demonstrates the applicability of the Fresnel diffraction model for BSI-CMOS, which provides a valuable reference for studying laser dazzling.
Article
Physical adversarial examples (AEs) have become an increasing threat to deploying deep neural network (DNN) models in the real world. Popular approaches adopt sticking-based or projecting-based strategies that stick the printed adversarial patches to objects or directly project the AE onto objects. Although effective, these methods require access to target objects and generate visible artifacts, which reduces the attack's stealthiness. In this paper, we propose MagShadow, a new attack vector that leverages imperceptible electromagnetic (EM) signals to realize physical AEs. MagShadow utilizes the CCD camera sensor's susceptibility to EM injection attacks and induces fine-grained adversarial perturbations on the camera's captured image by injecting carefully-crafted signals with a low-cost portable device. As MagShadow directly manipulates the image sensor's output with EM signals, the attack requires no access to the target object and can keep stealthy. We study the feasibility of MagShadow in two typical DNN application scenarios (image classification and object detection) and design a framework to implement four different types of attacks, i.e., untargeted, targeted, hiding, and appearing attacks. Extensive real-world experiments on five different cameras are conducted, which demonstrate MagShadow's effectiveness against different popular DNN models (Inception v3, ResNet101, YOLO v3/v4).
Article
Despite the impressive achievements of Deep Neural Networks (DNNs) in computer vision, their vulnerability to adversarial attacks remains a critical concern. Extensive research has demonstrated that incorporating sophisticated perturbations into input images can lead to a catastrophic degradation in DNNs' performance. This perplexing phenomenon not only exists in the digital space but also in the physical world. Consequently, it becomes imperative to evaluate the security of DNNs-based systems to ensure their safe deployment in real-world scenarios, particularly in security-sensitive applications. To facilitate a profound understanding of this topic, this paper presents a comprehensive overview of physical adversarial attacks. Firstly, we distill four general steps for launching physical adversarial attacks. Building upon this foundation, we uncover the pervasive role of artifacts carrying adversarial perturbations in the physical world. These artifacts influence each step. To denote them, we introduce a new term: adversarial medium. Then, we take the first step to systematically evaluate the performance of physical adversarial attacks, taking the adversarial medium as a first attempt. Our proposed evaluation metric, hiPAA , comprises six perspectives: Effectiveness , Stealthiness , Robustness , Practicability , Aesthetics , and Economics . We also provide comparative results across task categories, together with insightful observations and suggestions for future research directions.
Article
Unmanned aerial vehicle (UAV) imaging object detection systems based on deep neural networks are vulnerable to adversarial patch attacks. However, existing UAV image adversarial patch generation methods mainly target flat digital images, neglecting the adjustments to the adversarial patch morphology brought about by changes in the imaging projection matrix in a three-dimensional physical environment. This leads to adaptability issues with the attack patches, where changes in the UAV’s observation angle might cause dynamic variations in the vehicle target’s position and size, such geometric instability means that preset attack patches may no longer be applicable. To address these issues, this paper proposes an adversarial patch generation method based on a projection transformation model, which we call Projective-Patch attack. To tackle the adaptability problem of the attack patches, we employ an adversarial patch relocation strategy, adjusting the preset adversarial patch’s morphology, size, and positioning on the target vehicle through the projection transformation model to ensure omnidirectional adaptability at different angles and altitudes. We then train an adversarial patch using data captured from specific vehicle targets in multiple scenarios to enhance its attack universality across various real-world aerial photography scenarios. Results from the vanishing attack experiments show our method enhanced the attack success rate against YOLOv3 by 37.62% and 19.95% on the UAV and VisDrone datasets, respectively, compared to the baseline. For the Faster R-CNN detector, the success rates increased by 11.63% and 14.76%. Additionally, we investigated the transferability of projection patches and their attack performance at different sizes and angles, with our adversarial examples consistently performing well.
Chapter
Lane detection is one of the fundamental technologies for autonomous driving, but it faces many security threats from adversarial attacks. Existing adversarial attacks against lane detection often simplify it as a certain type of computer vision task and ignore its cross-task characteristic, resulting in weak transferability, poor stealthiness, and low executability. This paper proposes a cross-task physical adversarial attack scheme based on LED illumination modulation (AdvLIM) after analyzing the decision behavior coexisting in lane detection models. We generate imperceptible brightness flicker through fast intensity modulation of LED illumination and utilize the rolling shutter effect of the CMOS image sensor to inject brightness information perturbations into the captured scene image. According to different modulation parameters, this paper proposes two attack strategies: scatter attack and dense attack. Cross-task attack experiments are conducted on the Tusimple dataset on five lane detection models of three different task types, including segmentation-based models (SCNN, RESA), point detection-based model (LaneATT), and curve-based models (LSTR, BézierLaneNet). The experimental results show that AdvLIM causes a detection accuracy drop of 77.93%, 75.86%, 51.51%, 89.41%, and 73.94% for the above models respectively. Experiments are also conducted in the physical world, and the results indicate that AdvLIM can pose a real and significant security threat to diverse lane detection systems in real-world scenarios.
Article
In recent years, researches on adversarial attacks and defense mechanisms have obtained much attention. It’s observed that adversarial examples crafted with small malicious perturbations would mislead the deep neural network (DNN) model to output wrong prediction results. These small perturbations are imperceptible to humans. The existence of adversarial examples poses great threat to the robustness of DNN-based models. It is necessary to study the principles behind it and develop their countermeasures. This paper surveys and summarizes the recent advances in attack and defense methods extensively and in detail, analyzes and compares the pros and cons of various attack and defense schemes. Finally we discuss the main challenges and future research directions in this field.