Access to this full-text is provided by MDPI.
Content available from Sensors
This content is subject to copyright.
Academic Editor: Gaochang Wu
Received: 11 February 2025
Revised: 31 March 2025
Accepted: 2 April 2025
Published: 4 April 2025
Citation: Stein, Z.; Hazan, A.; Stern,
A. Invisible CMOS Camera Dazzling
for Conducting Adversarial Attacks on
Deep Neural Networks. Sensors 2025,
25, 2301. https://doi.org/10.3390/
s25072301
Copyright: © 2025 by the authors.
Licensee MDPI, Basel, Switzerland.
This article is an open access article
distributed under the terms and
conditions of the Creative Commons
Attribution (CC BY) license
(https://creativecommons.org/
licenses/by/4.0/).
Article
Invisible CMOS Camera Dazzling for Conducting Adversarial
Attacks on Deep Neural Networks
Zvi Stein , Adir Hazan * and Adrian Stern
School of Electrical and Computer Engineering, Ben-Gurion University of the Negev, Beer-Sheva 8410501, Israel;
tzviste@post.bgu.ac.il (Z.S.); stern@bgu.ac.il (A.S.)
*Correspondence: hazanad@post.bgu.ac.il
Abstract: Despite the outstanding performance of deep neural networks, they remain
vulnerable to adversarial attacks. While digital domain adversarial attacks are well-
documented, most physical-world attacks are typically visible to the human eye. Here, we
present a novel invisible optical-based physical adversarial attack via dazzling a CMOS
camera. This attack involves using a designed light pulse sequence spatially transformed
within the acquired image due to the camera’s shutter mechanism. We provide a detailed
analysis of the photopic conditions required to keep the attacking light source invisible to
human observers while effectively disrupting the image, thereby deceiving the DNN. The
results indicate that the light source duty cycle controls the tradeoff between the attack’s
success rate and the degree of concealment needed.
Keywords: adversarial attack; PSF; rolling shutter; CMOS
1. Introduction
Deep Neural Networks (DNNs) have revolutionized the field of image analysis and
processing, delivering state-of-the-art performance across a range of applications. However,
these systems are inherently vulnerable to adversarial attacks [
1
], which introduce subtle
perturbations to the input signal that cause the DNNs to make incorrect predictions. The
concept of adversarial examples, commonly known as attacked images, was first introduced
a decade ago by Szegedy et al. [
2
], demonstrating that DNNs could be easily misled
by seemingly minor modifications to input images. Since then, numerous approaches
for generating adversarial examples have been explored [
3
], highlighting the significant
security concerns surrounding DNN-based systems.
The underlying mechanism for adversarial susceptibility lies in the way DNNs process
images. Rather than learning the actual semantic content of the image, these networks
often rely on superficial or spurious features for classification, as described by Goodfellow
et al. as a “Potemkin village” of features [
4
]. This explains why two images that are visually
indistinguishable from human vision can be classified differently by a DNN, revealing
a vulnerability that adversarial attacks exploit. These attacks often aim to minimize the
perturbations applied to an image so that the changes are not noticeable to the human eye
while still causing a misclassification.
Adversarial attacks on DNNs can be divided into digital and physical attacks. While
digital attacks manipulate image pixels, they often struggle to transfer to the physical
world due to dynamic conditions and deployment challenges. Physical attacks alter real-
world objects’ visual characteristics and pose a threat but are typically invasive, requiring
visible changes that can be easily dismissed and detected by human vision. However,
Sensors 2025,25, 2301 https://doi.org/10.3390/s25072301
Sensors 2025,25, 2301 2 of 15
optical-based physical adversarial attacks are non-invasive and generate perturbations
that mimic natural effects, making them harder to detect and better suited for real-world
applications [
5
]. Despite advancements in imperceptibility, many of these attacks still have
an obvious trace in the physical domain, limiting their effectiveness and feasibility, with
achieving complete invisibility to the human eye remaining an unresolved challenge.
This paper introduces and demonstrates a novel optical-based physical adversarial
attack that leverages the rolling shutter mechanism of CMOS sensors. The proposed attack
is designed to be invisible in the physical domain, ensuring that the attacking light source
remains undetectable to the scene observer. This involves a designed light pulse sequence
spatially transformed during the image acquisition, effectively disrupting the camera’s
image processing to deceive DNNs with a high attack success rate. Furthermore, our
approach does not require precise alignment of the adversarial spatial pattern with the
target object location, offering greater flexibility in real-world scenarios. A successful
invisible attack is achieved when the beam of Attacking MOdulated Light Source (AMOLS)
covers the camera aperture, such that the following are achieved:
1. The peak irradiance is sufficient to dazzle the sensor temporarily;
2. The average irradiance remains below the sensitivity threshold of the human eye.
The following summarizes the primary contributions of this work:
•
We propose a physical domain adversarial attack on DNNs that receive images from
a CMOS camera. The attack involves directing a light source toward the camera;
however, the presence of the projected light is completely unnoticed by observers in
the scene.
•
We introduce an optical attack that is based on dazzling a camera sensor by sending
short pulses. We investigate the effect of the projected pulses on the image captured
by the CMOS camera. We evaluate the irradiance required to attack the image.
•
We explore the relationship between the human eye’s ability to distinguish the at-
tacking light source directed at the camera and the disruption of DNN performance
caused by the influence of the pulsed laser beam. We analyze the photopic conditions
required to ensure that the attacking light source remains invisible to human observers
while still effectively disrupting the acquired image to mislead the classifier model.
•
We evaluate the trade-off between the success of DNN attacks caused by dazzling
pulses and their invisibility to the human eye. Our findings indicate that the duty
cycle of the light source can be adjusted to manage the balance between the attack’s
success rate and the level of concealment required.
•
We present simulated and real experimental results to demonstrate the effectiveness
of our attack.
2. Related Works
While most studies on adversarial attacks have focused on the digital domain, where
perturbations are added to pixel values, growing efforts have expanded into the physical
domain [
6
]. Examples of physical-world attacks typically include using adversarial objects
or imaging system manipulations to fool DNN models. These modifications may include
simple changes, such as adding elements like stickers, eyeglasses, earrings, and others to a
real-world object [
7
], to more complex approaches. The more complex methods typically
involve optical-based techniques [
5
], including temporarily projecting specifically crafted
adversarial perturbations onto target objects [
8
], among others, or strategically illuminating
target objects using infrared light sources [
9
]. Furthermore, synthesizing Three-Dimensional
(3D) adversarial objects has been proposed to confuse classifier models [
10
], and imaging
projection transformation in a 3D physical environment was demonstrated to deceive object
Sensors 2025,25, 2301 3 of 15
detection systems effectively [
11
]. These examples highlight the growing applicability of
adversarial attacks in real-world settings.
Recent studies on physical adversarial examples have increasingly focused on manip-
ulating imaging systems themselves. For instance, Liu et al. [
12
] induced perturbations
in the captured image through an electromagnetic injection attack. They focused on CCD
sensors but noted that CMOS sensors, which have an independent measurement unit for
each pixel, provide greater resilience to electromagnetic interference, making them more
robust against such threats. Additionally, Duan et al. [
13
] employed a laser beam attack to
create spatially tailored perturbations; however, they noted that this approach has a limited
success rate in dynamic conditions. Many physical adversarial attacks require precise
alignment of the adversarial spatial pattern with the target object placement. Moreover,
Liu et al. [
14
] inject their attack after image acquisition, targeting the data lane between the
camera sensor and the endpoint device. This requires physical access to the sensor-enabled
system, which is practically infeasible in certain situations.
In this work, we develop an invisible camera dazzling attack that leverages the rolling
shutter mechanism inherent in CMOS sensors. Unlike the continuous-wave operation
of light sources, where the degree of dazzle on CMOS sensors can be depicted by the
dazzling area or the number of saturated pixels [
15
], temporally modulated light can pro-
duce adjustable stripes in a captured image—introducing a unique approach to injecting
adversarial spatial patterns. The rolling shutter effect is primarily studied in the context of
mitigating distortions caused by fast-moving objects that approach the camera’s scanning
frequency [
16
]. Accordingly, models have been developed to correct these distortions.
Moreover, it was proposed that a smartphone camera can be used for visible light commu-
nications to detect and convert a temporal signal into spatial patterns by exploiting the
rolling shutter effect of CMOS sensors [17].
Adversarial attacks leveraging the rolling shutter mechanism have been introduced in
references [
18
–
22
], where temporally modulated LEDs are used to illuminate a target object,
as shown in Figure 1a. This results in distortions in the acquired image due to the camera’s
row-wise scanning process. The first configuration [
18
] was introduced as a black-box
backdoor attack on face recognition systems, where illuminating the entire scene induces
perturbations employing the rolling shutter effect. While the first two studies [
18
,
19
]
utilize programmable RGB LEDs, resulting in an adversarial signal with three adjustable
components of Red, Green, and Blue, later work [
20
] demonstrated the use of a common
commercial LED with a modulator to control the frequency of the emitted white light. In
addition, further schemes [
21
,
22
] expanded the application of the white light attack method,
showcasing the generalization and transferability of adversarial samples across different
models and tasks, including traffic sign recognition systems and lane detection models.
However, these approaches require comprehensive illumination of the whole scene and
usually fail to remain invisible to the human eye. Despite the light pulse sequence being
designed with a modulation frequency that prevents flickering perceived by the human
eye, the illumination source still appears steady and is not stealthy to the human observer
in the scene.
Here, we propose to employ an AMOLS beam that directly illuminates the camera’s
aperture as shown in Figure 1b, taking advantage of the rolling shutter’s scanning process
to induce real-world adversarial perturbations on the acquired image. Since the pulsed light
beam is directed toward the camera rather than reflecting off a target object (see Figure 1),
the average power requirements are significantly reduced compared to previous methods.
While Kohler et al. [
23
] and Yan et al. [
24
] introduced such a camera attack utilizing a laser
and exploiting the rolling shutter mechanism, their approaches still leave an obvious trace
of the attack in the physical domain and remain visible to the human eye.
Sensors 2025,25, 2301 4 of 15
Sensors2025,25,xFORPEERREVIEW4of16
Figure1.Practicalphysical-worldadversarialaack.Theaackcanbecarriedouteither(a)bytem-
porallymodulatingalightsourcetoilluminatetheentirescene,whichreflectslightpulsesontothe
CMOSsensor,or(b)bydirectingapulsedlaserbeamspecificallyataCMOSsensor.Theredarrows
indicatethepropagationdirectionofthelight.
Here,weproposetoemployanAMOLSbeamthatdirectlyilluminatesthecamera’s
apertureasshowninFigure1b,takingadvantageoftherollingshuer’sscanningprocess
toinducereal-worldadversarialperturbationsontheacquiredimage.Sincethepulsed
lightbeamisdirectedtowardthecameraratherthanreflectingoffatargetobject(seeFig-
ure1),theaveragepowerrequirementsaresignificantlyreducedcomparedtoprevious
methods.WhileKohleretal.[23]andYanetal.[24]introducedsuchacameraaackuti-
lizingalaserandexploitingtherollingshuermechanism,theirapproachesstillleavean
obvioustraceoftheaackinthephysicaldomainandremainvisibletothehumaneye.
Sincetheintegrationtimeofthehumaneyeissignificantlylongerthantheacquisi-
tiontimeofeachrowinarollingshuerscanningprocess,ahigh-frequencymodulated
signalisseenascontinuousbythehumaneye.IfdenotingthedutycycleoftheAMOLS
as𝐷,theintensityperceivedbythehumaneyecanbeexpressedasfollows:
𝐼
𝐷⋅𝐼
.(1)
Thatis,thehumaneyeonlyperceivesthesignal’saveragepower.Consequently,itispos-
sibletocontrolthisintensitybyappropriatelyreducingthedutycycleoftheAMOLS.In
thispaper,weexploretherelationshipbetweentheeffectivenessofadutycycleduringa
directcameraaackandtheabilitytodistinguishtheAMOLSimplementation.First,we
reviewtheeffectofAMOLSonthecameraanddeterminetheirradianceneededtopro-
ducethedesireddisruptiveeffectonthecamera.Next,weevaluatethedazzlingirradi-
anceonthehumaneyeanddeterminetheconditionsthatinfluencetheeye’sabilityto
perceiveandrecognizethelightsource.Finally,afterestablishingtheirradiancerequire-
ments,weexaminetheefficiencyofimagedistortioncausedbythedesignedpulsese-
quenceonawell-knownclassifier,theResidualNeuralNetwork(ResNet50)architecture,
throughsimulationsandexperiments.
3.MaterialsandMethods
3.1.DazzleEffectwithRollingShuer
Camera
Thespatialspreadofapointsourceintheimageplaneisconventionallydescribed
bythediffractionofthePointSparedFunction(PSF),generallygivenbytheFouriertrans-
formationoftheentrancepupil.However,particularlyforbrightpowersources(e.g.,a
Figure 1. Practical physical-world adversarial attack. The attack can be carried out either (a) by
temporally modulating a light source to illuminate the entire scene, which reflects light pulses onto
the CMOS sensor, or (b) by directing a pulsed laser beam specifically at a CMOS sensor. The red
arrows indicate the propagation direction of the light.
Since the integration time of the human eye is significantly longer than the acquisition
time of each row in a rolling shutter scanning process, a high-frequency modulated signal
is seen as continuous by the human eye. If denoting the duty cycle of the AMOLS as
D
, the
intensity perceived by the human eye can be expressed as follows:
Ieye =D·Isource. (1)
That is, the human eye only perceives the signal’s average power. Consequently, it is
possible to control this intensity by appropriately reducing the duty cycle of the AMOLS.
In this paper, we explore the relationship between the effectiveness of a duty cycle during a
direct camera attack and the ability to distinguish the AMOLS implementation. First, we
review the effect of AMOLS on the camera and determine the irradiance needed to produce
the desired disruptive effect on the camera. Next, we evaluate the dazzling irradiance on
the human eye and determine the conditions that influence the eye’s ability to perceive
and recognize the light source. Finally, after establishing the irradiance requirements,
we examine the efficiency of image distortion caused by the designed pulse sequence on
a well-known classifier, the Residual Neural Network (ResNet50) architecture, through
simulations and experiments.
3. Materials and Methods
3.1. Dazzle Effect with Rolling Shutter Camera
The spatial spread of a point source in the image plane is conventionally described
by the diffraction of the Point Spared Function (PSF), generally given by the Fourier
transformation of the entrance pupil. However, particularly for bright power sources
(e.g., a laser source), other effects such as stray light scattering and halo [
25
] may occur in
addition to the PSF diffraction, which may be considerably more significant than the PSF.
The dazzling effect is demonstrated in Figure 2, where the measurement is acquired from a
laptop camera (installed on a DELL-INSPIRON laptop with 0.92 Megapixel, 88
◦
diagonal
viewing angle). The AMOLS average power was 5
mW
with
∼
3.5
mm
spot diameter. As
shown in Figure 2, a notable dazzling effect is observed when utilizing such a power level.
Sensors 2025,25, 2301 5 of 15
Sensors2025,25,xFORPEERREVIEW5of16
lasersource),othereffectssuchasstraylightscaeringandhalo[25]
mayoccurinaddi-
tiontothePSFdiffraction,whichmaybeconsiderablymoresignificantthanthePSF.The
dazzlingeffectisdemonstratedinFigure2,wherethemeasurementisacquiredfroma
laptopcamera(installedonaDELL-INSPIRONlaptopwith0.92Megapixel,88°diagonal
viewingangle).TheAMOLSaveragepowerwas5 mWwith~3.5 mm
spotdiameter.As
showninFigure2,anotabledazzlingeffectisobservedwhenutilizingsuchapowerlevel.
Figure2.ExperimentalPSFmeasurement.Thecamera’sresponsetoplacedpointsourcewithinthe
fieldofview.Theradiantfluxmeasuredintheobjectplaneis
~50 mW/cm
.
Previousstudiesoninfraredimagers[26,27]haveempiricallyshownthatthediame-
terofthesaturatedareaintheimageplane,denotedas𝑥
,canbeapproximatedasfol-
lows:
𝑥
∝
(2)
where𝐼
and𝐼
arethelaserirradianceandthesaturationlevel,respectively.Basedon
resultsforvisiblelightusingaCMOScamera[28,29]aminimumaverageirradianceof
50 mW/cm
duringeachrowexposureisrequired,andatleast0.1 mW/cm
pickirradi-
ancetoachievedazzlingwithshorterpulses.Weexperimentallyfoundthatsimilarcon-
ditionsholdforthecamerausedinthiswork,asobservedinFigure2.
Next,thedazzlingeffectformedintheaackedimageisexamined.Witharolling
shuercamera,everyrowintheframecollectsambientlightduringdifferentperiods.As
showninFigure3,the𝑖-throwofthesensorrecordsthelightintegratedduringtheperiod
from𝑡
𝑡
till𝑡
,whileforthefollowingrow𝑖1,theintegrationtimewillbeuntil
𝑡
𝑡
,where𝑡
denotesthereadingtimeofasinglerowand𝑡
denotestheex-
posuretimeofasinglerow.Thedurationofscanningeachframe,denotedby𝑡
,can
beexpressedasfollows[16]:
𝑡
𝑡
𝑁
𝑁
𝑡
.(3)
where𝑁
and𝑁
arethenumberofpixelrowsandthenumberofhiddenpixelrowsin
eachframe,respectively.
Figure 2. Experimental PSF measurement. The camera’s response to placed point source within the
field of view. The radiant flux measured in the object plane is ~50 mW/cm2.
Previous studies on infrared imagers [
26
,
27
] have empirically shown that the diameter
of the saturated area in the image plane, denoted as xsat , can be approximated as follows:
xsat ∝I0
Isat 1
3(2)
where
I0
and
Isat
are the laser irradiance and the saturation level, respectively. Based
on results for visible light using a CMOS camera [
28
,
29
] a minimum average irradiance
of 50
mW/cm2
during each row exposure is required, and at least 0.1
mW/cm2
pick
irradiance to achieve dazzling with shorter pulses. We experimentally found that similar
conditions hold for the camera used in this work, as observed in Figure 2.
Next, the dazzling effect formed in the attacked image is examined. With a rolling
shutter camera, every row in the frame collects ambient light during different periods.
As shown in Figure 3, the
i
-th row of the sensor records the light integrated during the
period from
ti−tex p
till
ti
, while for the following row
i+
1, the integration time will be
until
ti+tread
, where
tread
denotes the reading time of a single row and
tex p
denotes the
exposure time of a single row. The duration of scanning each frame, denoted by
tf ram e
, can
be expressed as follows [16]:
tf ram e =tread(Nr+NrH)+tex p. (3)
where
Nr
and
NrH
are the number of pixel rows and the number of hidden pixel rows in
each frame, respectively.
Sensors2025,25,xFORPEERREVIEW6of16
Figure3.AschematicillustrationoftherollingshuereffectcausedbydazzlingAMOLS.Theroll-
ingshuermechanismtransformsthetemporalsignalwithadesignedsequenceoflaserpulses
(markedinblueatthetop)intospatialdistortion.Thisdistortionoccursduringdifferentperiodsof
readingandexposureforthepixelrowsintheframe(indicatedbywhiteandgrayblocks).Asa
result,astripe-likepaernemergesintheacquiredimage(right).
Theratio𝑅
𝑡
/𝑡
determinesthenumberofexposedrowsatanygiventime
(seeFigure4).Thus,𝑅
isreferredtoastherow’sexposureconstant.Itisworthhigh-
lightingthatifthepulsedurationgeneratedbytheAMOLSisshorterthan𝑡
,exactly
𝑅
rowswillbedazzled,regardlessofthepulsewidth.Forinstance,bothpulseswitha
duration1 μsand2 μswillproducethesamepaernwhenusingatypicalcamerawith
areadingtimeof𝑡
30 μs.Theexperimentallyobtaineddazzlepaernfortherolling
shuersensorwhentheAMOLSisappliedisshowninFigure4,alongwiththesimulated
stripe-linepaernutilizing𝑅
37obtainedwithacalibrationprocess.Thesimulation
resultcorrespondswellwiththeexperimentalmeasurement,withastructuralsimilarity
of93%.
Figure4.Dazzleeffectofrollingshuersensorbyamodulatedlightsource.(a,b)Theresulting
dazzlepaernforAMOLSvia(a)experimentand(b)simulationwith𝑅
37.
3.2.PhotopicConditionsforInvisibility
Thissectionfocusesondeterminingthephotometricconditionsrequiredtokeepthe
AMOLSeffectivelyinvisible.TheaackscenarioisdepictedinFigure5,whereatarget
object(car)andtheAMOLSareplacedinfrontofacamerawhileanobserverisnearthe
cameraatanangle𝜃relativetotheopticalaxis.TheacquiredimageisthenfedtoDNN
toclassifythetarget.ConsiderthattheAMOLSpowerissettoproduceanirradianceof
𝑒50 mW/cm
atthesensorplanewhenactive.Theaveragepower𝐸receivedbythe
observer’seyesfromtheAMOLSisinfluencedbythedutycycleoftheAMOLSduring
Figure 3. A schematic illustration of the rolling shutter effect caused by dazzling AMOLS. The rolling
shutter mechanism transforms the temporal signal with a designed sequence of laser pulses (marked
in blue at the top) into spatial distortion. This distortion occurs during different periods of reading
and exposure for the pixel rows in the frame (indicated by white and gray blocks). As a result, a
stripe-like pattern emerges in the acquired image (right).
Sensors 2025,25, 2301 6 of 15
The ratio
Rn=tex p/tread
determines the number of exposed rows at any given time
(see Figure 4). Thus,
Rn
is referred to as the row’s exposure constant. It is worth highlighting
that if the pulse duration generated by the AMOLS is shorter than
tread
, exactly
Rn
rows
will be dazzled, regardless of the pulse width. For instance, both pulses with a duration
1
µs
and 2
µs
will produce the same pattern when using a typical camera with a reading
time of
tread ≈
30
µs
. The experimentally obtained dazzle pattern for the rolling shutter
sensor when the AMOLS is applied is shown in Figure 4, along with the simulated stripe-
line pattern utilizing
Rn=
37 obtained with a calibration process. The simulation result
corresponds well with the experimental measurement, with a structural similarity of 93%.
Sensors2025,25,xFORPEERREVIEW6of16
Figure3.AschematicillustrationoftherollingshuereffectcausedbydazzlingAMOLS.Theroll-
ingshuermechanismtransformsthetemporalsignalwithadesignedsequenceoflaserpulses
(markedinblueatthetop)intospatialdistortion.Thisdistortionoccursduringdifferentperiodsof
readingandexposureforthepixelrowsintheframe(indicatedbywhiteandgrayblocks).Asa
result,astripe-likepaernemergesintheacquiredimage(right).
Theratio𝑅
𝑡
/𝑡
determinesthenumberofexposedrowsatanygiventime
(seeFigure4).Thus,𝑅
isreferredtoastherow’sexposureconstant.Itisworthhigh-
lightingthatifthepulsedurationgeneratedbytheAMOLSisshorterthan𝑡
,exactly
𝑅
rowswillbedazzled,regardlessofthepulsewidth.Forinstance,bothpulseswitha
duration1 μsand2 μswillproducethesamepaernwhenusingatypicalcamerawith
areadingtimeof𝑡
30 μs.Theexperimentallyobtaineddazzlepaernfortherolling
shuersensorwhentheAMOLSisappliedisshowninFigure4,alongwiththesimulated
stripe-linepaernutilizing𝑅
37obtainedwithacalibrationprocess.Thesimulation
resultcorrespondswellwiththeexperimentalmeasurement,withastructuralsimilarity
of93%.
Figure4.Dazzleeffectofrollingshuersensorbyamodulatedlightsource.(a,b)Theresulting
dazzlepaernforAMOLSvia(a)experimentand(b)simulationwith𝑅
37.
3.2.PhotopicConditionsforInvisibility
Thissectionfocusesondeterminingthephotometricconditionsrequiredtokeepthe
AMOLSeffectivelyinvisible.TheaackscenarioisdepictedinFigure5,whereatarget
object(car)andtheAMOLSareplacedinfrontofacamerawhileanobserverisnearthe
cameraatanangle𝜃relativetotheopticalaxis.TheacquiredimageisthenfedtoDNN
toclassifythetarget.ConsiderthattheAMOLSpowerissettoproduceanirradianceof
𝑒50 mW/cm
atthesensorplanewhenactive.Theaveragepower𝐸receivedbythe
observer’seyesfromtheAMOLSisinfluencedbythedutycycleoftheAMOLSduring
Figure 4. Dazzle effect of rolling shutter sensor by a modulated light source. (a,b) The resulting
dazzle pattern for AMOLS via (a) experiment and (b) simulation with Rn=37.
3.2. Photopic Conditions for Invisibility
This section focuses on determining the photometric conditions required to keep the
AMOLS effectively invisible. The attack scenario is depicted in Figure 5, where a target
object (car) and the AMOLS are placed in front of a camera while an observer is near the
camera at an angle
θ
relative to the optical axis. The acquired image is then fed to DNN
to classify the target. Consider that the AMOLS power is set to produce an irradiance of
e=
50
mW/cm2
at the sensor plane when active. The average power
E
received by the
observer’s eyes from the AMOLS is influenced by the duty cycle of the AMOLS during the
frame exposure period. The light source duty cycle denoted by
D
determines the average
power of the light source, which can be expressed as
E=e·D
. In addition, assuming the
AMOLS is smaller than the human eye’s angular resolution, the strictest condition would
be the concentration of the seen power from any given source. Thus, a larger angular extent
covered by the AMOLS would yield a lower peak power.
Sensors2025,25,xFORPEERREVIEW7of16
theframeexposureperiod.Thelightsourcedutycycledenotedby𝐷determinestheav-
eragepowerofthelightsource,whichcanbeexpressedas𝐸𝑒⋅𝐷.Inaddition,assum-
ingtheAMOLSissmallerthanthehumaneye’sangularresolution,thestrictestcondition
wouldbetheconcentrationoftheseenpowerfromanygivensource.Thus,alargeran-
gularextentcoveredbytheAMOLSwouldyieldalowerpeakpower.
Figure5.InvisibleAMOLSimplementationfordirectcameraaack.Atargetobject(e.g.,acar)is
placedinthecamera’sfieldofview,andalightsourcedirectlyilluminatesthecamera(bysending
abeambetweentheredarrows).ThetaskoftheDNNistoclassifytheacquiredimage.Whenap-
plyingtheAMOLS,itmustremaininvisibletoanobserveratanangle𝜃relativetotheopticalaxis.
Bydenotingthebackgroundbrightnessby𝐿
andtheAMOLSbrightnessby𝐿
,the
contrastcanbegivenbythefollowing:
𝐶
, (4)
PreviousstudiesbyH.R.Blackwell[30]andW.Adrian[31]investigatedthethresh-
oldcontrast𝐶
requiredtodetectanobject.AccordingtoW.Adrian,atargetcontrastof
1atasmallangleissufficienttorecognizethetarget.Sinceradianceisaphysicalquantity
conservedthroughoutanopticalsystem,itdictatesthebrightness.Whenthesolidangle
coveredbythetargetissmallerthanthesystem’sresolvingpower,theAMOLSbrightness
hasthefollowingform[32]:
𝐿
𝐸∙683 ∙𝑉
∙Ω
cd ∙m
,(5)
where𝑉
denotesthephotopicefficacyandΩ
istheresolvingpowerofthehuman
eye(representingthestrictestconditionregardingthereceivedpower).Employingacam-
eramodeltorepresenttheeyemodel,C.A.WilliamsonandL.N.McLin[33,34]proposed
ascaeringfunctionbasedonempiricalfindingsbyJ.Vosetal.[35],withaneffective
solidanglecollectedbytheeye:
𝑓
𝜃,𝐴,𝑝,𝐿
𝑆∙𝐿
∙𝑔
𝜃,𝐴,𝑝 sr
,(6)
where𝑔
canbedeterminedbytheoff-axisangle𝜃(seeFigure5),theageA(inyears),
andtheeyepigment𝑝,whichisgivenbythefollowing:
𝑔
𝜃,𝐴,𝑝
.
1
.
0.0025𝑝 sr
,(7)
Substitutingthetermoftheaveragepower𝐸,andtheangularresolutionby𝑓
,the
AMOLSbrightnessexpressedinEquation(5)takesthefollowingform:
𝐿
𝑒∙𝐷 ∙683 ∙𝑉
∙
𝑓
cd ∙m
,(8)
Figure 5. Invisible AMOLS implementation for direct camera attack. A target object (e.g., a car) is
placed in the camera’s field of view, and a light source directly illuminates the camera (by sending a
beam between the red arrows). The task of the DNN is to classify the acquired image. When applying
the AMOLS, it must remain invisible to an observer at an angle θrelative to the optical axis.
Sensors 2025,25, 2301 7 of 15
By denoting the background brightness by
Lb
and the AMOLS brightness by
LAS
, the
contrast can be given by the following:
C=LAS −Lb
Lb
, (4)
Previous studies by H.R. Blackwell [
30
] and W. Adrian [
31
] investigated the threshold
contrast
Cthr
required to detect an object. According to W. Adrian, a target contrast of 1
at a small angle is sufficient to recognize the target. Since radiance is a physical quantity
conserved throughout an optical system, it dictates the brightness. When the solid angle
covered by the target is smaller than the system’s resolving power, the AMOLS brightness
has the following form [32]:
LAS =E·683·Vλ·Ω−2
eye hcd·m−2i, (5)
where
Vλ
denotes the photopic efficacy and
Ωeye
is the resolving power of the human eye
(representing the strictest condition regarding the received power). Employing a camera
model to represent the eye model, C.A. Williamson and L.N. McLin [
33
,
34
] proposed a
scattering function based on empirical findings by J. Vos et al. [
35
], with an effective solid
angle collected by the eye:
feye(θ,A,p,Lb)=S·LT
b·geye(θ,A,p)hsr−1i, (6)
where
geye
can be determined by the off-axis angle
θ
(see Figure 5), the age A(in years), and
the eye pigment p, which is given by the following:
geye(θ,A,p)=10
θ3+5
θ2+0.1p
θ"1+A
62.52#+0.0025p[sr−1], (7)
Substituting the term of the average power
E
, and the angular resolution by
feye
, the
AMOLS brightness expressed in Equation (5) takes the following form:
LAS =e·D·683·Vλ·feye hcd·m−2i, (8)
Finally, by substituting Equation (8) into Equation (4), the light source duty cycle can
be expressed by the following:
D=L1−T
b
Cthr (Lb)+1
e·683·Vλ·S·geye(θ,A,p). (9)
Figure 6shows the light source duty cycle
D
required for dazzling as a function of the
background illumination for various viewing aspect angles. As the aspect angle increases,
the effective radiance on the retina decreases. Consequently, the contrast decreases with
the increasing background brightness, requiring more power to exceed the threshold. It is
observed from the results shown in Figure 6that for observers placed at an angle greater
than 10 degrees, a duty cycle of 0.5% is sufficient to keep the source invisible, regardless of
the background illumination level. The following sections will present a technique that can
be operated even at lower duty cycle percentages.
Sensors 2025,25, 2301 8 of 15
Sensors2025,25,xFORPEERREVIEW8of16
Finally,bysubstitutingEquation(8)intoEquation(4),thelightsourcedutycyclecan
beexpressedbythefollowing:
𝐷𝐿
∙∙
∙∙
,,
. (9)
Figure6showsthelightsourcedutycycle𝐷requiredfordazzlingasafunctionof
thebackgroundilluminationforvariousviewingaspectangles.Astheaspectanglein-
creases,theeffectiveradianceontheretinadecreases.Consequently,thecontrastde-
creaseswiththeincreasingbackgroundbrightness,requiringmorepowertoexceedthe
threshold.ItisobservedfromtheresultsshowninFigure6thatforobserversplacedatan
anglegreaterthan10degrees,adutycycleof0.5%issufficienttokeepthesourceinvisi-
ble,regardlessofthebackgroundilluminationlevel.Thefollowingsectionswillpresenta
techniquethatcanbeoperatedevenatlowerdutycyclepercentages.
Figure6.Thedutycycle𝐷oftheAMOLSatthethresholdofhumandiscriminationasafunction
ofthebackgroundluminanceforvariousviewingangles𝜃.
3.3.GeneratingthePhysicalAdversarialAack
Followingtheformalismin[2],theproblemoffindinganadversarialexamplecanbe
formallydefinedasfollows:
minimize |𝑥
𝑥|
s. t. 𝐶𝑥
𝑙
𝑥
∈0, 1
,
(10)
where𝑥 istheundistributedimage,𝑥′ istheperturbatedimage,𝑙 representsthe
groundtruthlabeloftheimage𝑥,and𝐶𝑥denotestheDNNusedasaclassifier.Peri-
odically,solvingsuchaproblemcanbeincrediblycomplex,whichleadstosolvingamore
straightforwardprobleminstead,assuggestedin[36].Inbrief,thegoalistofindasmall
perturbation𝛿 𝑥
𝑥,whichcanbeappliedtoanimage𝑥 toalteritsclassification
whileensuringthattheresultingimageremainsvalid.ConsideringthattheSoftmax𝑉
isappliedontopoftheDNNlogits,thelossfunctionmappinganimage𝑥toapositive
realnumbercanbedescribedasfollows
𝑓
𝑥 𝐿𝑜𝑠𝑠
,
𝑉
𝑥,(11)
Accordingly,insteadofformulatingtheconstraintminimizationproblemasinEquation
(10),onecanuseanalternativeformulationandsolvethefollowingproblem:
Figure 6. The duty cycle
D
of the AMOLS at the threshold of human discrimination as a function of
the background luminance for various viewing angles θ.
3.3. Generating the Physical Adversarial Attack
Following the formalism in [
2
], the problem of finding an adversarial example can be
formally defined as follows:
minimize ||x′−x||22
s.t. C(x′)=l
x′∈[0, 1]n,
(10)
where
x
is the undistributed image,
x′
is the perturbated image,
l
represents the ground
truth label of the image
x
, and
C(x)
denotes the DNN used as a classifier. Periodically,
solving such a problem can be incredibly complex, which leads to solving a more straightfor-
ward problem instead, as suggested in [
36
]. In brief, the goal is to find a small perturbation
δ=x′−x
, which can be applied to an image
x
to alter its classification while ensuring that
the resulting image remains valid. Considering that the Softmax
Vn
is applied on top of
the DNN logits, the loss function mapping an image
x
to a positive real number can be
described as follows
f(x)=LossC,l(Vn(x)), (11)
Accordingly, instead of formulating the constraint minimization problem as in Equation
(10), one can use an alternative formulation and solve the following problem:
minimize ||δ||0−α·f(x+δ)
s.t. x+δ∈[0, 1]n.(12)
where
α
represents the ratio between the magnitude of the disturbance and its effect’s
intensity on the output, and ∥ · ∥0denotes the zero norm.
In our case, we aim to establish a relation between the pulsed laser activity and the
resulting adversarial perturbation caused by the rolling shutter mechanism of the CMOS
camera. This mechanism converts the temporal signal of the designed laser pulse se-
quence into a spatial distortion within the acquired image.
Ee f f
is an
N
-dimensional
binary row vector representing the pulsed laser activity, which can be expressed by
N=(Nr+NrH)/Rn
, where
Rn
denotes the number of dazzled pixel rows by each pulse
and
(Nr+NrH)
indicates the sensor’s total number of pixel rows (see Section 3.1). Specifi-
cally, a unit value at the
i
-th component of this vector
Ee f f [i]=
1, indicates a pulse occurring
Sensors 2025,25, 2301 9 of 15
at the time
t=i·tf ram e/N
, and dazzles the sensor’s pixel rows from
i·Rn
to
(i+1)·Rn
.
Thus, the indices of the dazzled pixel rows in the acquired image can be obtained by
substituting each unit entry of the pulsed laser activity vector
ET
e f f
with a size
Rn
vector of
ones, which is given by the following:
ET
r=ET
e f f O1T
Rn(13)
where
⊗
is the Kronecker product and
1T
Rn
is an
Rn
-dimensional column vector of ones.
Consequently,
ET
r
is an
N·Rn
-dimensional binary column vector in which unit entries
indicate the dazzled sensor’s pixel rows. Next, the resulting dazzle pattern in the acquired
N×Mimage (e.g., Figure 4) can be obtained by the following:
δ=ET
rO1M=ET
e f f O1T
RnO1M, (14)
where
1M
is a size
M
vector of ones corresponding to the number of pixel columns in the
acquired image. Instead of formulating the minimization problem following Equation (12),
we now use an alternative formulation expressed in terms of the pulsed laser activity vector
Ee f f —the problem then becomes as follows: given x, find δthat satisfies the following:
minimize ET
e f f 0−α·f(x+δ)
s.t. δ∈[0, 1]n,(15)
In practice, to implement a typical gradient-based optimization algorithm (such as
SGD or ADAM) for solving Equation (15), we replace the binary vector derived from
Equation (14). Rather than optimizing over the variable
δ
defined above, we change the
variables and optimize over ωT, which has the following form:
δ=1
2tanhωT+1O1T
RnO1M, (16)
where δ∈[0, 1]n, and ωThas the same dimensions as ET
e f f .
Since the exact moment of camera exposure is unknown to the attacker in a real-world
setting, applying the AMOLS, consisting of a designed sequence of laser pulses, yields a
dazzling pattern with a random horizontal shift. Considering the asynchrony between
the attacking light pulse sequence and the camera’s exposure moment, we utilize the
Expectation over Transformation (EoT) method [10] as follows:
minimize Et0∼TnET
e f f 0−α·f(x+δ)o. (17)
where Tis the space of all possible instances of frame exposure, denoted as t0.
4. Results and Discussion
This section presents the feasibility of conducting invisible adversarial attacks on
DNNs in the physical domain by dazzling the camera. In addition, we evaluate the
AMOLS performance using optimal dazzle patterns following the method described in
Section 3.3, considering the pulsed laser activity depicted in Section 3.1. In the following
sections, we employ both simulations and real experiments. First, we conduct simulations
to investigate the effect of the AMOLS duty cycle while maintaining a constant pulse width.
Next, we optically demonstrate the attack and examine its sensitivity to the pulse width.
Sensors 2025,25, 2301 10 of 15
4.1. Effectiveness of the AMOLS
We evaluate the effectiveness of the AMOLS based on the duty cycle of a pulsed
laser (as discussed in Section 3.2) while keeping a constant pulse width. The ResNet50
classifier [
37
] and the standard cross-entropy loss function are utilized to simulate the
adversarial attacks on the image classification model. Figure 7shows simulation results
of the loss function depending on the AMOLS duty cycle for two cases: where an object
covers (1) approximately 40% of the field of view (FOV), and (2) approximately 85% of the
FOV. These results focus on the “Coffee mug” as the target object, with the highest obtained
values for each examined duty cycle as a result of optimizing the attack (as discussed in
Section 3.3). It can be empirically determined that loss function values exceeding 2 exhibit
poor classifier performances, resulting in misclassification across a significant number
of input images—specifically, this enhances the effectiveness of the AMOLS. The results
presented in Figure 7indicate that when the duty cycle is set lower than 0.2%, the attack
remains feasible—yet the classification model tends to yield better results when the target
object covers
∼
40% of the FOV. Conversely, increasing the AMOLS duty cycle substantially
raises the loss, thereby enhancing the effectiveness of the attack in the case of an object
occupying
∼
40% of the FOV. Additionally, for a target object that covers
∼
85% of the
FOV, the attack proves effective across the entire duty cycle range examined, with a milder
dependence on changes in the AMOLS duty cycle.
Sensors2025,25,xFORPEERREVIEW10of16
4.ResultsandDiscussion
Thissectionpresentsthefeasibilityofconductinginvisibleadversarialaackson
DNNsinthephysicaldomainbydazzlingthecamera.Inaddition,weevaluatethe
AMOLSperformanceusingoptimaldazzlepaernsfollowingthemethoddescribedin
Section3.3,consideringthepulsedlaseractivitydepictedinSection3.1.Inthefollowing
sections,weemploybothsimulationsandrealexperiments.First,weconductsimulations
toinvestigatetheeffectoftheAMOLSdutycyclewhilemaintainingaconstantpulse
width.Next,weopticallydemonstratetheaackandexamineitssensitivitytothepulse
width.
4.1.EffectivenessoftheAMOLS
WeevaluatetheeffectivenessoftheAMOLSbasedonthedutycycleofapulsedlaser
(asdiscussedinSection3.2)whilekeepingaconstantpulsewidth.TheResNet50classifier
[37]andthestandardcross-entropylossfunctionareutilizedtosimulatetheadversarial
aacksontheimageclassificationmodel.Figure7showssimulationresultsoftheloss
functiondependingontheAMOLSdutycyclefortwocases:whereanobjectcovers(1)
approximately40%ofthefieldofview(FOV),and(2)approximately85%oftheFOV.
Theseresultsfocusonthe“Coffeemug”asthetargetobject,withthehighestobtained
valuesforeachexamineddutycycleasaresultofoptimizingtheaack(asdiscussedin
Section3.3).Itcanbeempiricallydeterminedthatlossfunctionvaluesexceeding2exhibit
poorclassifierperformances,resultinginmisclassificationacrossasignificantnumberof
inputimages—specifically,thisenhancestheeffectivenessoftheAMOLS.Theresultspre-
sentedinFigure7indicatethatwhenthedutycycleissetlowerthan0.2%,theaack
remainsfeasible—yettheclassificationmodeltendstoyieldbeerresultswhenthetarget
objectcovers~40%oftheFOV.Conversely,increasingtheAMOLSdutycyclesubstan-
tiallyraisestheloss,therebyenhancingtheeffectivenessoftheaackinthecaseofan
objectoccupying~40%oftheFOV.Additionally,foratargetobjectthatcovers~85%
oftheFOV,theaackproveseffectiveacrosstheentiredutycyclerangeexamined,with
amilderdependenceonchangesintheAMOLSdutycycle.
Figure7.Theeffectivenessoftheproposedaackonthelossfunctionanditsdependencyonthe
dutycycle𝐷ofthepulsedlaserbeam.
Inaddition,weexaminedarangeoftargetobjectsduringtheaack,imagedfrom
variousanglesofviewcorrespondingtodifferentclasses—severalsamplesareshownin
Figure8a.AnanalysisoftheeffectoftheAMOLSdutycycle,whilemaintainingconstant
pulsewidth,ontheclassifier’slossfunctionacrossdiverseinputimagesisshownin
Figure 7. The effectiveness of the proposed attack on the loss function and its dependency on the
duty cycle Dof the pulsed laser beam.
In addition, we examined a range of target objects during the attack, imaged from
various angles of view corresponding to different classes—several samples are shown in
Figure 8a. An analysis of the effect of the AMOLS duty cycle, while maintaining constant
pulse width, on the classifier’s loss function across diverse input images is shown in
Figure 8b. We empirically found the critical values of the cross-entropy loss function at
which the DNN begins to misclassify objects across different classes, considering an offset
in the obtained loss curves above these critical values. It is observed from the results shown
in Figure 8b that an AMOLS duty cycle of 0.4%, which corresponds to a designed sequence
of 4 laser pulses, successfully fools the classifier in all cases.
Sensors 2025,25, 2301 11 of 15
Sensors2025,25,xFORPEERREVIEW11of16
Figure8b.Weempiricallyfoundthecriticalvaluesofthecross-entropylossfunctionat
whichtheDNNbeginstomisclassifyobjectsacrossdifferentclasses,consideringanoffset
intheobtainedlosscurvesabovethesecriticalvalues.Itisobservedfromtheresults
showninFigure8bthatanAMOLSdutycycleof0.4%,whichcorrespondstoadesigned
sequenceof4laserpulses,successfullyfoolstheclassifierinallcases.
(a)(b)
Figure8.TheAMOLSisappliedtodifferentobjects.(a)Examplesofaackedimages.(b)Thede-
pendenceofthelossfunctionontheaackinglightsourcedutycycleforvariousobjects.
4.2.RealExperimentsonPhysical-WorldAdversarialAack
Wecarriedoutrealexperimentstoevaluatethefeasibilityoftheproposedoptical-
basedphysicaladversarialaacksinreal-worldscenariosformedbyconvertingthelight
temporalsignaltoaspatialdistortionwithintheacquiredimage.Acoffeemugisusedas
thetargetobjectandplacedinsidetheFOVofalaptopcamera(installedonaDELL-IN-
SPIRONlaptopwith0.92Megapixel,88° diagonalviewingangle).Fortheaack,a
pulsedlaserbeamisdirectedatthecamerafromapositionadjacenttotheobject,pro-
ducedfroma650nmdotdiodelaser,withanaveragepowerof5mWandaspotsizeof
3.5mm.Asequenceofpulsesisdesignedtogeneratetheadversarialdazzlepaernfol-
lowingtheoptimizationmethoddescribedinSection3.2,wherethetemporalmodulating
signalisproducedutilizingtheArduino-Unomicrocontrollerboard.Thecameracaptures
boththelightreflectedfromtheobjectandthelightemiedbytheAMOLS.Theacquired
imagesarethenfedtotheDNNforclassification.Weconductedourexperimentswithno
ambientlight,asthisrepresentsthemostchallengingconditionforourproblemseing,
whichrequiresthelightsourcetoremaininvisibletoahumanobserver.Asillustratedin
Figure6,asthebackgroundilluminationdecreases,theallowableAMOLSillumination
budgetthatcanremaininvisiblealsodecreases.Conversely,alowerAMOLSillumination
budgetchallengesthesuccessofaacks,asindicatedbythereducedclassificationloss
showninFigure8b.
Figure9a,bshowstwooptical-basedphysicaladversarialexamplesandtheircorre-
spondingpredictionsfromtheimageclassificationmodel.Theseexamplesweregener-
atedfromtwoseparateexposureshots,wheretheAMOLSuseddifferentpulsewidths.It
isworthmentioningthattheaackinglightpulsesequenceisnotsynchronizedwiththe
camera’sexposuremoment(seeSection3.3),leadingtovariationsinthedazzlepaern
acrosseachframe,specificallyintroducingahorizontalshift.Videosshowingthefootage
fromtheaackedcamerasequenceareprovidedintheSupplementaryMaterials.Addi-
tionalexamplescanbefoundontheGitHubrepositoryassociatedwiththispaperat
hps://github.com/ZviSteinOpt/RollingShuerAack/tree/main(accessedon1April
2025).TheinvisibleCMOScameradazzlingaackinducesmisclassificationacrossthe
Figure 8. The AMOLS is applied to different objects. (a) Examples of attacked images. (b) The
dependence of the loss function on the attacking light source duty cycle for various objects.
4.2. Real Experiments on Physical-World Adversarial Attack
We carried out real experiments to evaluate the feasibility of the proposed optical-
based physical adversarial attacks in real-world scenarios formed by converting the light
temporal signal to a spatial distortion within the acquired image. A coffee mug is used
as the target object and placed inside the FOV of a laptop camera (installed on a DELL-
INSPIRON laptop with 0.92 Megapixel, 88
◦
diagonal viewing angle). For the attack, a
pulsed laser beam is directed at the camera from a position adjacent to the object, produced
from a 650 nm dot diode laser, with an average power of 5 mW and a spot size of 3.5 mm.
A sequence of pulses is designed to generate the adversarial dazzle pattern following the
optimization method described in Section 3.2, where the temporal modulating signal is
produced utilizing the Arduino-Uno microcontroller board. The camera captures both the
light reflected from the object and the light emitted by the AMOLS. The acquired images
are then fed to the DNN for classification. We conducted our experiments with no ambient
light, as this represents the most challenging condition for our problem setting, which
requires the light source to remain invisible to a human observer. As illustrated in Figure 6,
as the background illumination decreases, the allowable AMOLS illumination budget that
can remain invisible also decreases. Conversely, a lower AMOLS illumination budget
challenges the success of attacks, as indicated by the reduced classification loss shown in
Figure 8b.
Figure 9a,b shows two optical-based physical adversarial examples and their corre-
sponding predictions from the image classification model. These examples were generated
from two separate exposure shots, where the AMOLS used different pulse widths. It
is worth mentioning that the attacking light pulse sequence is not synchronized with
the camera’s exposure moment (see Section 3.3), leading to variations in the dazzle pat-
tern across each frame, specifically introducing a horizontal shift. Videos showing the
footage from the attacked camera sequence are provided in the Supplementary Materi-
als. Additional examples can be found on the GitHub repository associated with this
paper at https://github.com/ZviSteinOpt/RollingShutterAttack/tree/main (accessed on
1 April 2025).
The invisible CMOS camera dazzling attack induces misclassification across
the input images, significantly reducing the classifier’s confidence in the correct 500th class,
as shown in Figure 9c.
Sensors 2025,25, 2301 12 of 15
Sensors2025,25,xFORPEERREVIEW12of16
inputimages,significantlyreducingtheclassifier’sconfidenceinthecorrect500thclass,
asshowninFigure9c.
Figure9.ResultsofAMOLSrealizationonanimageclassificationmodel.Physical-worldadversar-
ialexamplesgeneratedviatwoshotsrecordingwhenseingdifferentAMOLSactivities:(a)pulse
widthof1 μswith𝐷0.01%,and(b)pulsewidthof70 μswith𝐷0.85%.(c)TheDNN’scon-
fidenceinthepredictedresultsacrossthe1000classesitwastrainedon,withtheindexforthecor-
rect“coffeemug”labelbeing#500.
ThedistributionofpredictionsmadebyatargetedDNNmodelacrossvariousclasses
duringtheoptical-basedphysicaladversarialaacksisdepictedinFigure10.Itisbased
on254repeatedtrials,wheretheAMOLSoperatesfourpulses,havingapulsedurationof
1 μs.Theresultsindicatethatthedesignedaackachievedan85%successrateunder
theseconditions.TheresultsshowninFigure11indicatethatahigheraacksuccessrate
canbeachievedbyincreasingthepulsewidth.WhentheAMOLSpulsewidthexceeds
approximately70 μs,thephysical-worldaacksuccessrateapproaches98%.However,
followingSection3.2,increasingthepulsewidthreducestherangeofconcealedviewing
angles(seeFigure6).Theseexhibitatradeoffbetweentheangularrealmachievinginvis-
ibilityandthesuccessrateofthephysical-worldaackastheAMOLSdutycyclevaries.
Consideringthatthecameracaptures30framespersecond,apulseof1 μscorresponds
toalowdutycycleof0.012%(𝐷100 ∙4∙1 μs∙30 s
0.012%),whereas
apulsedura-
tionof 70 μsresultsinahigherdutycycleof0.84%.Itcanbeobservedfromtheresults
showninFigure6thatseingadutycycleof0.01%ensurestheAMOLSactivityremains
invisibletotheobserverlocatedatanglesgreaterthanapproximately5°fromtheoptical
axis.Incomparison,adutycycleof0.85%couldbesufficienttomaintaintheinvisibility
ofoptical-basedphysicaladversarialaacksataviewingangleof15°.
TheperformanceandpropertiesofouraackaresummarizedinTableA1inAppen-
dixA,togetherwithacomparisontothatofotherphysicaladversarialaacksinvolving
imagesensors.
Figure 9. Results of AMOLS realization on an image classification model. Physical-world adversarial
examples generated via two shots recording when setting different AMOLS activities: (a) pulse width
of 1
µs
with
D=
0.01%, and (b) pulse width of 70
µs
with
D=
0.85%. (c) The DNN’s confidence in
the predicted results across the 1000 classes it was trained on, with the index for the correct “coffee
mug” label being #500.
The distribution of predictions made by a targeted DNN model across various classes
during the optical-based physical adversarial attacks is depicted in Figure 10. It is based
on 254 repeated trials, where the AMOLS operates four pulses, having a pulse duration
of 1
µs
. The results indicate that the designed attack achieved an 85% success rate under
these conditions. The results shown in Figure 11 indicate that a higher attack success
rate can be achieved by increasing the pulse width. When the AMOLS pulse width
exceeds approximately 70
µs
, the physical-world attack success rate approaches 98%.
However, following Section 3.2, increasing the pulse width reduces the range of concealed
viewing angles (see Figure 6). These exhibit a tradeoff between the angular realm achieving
invisibility and the success rate of the physical-world attack as the AMOLS duty cycle varies.
Considering that the camera captures 30 frames per second, a pulse of 1
µs
corresponds to
a low duty cycle of 0.012% (
D=
100
·
4
·
1
µs·
30
s−1=
0.012
%
), whereas a pulse duration of
70
µs
results in a higher duty cycle of 0.84
%
. It can be observed from the results shown in
Figure 6that setting a duty cycle of 0.01% ensures the AMOLS activity remains invisible
to the observer located at angles greater than approximately 5
◦
from the optical axis.
In comparison, a duty cycle of 0.85% could be sufficient to maintain the invisibility of
optical-based physical adversarial attacks at a viewing angle of 15◦.
Sensors2025,25,xFORPEERREVIEW13of16
Figure10.ThefrequencydistributionoftheDNNpredictionsduringtheaack.Whiletheobject’s
correctlabelisa“coffeemug”,theaackexhibitsanaacksuccessrateof85%.
Figure11.TheaverageaacksuccessrateasafunctionoftheAMOLSpulsewidth.
5.Conclusions
Insummary,weintroducedanovelmethodforconductingoptical-basedphysical
adversarialaacksonDNN.Theaackisdemonstratedbydirectingapulsedlightata
CMOScamera.Therollingshuermechanismofthecameraconvertsthetemporalsignal,
whichconsistsofthedesignedsequenceoflightpulses,intoaspatialdistortionwithinthe
physical-worldadversarialimage.Thephotometricconditionsandlightpulsecharacter-
isticsareanalyzedtodazzletheCMOScamerasufficiently,therebyfoolingtheDNN
modelwhilekeepingtheAMOLSactivityinvisibletoobserversintheenvironment.
Wedemonstratedthatthelightsourcedutycycleenablesthecontrolofthetradeoff
betweentheaack’ssuccessrateandtherequiredangulardegreeofconcealment.For
instance,withtheproposedmethod,an85%successrateforthephysical-worldaackcan
beachievedwhileensuringtheinvisibilityoflightsourceactivitytotheobserverexcept
foranarrowangularrangeof5°fromtheopticalaxis.However,theaacksuccessrate
couldbeincreasedto98%byallowingaslightreductionof10°intheangularconceal-
mentrange.
SupplementaryMaterials:Thefollowingsupportinginformationcanbedownloadedat:
www.mdpi.com/xxx/s1.Thefollowingsupportingvideosshowthefootagefromtheaackedcam-
erasequencewhenseingdifferentAMOLSactivities.VideoS1:pulsewidthof1 μs with𝐷
0.01%,andVideoS2:pulsewidthof70 μswith𝐷0.85%.
Chain
Coffe mug
Cup
Digital cloack
Hodometer
Knot
Ruler
Website
Class
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
0.45
0.5
Frequency [%]
Figure 10. The frequency distribution of the DNN predictions during the attack. While the object’s
correct label is a “coffee mug”, the attack exhibits an attack success rate of 85%.
Sensors 2025,25, 2301 13 of 15
Sensors2025,25,xFORPEERREVIEW13of16
Figure10.ThefrequencydistributionoftheDNNpredictionsduringtheaack.Whiletheobject’s
correctlabelisa“coffeemug”,theaackexhibitsanaacksuccessrateof85%.
Figure11.TheaverageaacksuccessrateasafunctionoftheAMOLSpulsewidth.
5.Conclusions
Insummary,weintroducedanovelmethodforconductingoptical-basedphysical
adversarialaacksonDNN.Theaackisdemonstratedbydirectingapulsedlightata
CMOScamera.Therollingshuermechanismofthecameraconvertsthetemporalsignal,
whichconsistsofthedesignedsequenceoflightpulses,intoaspatialdistortionwithinthe
physical-worldadversarialimage.Thephotometricconditionsandlightpulsecharacter-
isticsareanalyzedtodazzletheCMOScamerasufficiently,therebyfoolingtheDNN
modelwhilekeepingtheAMOLSactivityinvisibletoobserversintheenvironment.
Wedemonstratedthatthelightsourcedutycycleenablesthecontrolofthetradeoff
betweentheaack’ssuccessrateandtherequiredangulardegreeofconcealment.For
instance,withtheproposedmethod,an85%successrateforthephysical-worldaackcan
beachievedwhileensuringtheinvisibilityoflightsourceactivitytotheobserverexcept
foranarrowangularrangeof5°fromtheopticalaxis.However,theaacksuccessrate
couldbeincreasedto98%byallowingaslightreductionof10°intheangularconceal-
mentrange.
SupplementaryMaterials:Thefollowingsupportinginformationcanbedownloadedat:
www.mdpi.com/xxx/s1.Thefollowingsupportingvideosshowthefootagefromtheaackedcam-
erasequencewhenseingdifferentAMOLSactivities.VideoS1:pulsewidthof1 μs with𝐷
0.01%,andVideoS2:pulsewidthof70 μswith𝐷0.85%.
Chain
Coffe mug
Cup
Digital cloack
Hodometer
Knot
Ruler
Website
Class
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
0.45
0.5
Frequency [%]
Figure 11. The average attack success rate as a function of the AMOLS pulse width.
The performance and properties of our attack are summarized in Table A1 in
Appendix A
, together with a comparison to that of other physical adversarial attacks
involving image sensors.
5. Conclusions
In summary, we introduced a novel method for conducting optical-based physical
adversarial attacks on DNN. The attack is demonstrated by directing a pulsed light at
a CMOS camera. The rolling shutter mechanism of the camera converts the temporal
signal, which consists of the designed sequence of light pulses, into a spatial distortion
within the physical-world adversarial image. The photometric conditions and light pulse
characteristics are analyzed to dazzle the CMOS camera sufficiently, thereby fooling the
DNN model while keeping the AMOLS activity invisible to observers in the environment.
We demonstrated that the light source duty cycle enables the control of the tradeoff
between the attack’s success rate and the required angular degree of concealment. For
instance, with the proposed method, an 85% success rate for the physical-world attack can
be achieved while ensuring the invisibility of light source activity to the observer except for
a narrow angular range of 5
◦
from the optical axis. However, the attack success rate could
be increased to 98% by allowing a slight reduction of 10
◦
in the angular concealment range.
Supplementary Materials: The following supporting information can be downloaded at: https:
//www.mdpi.com/article/10.3390/s25072301/s1. The following supporting videos show the footage
from the attacked camera sequence when setting different AMOLS activities. Video S1: pulse width
of 1 µs with D=0.01%, and Video S2: pulse width of 70 µs with D=0.85%.
Author Contributions: Conceptualization, A.S. and A.H.; methodology, Z.S.; software, Z.S.; val-
idation, Z.S., A.H. and A.S.; investigation, Z.S.; data curation, Z.S.; writing—original draft, Z.S.;
writing—review and editing, A.H. and A.S.; supervision, A.H. and A.S.; project administration, A.S.
All authors have read and agreed to the published version of the manuscript.
Funding: This research received no external funding.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The original data presented in the study are openly available in the
GitHub repository at https://github.com/ZviSteinOpt/RollingShutterAttack/tree/main (accessed
on 1 April 2025).
Conflicts of Interest: The authors declare no conflicts of interest.
Sensors 2025,25, 2301 14 of 15
Appendix A
Table A1. Comparison of physical adversarial attacks involving the image sensors.
Physical World Attack Attack Mechanism Targeting Camera
Sensors
Adversary Physical
Access
Achievable Attack
Success Rate
Invisibility
Criterion
EM Injection [12] CCD interface ✓X Near distances 50% ∼94% a✓
AdvLB [13] Spatial laser beam X X 77.43% ∼100% bX
CamData Lane [14] Camera data lane X ✓Camera interface 89.2% ∼96% c−
RS Backdoor Attack [18] CMOS dazzling ✓X40% ∼88% dXf
Adversarial RS [19] CMOS dazzling ✓X∼84% Xf
Our Attack Invisible AMOLS ✓X 85% ∼98% e✓
a
Average performance from various viewpoints depending on the threat model.
b
Depending on indoor or
outdoor attacks.
c
Depending on the DNN model.
d
Based on simulation study or physical-domain study.
e
Depending on the observer zone location restriction (Figure 5).
f
Designed to prevent visible flickering, although
the illumination source may be seen shining. RS—Rolling Shutter. X and
✓
represent whether the attacks target
the camera sensors to inject their perturbations, require physical access by the adversary, or satisfy the invisibility
criterion in the physical domain. — represents a designed attack assuming adversary physical access.
References
1. Heaven, D. Why deep-learning AIs are so easy to fool. Nature 2019,574, 163–166. [CrossRef] [PubMed]
2.
Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks.
In Proceedings of the International Conference on Learning Representations (ICLR), Banff, AB, Canada, 14–16 April 2014.
3. Yuan, X.; He, P.; Zhu, Q.; Li, X. Adversarial Examples: Attacks and Defenses for Deep Learning. IEEE Trans. Neural Netw. Learn.
Syst. 2019,30, 2805–2824. [CrossRef] [PubMed]
4.
Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. In Proceedings of the International
Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015.
5.
Fang, J.; Jiang, Y.; Jiang, C.; Jiang, Z.L.; Liu, C.; Yiu, S.M. State-of-the-art optical-based physical adversarial attacks for deep
learning computer vision systems. Expert Syst. Appl. 2024,250, 123761. [CrossRef]
6.
Wang, J.; Wang, C.; Lin, Q.; Luo, C.; Wu, C.; Li, J. Adversarial attacks and defenses in deep learning for image recognition: A
survey. Neurocomputing 2022,514, 162–181.
7.
Wei, H.; Tang, H.; Jia, X.; Wang, Z.; Yu, H.; Li, Z.; Satoh, S.I.; Van Gool, L.; Wang, Z. Physical adversarial attack meets computer
vision: A decade survey. IEEE Trans. Pattern Anal. Mach. Intell. 2024,46, 9797–9817. [CrossRef] [PubMed]
8.
Lovisotto, G.; Turner, H.; Sluganovic, I.; Strohmeier, M.; Martinovic, I. SLAP: Improving Physical Adversarial Examples with
Short-Lived Adversarial Perturbations. arXiv 2020, arXiv:2007.04137.
9.
Zhou, Z.; Tang, D.; Wang, X.; Han, W.; Lu, X.; Zhang, K. Invisible Mask: Practical attacks on face recognition with infrared. arXiv
2018, arXiv:1803.04683v1.
10.
Athalye, A.; Engstrom, L.; Ilyas, A.; Kwok, K. Synthesizing Robust Adversarial Examples. In Proceedings of the 35th International
Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018.
11.
Cui, J.; Guo, W.; Huang, H.; Lv, X.; Cao, H.; Li, H. Adversarial examples for vehicle detection with projection transformation.
IEEE Trans. Geosci. Remote Sens. 2024,62, 5632418. [CrossRef]
12.
Liu, Z.; Lin, F.; Ba, Z.; Lu, L.; Ren, K. MagShadow: Physical Adversarial Example Attacks via Electromagnetic Injection. IEEE
Trans. Dependable Secur. Comput. 2025, 1–17. [CrossRef]
13.
Duan, R.; Mao, X.; Qin, A.K.; Chen, Y.; Ye, S.; He, Y.; Yang, Y. Adversarial laser beam: Effective physical-world attack to dnns in a
blink. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June
2021; pp. 16062–16071.
14.
Liu, W.; He, W.; Hu, B.; Chang, C.H. A practical man-in-the-middle attack on deep learning edge device by sparse light strip
injection into camera data lane. In Proceedings of the 2022 IEEE 35th International System-on-Chip Conference (SOCC), Belfast,
UK, 5–8 September 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6.
15.
Wang, X.; Xu, Z.; Zhong, H.; Cheng, X.A.; Xing, Z.; Zhang, J. Fresnel Diffraction Model for Laser Dazzling Spots of Complementary
Metal Oxide Semiconductor Cameras. Sensors 2024,24, 5781. [CrossRef] [PubMed]
16.
Chia-Kai, L. Analysis and Compensation of Rolling Shutter Effect. IEEE Trans. Image Process. 2008,17, 1323–1330. [CrossRef]
[PubMed]
17.
Danakis, C.; Afgani, M.; Povey, G.; Underwood, I.; Haas, H. Using a CMOS Camera Sensor for Visible Light Communication. In
Proceedings of the IEEE Globecom Workshops, Anaheim, CA, USA, 3–7 December 2012.
18.
Li, H.; Wang, Y.; Xie, X.; Liu, Y.; Wang, S.; Wan, R.; Chau, L.P.; Kot, A.C. Light Can Hack Your Face! Black-box Backdoor Attack on
Face Recognition Systems. arXiv 2020, arXiv:2009.06996.
Sensors 2025,25, 2301 15 of 15
19.
Sayles, A.; Hooda, A.; Gupta, M.; Chatterjee, R.; Fernandes, E. Invisible Perturbations: Physical Adversarial Examples Exploiting
the Rolling Shutter Effect. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville,
TN, USA, 20–25 June 2021; pp. 14666–14675.
20.
Chen, Z.; Lin, P.; Jiang, Z.L.; Wei, Z.; Yuan, S.; Fang, J. An illumination modulation-based adversarial attack against automated
face recognition system. In Proceedings of the Information Security and Cryptology: 16th International Conference, Inscrypt
2020, Guangzhou, China, 11–14 December 2020; Revised Selected Papers. Springer International Publishing: Cham, Switzerland,
2021; pp. 53–69.
21.
Shen, Y.; Cheng, Y.; Lin, Y.; Long, S.; Jiang, C.; Li, D.; Dai, S.; Jiang, Y.; Fang, J.; Jiang, Z.L.; et al. MLIA: Modulated LED
illumination-based adversarial attack on traffic sign recognition system for autonomous vehicle. In Proceedings of the 2022 IEEE
International Conference on Trust, Security and Privacy in Computing and Communications (TrustCom), Wuhan, China, 9–11
December 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1020–1027.
22.
Fang, J.; Yang, Z.; Dai, S.; Jiang, Y.; Jiang, C.; Jiang, Z.L.; Liu, C.; Yiu, S.M. Cross-task physical adversarial attack against lane
detection system based on LED illumination modulation. In Proceedings of the Chinese Conference on Pattern Recognition and
Computer Vision (PRCV), Xiamen, China, 13–15 October 2023; Springer Nature: Singapore, 2023; pp. 478–491.
23.
Köhler, S.; Lovisotto, G.; Birnbach, S.; Baker, R.; Martinovic, I. They See Me Rollin’: Inherent Vulnerability of the Rolling Shutter
in CMOS Image Sensors. In Proceedings of the 37th Annual Computer Security Applications Conference, Virtual, 6–10 December
2021; pp. 399–413.
24.
Yan, C.; Xu, Z.; Yin, Z.; Mangard, S.; Ji, X.; Xu, W.; Zhao, K.; Zhou, Y.; Wang, T.; Gu, G.; et al. Rolling colors: Adversarial laser
exploits against traffic light recognition. In Proceedings of the 31st USENIX Security Symposium (USENIX Security 22), Boston,
MA, USA, 10–12 August 2022; pp. 1957–1974.
25.
Nilson, D.G.; Hill, D.N.; Evans, J.C. Thomson Scattering Stray Light Reduction Techniques Using a CCD Camera; Lawrence Livermore
National Laboratory: Livermore, CA, USA, 1997.
26.
Schleijpen, R.H.M.A.; Dimmeler, A.; Eberle, B.; van den Heuvel, J.C.; Mieremet, A.L.; Beckman, H.; Mellier, B. Laser Dazzling of
Focal Plane Array Cameras. In Proceedings of the Defense and Security Symposium, Orlando, FL, USA, 10 October 2007.
27.
Schleijpen, H.M.A.; Carpenter, S.R.; Mellier, B.; Dimmeler, A. Imaging Seeker Surrogate for IRCM evaluation. In Proceedings of
the Optics/Photonics in Security and Defence, Stockholm, Sweden, 5 October 2006.
28.
Santos, C.N.; Chrétien, S.; Merella, L.; Vandewal, M. Visible and near-infrared laser dazzling of CCD and CMOS cameras. In
Proceedings of the Technologies for Optical Countermeasures XV, Berlin, Germany, 9 October 2018.
29.
Eberle, B.; Kinerk, W.T.; Koerber, M.; Öhgren, J.; Ritt, G.; Santos, C.N.; Schwarz, B.; Steinvall, O.; Tipper, S.M.; Vandewal, M.;
et al. NATO SET-249 joint measurement campaign on laser dazzle effects in airborne scenarios. In Proceedings Volume 11161,
Technologies for Optical Countermeasures XVI; SPIE: Bellingham, WA, USA, 2019; pp. 119–138.
30. Blackwell, H.R. Contrast Thresholds of the Human Eye. J. Opt. Soc. Am. 1946,36, 624–643. [CrossRef] [PubMed]
31. Adrian, W. Visibility of targets: Model for calculation. Light Res. Technol. 1989,21, 181–188.
32.
Paschotta, R. RP Photonics Encyclopedia. Available online: https://www.rp-photonics.com/radiance.html (accessed on 1 April 2025).
33. Williamson, C.A.; McLin, L.N. Nominal ocular dazzle distance (NODD). Appl. Opt. 2015,54, 1564–1572. [CrossRef]
34.
McLin, L.N.; Smith, P.A.; Barnes, L.E.; Dykes, J.R. Scaling laser disability glare functions with “K” factors to predict dazzle. In
International Laser Safety; AIP Publishing: Albuquerque, NM, USA, 2015.
35.
Vos, J.; Cole, B.; Bodmann, H.-W.; Colombo, E.; Takeuchi, T.; van den Berg, T.J.T.P. CIE Equations for Disability Glare; CIE TC:
Vienna, Austria, 2002.
36.
Carlini, N.; Wagner, D. Towards Evaluating the Robustness of Neural Networks. In Proceedings of the 2017 IEEE Symposium on
Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017.
37.
He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on
Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.