ArticlePDF Available

Classification with electromagnetic waves

Wiley
IET Microwaves, Antennas & Propagation
Authors:

Abstract and Figures

The integration of neural networks and machine learning techniques has ushered in a revolution in various fields, including electromagnetic inversion, geophysical exploration, and microwave imaging. While these techniques have significantly improved image reconstruction and the resolution of complex inverse scattering problems, this paper explores a different question: Can near‐field electromagnetic waves be harnessed for object classification? To answer this question, we first create a dataset based on the MNIST dataset, where we transform the grayscale pixel values into relative electrical permittivity values to form scatterers and calculate the electromagnetic waves scattered from these objects using a 2D electromagnetic finite‐difference frequency‐domain solver. Then, we train various machine learning models with this dataset to classify the objects. When we compare the classification accuracy and efficiency of these models, we observe that the neural networks outperform others, achieving a 90% classification accuracy solely from the data without a need for projecting the input data into a latent space. The impacts of the training dataset size, the number of antennas, and the location of antennas on the accuracy and time spent during training are also investigated. These results demonstrate the potential for classifying objects with near‐field electromagnetic waves in a simple setup and lay the groundwork for further research in this exciting direction.
This content is subject to copyright. Terms and conditions apply.
Received: 30 March 2024
-
Revised: 16 September 2024
-
Accepted: 27 September 2024
-
IET Microwaves, Antennas & Propagation
DOI: 10.1049/mia2.12522
ORIGINAL RESEARCH
Classication with electromagnetic waves
Ergun Simsek
|Harish Reddy Manyam
Department of Computer Science and Electrical
Engineering, University of Maryland Baltimore
County, Baltimore, Maryland, USA
Correspondence
Ergun Simsek.
Email: simsek@umbc.edu
Funding information
UMBC, Grant/Award Number: 7040330
Abstract
The integration of neural networks and machine learning techniques has ushered in a
revolution in various elds, including electromagnetic inversion, geophysical exploration,
and microwave imaging. While these techniques have signicantly improved image
reconstruction and the resolution of complex inverse scattering problems, this paper
explores a different question: Can near‐field electromagnetic waves be harnessed for
object classification? To answer this question, we rst create a dataset based on the
MNIST dataset, where we transform the grayscale pixel values into relative electrical
permittivity values to form scatterers and calculate the electromagnetic waves scattered
from these objects using a 2D electromagnetic nitedifference frequencydomain solver.
Then, we train various machine learning models with this dataset to classify the objects.
When we compare the classication accuracy and efciency of these models, we observe
that the neural networks outperform others, achieving a 90% classication accuracy
solely from the data without a need for projecting the input data into a latent space.
The impacts of the training dataset size, the number of antennas, and the location of
antennas on the accuracy and time spent during training are also investigated. These
results demonstrate the potential for classifying objects with neareld electromagnetic
waves in a simple setup and lay the groundwork for further research in this exciting
direction.
KEYWORDS
electromagnetic wave scattering, learning (articial intelligence), neural nets
1
|
INTRODUCTION
In the elds of electromagnetic inversion, geophysical explo-
ration, and microwave imaging, the primary objective is to
extract valuable information about the internal structure of
objects or scenes by analysing their interactions with electro-
magnetic waves [1–12]. Traditional methods in these areas
often encounter challenges related to nonlinearity, ill
posedness, and high computational costs [1–12]. In recent
years, many novel machine learningbased approaches have
been proposed to address these issues in applications related to
electromagnetics [13–25], geophysics [14, 26–30], and imaging
[17, 31, 32]. For instance, in ref. [13], a cascade of multilayer
complexvalued residual convolutional neural network (NN)
modules is used to learn a general model for approximating the
underlying electromagnetic inverse scattering system and then
utilise this general model for solving highly nonlinear inverse
scattering problems accurately. In another study [17], a novel
NN architecture, termed the contrast source network, is
introduced to address the issue of traditional techniques getting
trapped in false local minima when recovering high permittivity
objects. In ref. [23], the authors provide a very broad review of
the recently developed deep learningbased approaches for
solving electromagnetic inversion problems and recommend a
learningassisted objectivefunction approach to achieve ac-
curacy with a desired level of condence.
Abbreviations: BIM, Born Iterative Method; CSI, Contrast Source Inversion; DBIM, Distorted Born Iterative Method; GBG, Gaussian Naive Bayes; kNN, knearestneighbour;
MNIST, Modied National Institute of Standards and Technology; NN, neural network; PCA, principal components analysis; RF, random forest; SVM, support vector machine; tSNE,
tdistributed stochastic neighbour embedding; XGB, gradient boosting.
This is an open access article under the terms of the Creative Commons AttributionNonCommercialNoDerivs License, which permits use and distribution in any medium, provided the
original work is properly cited, the use is noncommercial and no modications or adaptations are made.
© 2024 The Author(s). IET Microwaves, Antennas & Propagation published by John Wiley & Sons Ltd on behalf of The Institution of Engineering and Technology.
898
-
IET Microw. Antennas Propag. 2024;18:898910. wileyonlinelibrary.com/journal/mia2
The integration of neural networks and machine learning
techniques has indeed signicantly improved the quality and
speed of image reconstruction, making it possible to tackle
problems that were previously deemed impractical when using
conventional methods. However, the central research question
of this work, which differs from what has previously been
considered, is, Can we achieve the task of classification using
electromagnetic data? So instead of determining, for example,
the electrical permittivity and conductivity distribution over a
certain domain, we are primarily interested in determining the
kinds (classes) of the objects that we have in that domain of
interest.
In the eld of computer vision, researchers— [33] and ref-
erences therein—have developed computer systems that can
automatically recognise and classify handwritten digits in images,
nding applications in postal mail sorting, bank check process-
ing, and form processing. The electromagnetic counterpart,
which involves recognising and classifying objects using elec-
tromagnetic waves, has potential applications in robotics for
enabling robots to perceive their environment and interact with
objects. The use of electromagnetic waves for object recognition
can offer advantages over conventional imaging systems, espe-
cially in scenarios where optical methods (e.g., cameras) are
insufcient. For example, electromagnetic waves can penetrate
certain materials that are opaque to visible light, enabling the
detection of hidden or occluded objects. Furthermore, cameras
rely on lighting conditions and may struggle to detect objects in
lowvisibility environments (such as fog, darkness, or smoke),
whereas electromagnetic waves, particularly in specic frequency
ranges like microwave or millimetrewave, can provide more
robust detection capabilities in these challenging environments.
In this work, we aim to determine whether it is possible to
assign labels or categories to objects based on the electro-
magnetic waves scattered from these objects in a simple setup.
We stress the emphasis on a simple setup, as some researchers
have proposed sophisticated approaches to achieve classica-
tion using light in reservoir computing applications. For
instance, in ref. [34], a new method for extreme deep learning
using electromagnetic waves is proposed, utilising specially
designed materials to create a nonlinear interaction between
light waves of different frequencies. This interaction is then
employed to perform complex learning tasks, such as fore-
casting chaotic time series or classifying various types of data.
This describes a signicant development in the eld of wave
based computing and has the potential to lead to new types of
optical computers that are much faster and more energy
efcient than traditional electronic computers. Nevertheless,
the primary aim of our study is more straightforward: to
classify individual objects based on electromagnetic data
collected by antennas situated in close proximity to the objects
scattering the waves. Successfully demonstrating this concept
might open up the possibility of exploring the practical ap-
plications of electromagnetic wavebased classication in
various realworld scenarios, particularly in the domains of
robotics and object recognition. The key potential advantages
of electromagnetic wavebased object classication over
imagebased recognition can be summarised as follows.
One of the primary advantages of using electromagnetic
waves, particularly at certain frequencies, is their ability to
penetrate materials that are opaque to visible light. This
means that our approach can detect and classify objects that
are hidden behind obstructions, covered by certain mate-
rials, or located in environments where optical methods fail
(e.g., smoke, fog, or darkness). Conventional imagebased
recognition systems, which rely on visible or nearvisible
light, would be severely limited or ineffective in such
scenarios.
Imagebased pattern recognition systems rely on visual
features (shape, colour, texture, etc.) to classify objects.
However, they cannot directly discern the material proper-
ties of the objects they are identifying. In contrast, our
method is sensitive to the dielectric properties of objects,
which allows us to distinguish between materials that may
appear visually similar but have different electromagnetic
properties. This could be particularly useful in applications
where material composition is a key factor, such as detecting
hazardous materials, distinguishing between different types
of plastics, or even identifying materials based on their
electromagnetic response.
Optical systems can be affected by environmental condi-
tions such as lighting, shadows, or reections, which can
introduce noise or distortions in image recognition tasks.
Electromagnetic wavebased systems, on the other hand, are
less sensitive to such factors. They can operate effectively
across various environments without the need for specic
lighting conditions, making them more robust for realworld
applications, especially in challenging or uncontrolled
environments.
Unlike conventional image recognition, which only captures
surface information, electromagnetic wave scattering tech-
niques have the potential for subsurface imaging, allowing
for classication of objects that are buried or embedded
within other materials. This could be particularly useful in
nondestructive testing, medical imaging, or geophysical
exploration, where it is crucial to identify the objects
beneath the surface.
It should also be noted that there are some other recent
studies that have shown the integration of machine learning
and electromagnetic wave analysis for material identication
and classication. For example, Harrison et al., proposed a
novel methodology for material identication using a micro-
wave sensor array [35]. Unlike traditional systems that use a
single resonating sensor, this method employs an array reso-
nating at different frequencies to improve identication accu-
racy. Machine learning algorithms were applied to the collected
data, and the approach was validated on various materials such
as wood, cardboard, and plastics. In another work, Cova-
rrubiasMartínez et al., introduced a method for classifying
plastic materials using a microwave resonant sensor [36]. They
evaluated several machine learning classiers in material clas-
sication, accounting for uncertainties such as air gaps and
pellet positions. Their approach presents a fast, non
destructive method for identifying plastic raw materials with
SIMSEK and MANYAM
-
899
industrial potential. Both in refs. [35, 36], the resonances in the
collected data play a crucial role because the primary intention
is the material identication. In a very recent study, Ting et al.,
proposed a material classication system utilising an embedded
random forest (RF) antenna array, which measures changes in
the received signal strength indicator values [37]. The study
combined a Kalman lter with a support vector machine
(SVM) classier, achieving over 96% accuracy in material
classication within a 2m range. Their system, designed for
mobile robotic applications such as warehouse automation,
focuses on realtime, proximal remote sensing of materials.
Our work might be considered a numerical version of their
experimental study.
The structure of this paper is as follows. First, we provide a
brief overview of electromagnetic inversion and explain its main
difference from electromagnetic classication. Then, we
describe how we create our dataset. Note that the sample codes
to produce the results presented in this work and the dataset can
be found at ref. [38]. This dataset can be used for both electro-
magnetic classication and inversion problems. Third, we
compare the classication accuracy and efciency of several
machine learning models trained to classify 10 labels from the
electromagnetic data recorded by antennas placed around one
half of the domain of interest. Subsequently, we investigate
how accuracy and training time change with the training dataset
size, the number of antennas, and the location of antennas.
Following some discussions, we present our conclusions.
2
|
ELECTROMAGNETIC INVERSION
VS. ELECTROMAGNETIC
CLASSIFICATION
Consider a black box, as depicted in Figure 1a, surrounded by
an array of transmitter and receiver antennas. Inside this black
box, assume the presence of a concealed object characterised
by relative electrical permittivity 3.5, as illustrated in Figure 1b.
The surrounding medium is assumed to be air. In the context
of this simplied electromagnetic inversion problem, wherein
all materials involved are treated as lossless and nonmagnetic
dielectrics, the objective is to infer the permittivity distribution
across the domain of interest based on the electromagnetic
data acquired by the receiver antennas. One can solve this
electromagnetic inversion problem numerically and iteratively
[1–11, 23, 39] as follows.
Consider a scenario where there exist MTilluminating
sources for exciting the medium and MRreceivers for collecting
the scattered eld. Consequently, the overall number of ac-
quired data points is denoted as M¼MTMR. Let the
reconstruction domain Dbe discretised into Nsmall cells, with
constant eld quantities and contrast function within each cell.
The total electric eld at position rwithin the dielectric object,
induced by an exciting source situated at rT, can be expressed
as a sum of the incident and scattered elds. This summation is
governed by the superposition principle and can be repre-
sented as follows:
E r;rT
ð Þ ¼ Einc r;rT
ð Þ þ k2þ
ZDGAJ r;r0
ð Þ χr0
ð ÞE r0;rT
ð Þdr0;r2Dð1Þ
where GAJ r;r0
ð Þ is an auxiliary dyadic Green's function rep-
resenting the magnetic vector potential, the wavenumber is
given by k2¼ω2μ~
eand χðrÞis the contrast, that is,
χðrÞ ¼ ~
eðrÞeð Þ=~
e. Equation (1) is called the object equation,
which is a Fredholm integral equation of the second kind for
the unknown eld inside the object. Once the total electric
eld is obtained, then the scattered eld recorded at any of the
receiver antennas, let's say located on path can be calculated
with the data equation which denes the scattered eld at the
observation point, that is,
FIGURE 1 (a) Domain of interest is surrounded by a group of transmitter and receiver antennas, (b) an object embedded in the domain of interest, and a
typical output of (c) an electromagnetic inversion and (d) electromagnetic classication.
900
-
SIMSEK and MANYAM
Esca
mr;rT
ð Þ ¼ jω~
eZDGEJ r;r0
ð Þ χr0
ð ÞE r0;rT
ð Þdr0;r2ð2Þ
where GEJ r;r0
ð Þ is the electric dyadic Green's function at the
observation point rrelated to a unit current source at the
point r0.
The data equation, establishing the connection between the
measured data and the unknown contrast of the material, can
be expressed in a discretised form as follows:
f riR;riT
ð Þ ¼ ω~
eqX
N
k¼1
GEJ riR;r0
k
E r0
k;riT
χr0
k
ΔSð3Þ
where fis a 2Mdimensional data column vector whose ele-
ments are given measured scattered electric eld data, ΔSis the
surface element, iR¼1;;MRand iT¼1;;MTdenote
indices for the receiver and transmitter, respectively. Note that
if this was a threedimensional problem, then fwould be a 3M
dimensional data and instead of surface element ðΔSÞ, we
would be using volume elements ðΔVÞ. For Mmeasurements
and Ndiscretised cells, Equation (3) can be written compactly
as follows:
f¼Mx;ð4Þ
where xis a Ndimensional column vector of the contrast
function χ, and Mis a 2MNmatrix whose elements are
given using the following equation:
Mik ¼jω~
eqGEJ
mq riR;r0
k
E r0
k;riT
ΔSð5Þ
where i¼iRþiT1ð ÞMR, and k¼1;;N.
Since the total electric eld Ewithin the objects is an un-
known function of the material contrast χ, Equation (4) becomes
a nonlinear equation. Furthermore, the limited amount of
available information renders the problem nonunique. This
equation can be solved iteratively using either the Contrast
Source Inversion (CSI) method [8, 9, 12], or the Born Iterative
Method (BIM) [2, 7, 39], or the Distorted Born Iterative Method
(DBIM) [10, 11, 39]. The CSI method constructs a sequence of
contrast sources and contrasts iteratively without relying on a
forward solver. While it is a stable method, it requires a large
number of iterations to achieve the desired accuracy. Born
Iterative Method and DBIM are commonly employed iterative
methods for solving nonlinear inverse scattering problems, as
they typically demand fewer iterations. The primary distinction
between BIM and DBIM is that the latter utilises an updated
background Green's function for each iteration, a necessity when
addressing electromagnetic inversion problems involving inho-
mogeneous backgrounds, such as multilayered media. Due to
this difference, DBIM is computationally more expensive than
BIM, but it boasts secondorder convergence, whereas BIM only
achieves rstorder convergence. This computational cost/
convergence order tradeoff can be managed through a twostep
algorithm, as elucidated in ref. [7].
Over the past two decades, both computational electro-
magnetics and geophysics societies have witnessed a growing
interest in enhancing the efciency of existing methodologies
through the incorporation of machine learning techniques,
with a particular emphasis on the application of deep learning.
This convergence marks a signicant paradigm shift in the
pursuit of optimising electromagnetic and geophysical simu-
lations for diverse applications. One major avenue of explo-
ration involves training neural networks to directly reconstruct
the contrast of scatterers based on the measured scattered
elds [13–16, 28, 40, 41]. This approach capitalises on the
inherent capacity of deep learning models to discern complex
patterns and relationships within datasets. By leveraging these
capabilities, researchers aim to enhance the accuracy and speed
of contrast reconstruction, enabling a more precise represen-
tation of the underlying electromagnetic or geophysical prop-
erties. In parallel, another compelling avenue unfolds where
neural networks are harnessed to learn dynamically throughout
the iterative solution process [17, 18, 20–22, 24, 25, 30, 31].
This innovative approach stands in contrast to traditional
methods by enabling the network to adapt and rene its un-
derstanding of the problem space as the solution evolves.
Through this adaptive learning process, neural networks can
capture intricate features and nuances in the data, thereby
contributing to more robust and efcient solutions.
However, the focus here is not on determining the
permittivity distribution across the domain of interest. Instead,
the objective is to identify the object within the black box,
exemplied in Figure 1d. One possible approach to tackle this
classication task involves training neural networks, with the
input data being the electromagnetic data recorded by the
receiver antennas and the outputs corresponding to the object
types (classes). Given the absence of an extensive dataset
conducive to such a study, the subsequent section outlines our
efforts to generate one from a wellestablished dataset, as
detailed below.
3
|
THE DATASET
The Modied National Institute of Standards and Technology
(MNIST) dataset [42] is a collection of handwritten digits and
widely used in the eld of machine learning and computer
vision for training and testing various machine learning algo-
rithms, particularly for tasks like image classication and
character recognition. It consists of 28 28pixel square
grayscale images of handwritten digits, ranging from 0 to 9.
Each image is associated with a corresponding label that
species which digit the handwritten image represents.
In this work, the 60,000 digital images of the MNIST
dataset are used to create a scatterer database as follows. In the
MNIST dataset, the value of each pixel in an image represents
the grayscale intensity of that pixel. The pixel values, du;vfor
u;v¼1;2;;28, are typically scaled to fall within a range of
0–255, with 0 indicating white (no ink) and 255 indicating black
(maximum ink saturation). Values between 0 and 255 represent
SIMSEK and MANYAM
-
901
various shades of grey. These values can be converted into
relative electrical permittivity values ranging from emin
rto emax
r
using the following equation:
erðu;vÞ ¼ emin
rþemax
remin
r
du;v
255 ð6Þ
where uand vare the row and column numbers. We set
emin
r¼1 and emax
r¼4 for the reasons that are discussed below.
To create the electromagnetic scattering dataset, we utilise a
freely available 2D electromagnetic nite difference frequency
domain simulation tool called Ceviche [43]. The computation
domain, shown in Figure 2a, has dimensions of 2λ2λand is
uniformly meshed along the xand ydirections, with
Δx¼Δy¼λ=150, where λrepresents the wavelength of the
electromagnetic waves generated by a transmitter antenna with
a length of λ. The thickness of the perfectly matched layers
along all directions is λ=7:5. Initially, the permittivity of each
cell is assumed to be 1.
Since the original les in the MNIST dataset (28 28
pixels) are too small compared to our computation domain
(300 300 pixels), we use a 2D cubic interpolation to update
the permittivity of the 140pixel by 140pixel region at the
centre of the computation domain with the permittivity values
obtained using Equation (6), as illustrated in Figure 2a.
Figure 2b displays the permittivity distribution across the
computation domain for one of the examples studied in this
work. In this setup, we consider two groups of receiver an-
tennas. The rst group comprises Nr=2 receiver antennas
uniformly placed between x¼0:4λand x¼1:57λat y¼1:6λ,
while the second group consists of Nr=2 receiver antennas
uniformly placed between y¼0:4λand y¼1:57λat x¼1:6λ,
where Nris an even integer that represents the total number of
receiver antennas. Note that these antennas are not intended to
physically capture the permittivity values of each pixel but
rather capture the scattered electromagnetic elds resulting
from the permittivity distributions created with Equation (6)
from the pixelated MNIST images. The connection lies in how
the pixel values of the MNIST images affect the permittivity
distribution, which in turn inuences the electromagnetic elds
that the antennas record. Additionally, there are two trans-
mitter antennas: transmitter antenna1 is positioned at
x¼0:2λbetween y¼0:5λand y¼1:5λ, while transmitter
antenna2 is situated at y¼0:2λbetween x¼0:5λand
x¼1:5λ. Since we treat the problem as a 2D problem, we only
deal with three eld components: the zcomponent of the
electric eld Ez
ð Þ and the xand ycomponents of the magnetic
eld (Hxand Hy). We begin by assuming that transmitter
antenna1 is active and proceed to calculate Ez,Hx, and Hyat
Nrreceiver antennas. We then repeat this process, assuming
that only transmitter antenna2 is active. The real and imagi-
nary parts of the calculated electric and magnetic elds are
stored in separate columns of the dataset. Consequently, the
input section of the dataset consists of 60,000 rows and 624
columns. These 624 columns represent the maximum number
of receiver antennas, Nmax
r¼52, two transmitter antennas,
three elds, and two components (real and imaginary parts).
The output section of the dataset consists of the labels (digits),
resulting in a 60;000 1 vector. Note that in the rst part of
the numerical studies, we only 10 receiver antennas. The reason
we record the electric eld intensities at 52 different locations
is that in the second part of the numerical studies, we aim to
FIGURE 2 (a) The computation domain covers a 2λby 2λregion that is uniformly meshed along the xand yaxes at a mesh sampling density of λ=150.
Greyshaded areas represent the perfectly matched layers. (b) Permittivity distribution for one of the example geometries studied. The purple regions have a relative
permittivity of 1. The regions with higher relative permittivity values are represented by lighter colours. The locations of transmitter antennas are indicated by yellow
dashed lines, while white circles depict the positions of the 52 receiver antennas.
902
-
SIMSEK and MANYAM
determine the affect of antenna locations and interantenna
spacing on the classication accuracy.
For the sake of brevity, we do not provide a gure that
depicts the electric eld intensities jEzjð Þ recorded by the
receiver antennas for each label, but we would like to briey
emphasise that there is no clear pattern recognisable by
humans for identifying the labels from the eld proles. Also,
to investigate the complexity of our data, we project 5000
samples from our dataset to two latent spaces using principal
components analysis (PCA) [44, 45] and tdistributed sto-
chastic neighbour embedding (tSNE) [45, 46]. The visual-
isations are provided in Figure 3. Unlike for the original
MNIST dataset [45], here we do not have a clear interpretation
even after projections.
4
|
MACHINE LEARNING MODELS
We employ various machinelearning models, namely knearest
neighbour (kNN), RF, Gaussian Naive Bayes (GNB), SVM,
gradient boosting (XGB), and a NN, to assess whether they
exhibit similar levels of learning from the data or not. For the
rst four models (kNN, RF, GNB, and SVM), we utilise scikit
learn [47], which is a free software machine learning library for
the Python programming language. For the XGB imple-
mentation, we use another freely available library [48], which is
developed based on XGBoost [49]. Hyperparameter tuning is
done using the Grid Search CrossValidation (GridSearchCV)
module [50] that is available in the scikitlearn [47] library.
For the NN implementations, we employ Keras' [51]
functional application programme interface running on top of
TensorFlow [52]. The model starts with an input layer of Nc
dimensions, where Nc¼12 Nris the number of columns
of our input data, followed by a dense layer with 500 units,
utilising the HeNormal weight initialisation [53] and
employing the rectied linear unit activation function [54].
Subsequently, another dense layer with 250 units continues
the ow, maintaining the same initialisation and activation
choices. To enhance the model's robustness, a dropout layer
with a dropout rate of 0.2 is introduced to mitigate over-
tting. The architecture then proceeds with a dense layer of
100 units, followed by a smaller layer with 30 units, both
sharing the same initialisation and activation congurations.
To further optimise the network, a batch normalisation layer
is integrated, contributing to the stability and efciency of the
learning process. The nal touch is a dense layer with 10
units and a softmax activation function [55], tailored for a
classication task with 10 output classes. The model's learning
process is guided by the Adam optimiser [56], with a learning
rate set at 0.001. The categorical crossentropy loss [57] is
chosen as the optimisation objective, and categorical accuracy
is monitored as a metric to gauge performance. During
training, adaptive strategies are implemented. Specically,
a learning rate reduction is triggered by the Reduc‐
eLROnPlateau callback, which responds to changes in vali-
dation categorical accuracy. Additionally, the EarlyStopping
callback monitors the same metric, terminating training early
if improvements plateau for an extended period. To ensure
the preservation of the bestperforming model, the Mod‐
elCheckpoint callback is employed. It saves the model
whenever a new high in validation categorical accuracy is
achieved, promoting the retention of optimal weights. The
architecture as a whole embodies a holistic approach,
combining architectural choices and training strategies for an
effective and adaptive NN. All the code is executed on
Google Colaboratory using T4 GPU accelerators.
5
|
NUMERICAL RESULTS
In this section, we rst evaluate the accuracy of two machine
learning models and then investigate how different factors,
such as dataset size and the number and location of antennas,
inuence the results.
FIGURE 3 Visualisations of the labels in two latent spaces obtained with (a) principal components analysis (PCA) and (b) tSNE.
SIMSEK and MANYAM
-
903
5.1
|
Accuracy of different ML methods
For the initial set of calculations, we use 10 receiver antennas
and allocate 50% of the dataset for training and the other 50%
for testing. Table 1presents the time spent during training for
each model, as well as their accuracy, precision, recall, and F1
scores. We observe that SVM, XGB, and NN implementations
achieve higher accuracy (>80%), precision, recall, and F1
scores compared to the kNN, RF, and GNB implementations.
Although the NN's training time is longer than the SVM and
XGB, it delivers the highest classication accuracy (90.17%).
This result is promising but not surprising, given the NN's
ability to identify extremely complex patterns when trained on
sufciently large datasets. It is worth noting that the classi-
cation accuracy obtained with the NN is higher than the pre-
viously reported value [34] of 85.3%.
Figure 4displays the confusion matrix of the NN imple-
mentation, revealing that the most accurately classied label is
1, while eight is the least accurately classied. Notably, 5.7%
and 3.8% of objects labelled as 8 are in fact 3 and 5s.
One of the signicant advantages of working with neural
networks is the ease with which you can quantify and visualise
the learning process's efciency. Figure 5a,b illustrate how
accuracy and loss change with respect to the epoch number for
both training (solid curves) and validation (dashed curves). The
little misalignment of these curves informs us that the number
of neurons in the NN architecture is slightly higher than the
optimal but the overtting is not serious. The convergent
behaviour observed in both gures indicates that the number
of epochs used during the training is sufcient. In the same
gure, we also observe that the training is ended at the 68th
TABLE 1The time spent during training, accuracy, precision, recall,
and F1score of the knearest neighbour (kNN), random forest (RF),
Gaussian Naive Bayes (GNB), support vector machine (SVM), gradient
boosting (XGB), and neural network (NN) implementations using the 50%
of the dataset for training and the remaining 50% for testing.
Method Time (s) Accuracy Precision Recall F1 score
kNN 0.1 0.66 0.67 0.66 0.65
RF 495 0.76 0.75 0.75 0.75
GNB 0.52 0.61 0.61 0.60 0.60
SVM 296 0.85 0.85 0.85 0.85
XGB 183.7 0.84 0.84 0.84 0.84
NN 434 0.90 0.90 0.90 0.90
FIGURE 4 Confusion matrix of the neural network (NN) implementation. Since none of the 7s is predicted as a 6, the corresponding cell is blank.
904
-
SIMSEK and MANYAM
epoch even the number of epochs is set to 100 thanks to the
EarlyStopping callback monitor. Also, the learning rate is
dropped from 103to 4:398 105in 14 steps triggered by
the ReduceLROnPlateau callback.
Machine learningbased computer vision algorithms can
achieve a classication accuracy of over 97% for handwritten
digits [33]. However, these studies typically use all pixels to
achieve such high accuracy, while we use only electromagnetic
data recorded by antennas placed on two onedimensional lines
near the objects. Therefore, we consider a 90% classication
accuracy using electromagnetic waves to be a successful
approach.
5.2
|
Inuence of dataset size and number
and location of antennas on the accuracy
For the results presented in Table 1and Figures 4and 5, we use
50% of the data for training, with the remaining 50% used for
testing. In other words, we use 30,000 samples for training and
other 30,000 samples that have not been seen by the NN
before for testing. To examine the impact of the training
dataset size Ntrain
ð Þ on accuracy and training time, we conduct
an additional two sets of calculations. For the rst set, we use
10% of the data for training Ntrain ¼6000ð Þ and the remaining
90% for testing. For the second set, we use 1% of the data for
training Ntrain ¼600ð Þ and the remaining 99% for testing. The
accuracy of each method is listed in Table 2. As expected, the
accuracies drop for all methods. However, we would like to
emphasise one important detail: NN is the most accurate when
the dataset is large enough (Ntrain is 6000 or higher in this case),
however, their accuracy drops signicantly for small datasets
and other methods–such as XGB in this case–can perform
better. We might conclude that if we are going to implement a
NN for object classication, then we should have a large
enough dataset to achieve high accuracy.
Next, we investigate the inuence of the number of antennas
over the accuracy as follows. For the three cases above, where the
training dataset size is changed from 50% to 10% rst and then
to 1%, we decrease the number of receiver antennas from 10 to 8,
4, and 2, and monitor the accuracy and training times. The results
are plotted in Figure 6. Note that the number of transmitter
antennas is still 2, which means that the number of data points is
2Nr. As we examine Figure 6a–c, we rst observe that both
the accuracy and training times decrease with decreasing dataset
size but the accuracy remains almost constant when the receiver
antenna number is reduced from 10 to 4. Hence, we might claim
that having an interantenna spacing of λ=4 is adequate for ac-
curate classication as long as we have a large training dataset. In
terms of training time, we observe a slight decrease in training
times with a decreasing number of antennas, but this decrease is
not linear and the main factor that determines the training time is
the dataset size.
Next, we investigate the inuence of antenna locations
over the accuracy as follows. Since the accuracy signicantly
drops when we have two receiver antennas only, we rst start
FIGURE 5 (a) Accuracy and (b) loss versus epoch number of the neural network (NN) implementation, which uses 50% of the dataset for training and the
remaining 50% for testing.
TABLE 2Accuracy (in percentile) of the knearest neighbour (kNN),
random forest (RF), Gaussian Naive Bayes (GNB), support vector machine
(SVM), XGB, and neural network (NN) implementations using 50%, 10%,
and 1% of the dataset for training and the remaining data for testing.
Training/Testing ratio kNN RF GNB SVM XGB NN
50%/50% 66 76 61 85 84 90
10%/90% 56 67 55.5 76 79 81
1%/99% 29 48 37 49 57 54
SIMSEK and MANYAM
-
905
investigating this scenario by creating pairs of antennas by
selecting one antenna from the top antenna group and one
antenna from the side antenna group at a time. As mentioned
before, our dataset includes eld intensities recorded at 52
different locations. Here, for the pairi, we select antennas Ri
and Riþ26 for i¼1;2;;13 as shown in Figure 7a for the rst
two pairs, and we monitor the accuracy. The blue curve in
Figure 7c shows how the accuracy changes with the pair
number. Even though the results change slightly (0:6%), we
can understand the trend (peaking near the middle), by
considering the following two facts: (i) the scatterers are placed
parallel to the xaxis, so the antennas from the top group are
likely to carry more information about the scatterer than the
side antennas, and (ii) when we increase i, we get closer to
the scatterer centre but at the same time, the distance between
the left transmitter antenna and ith receiver antenna increases
while the distance between the bottom transmitter antenna and
iþ26th receiver antenna decreases.
Then we repeat the same procedure for groups of 4 an-
tennas at a time. In this case, the groupiis formed by select
antennas Riand Riþ13 from the top group and antennas Riþ26
and Riþ39 from the side group as shown in Figure 7b for the
rst two groups. How accuracy changes with the group
number is depicted with the red curve in Figure 7c. In this
case, the accuracy decreases almost steadily. Again, we believe
that the increasing distance from the left transmitter antenna
with increasing iand hence the weakening scattering data is the
main cause of this decaying accuracy.
6
|
DISCUSSIONS
In this work, the main aim was to achieve high classication
accuracy without projecting the input data into a latent space.
If such projections, for example, PCA, are used, both neural
networks and gradientboosted decision trees achieve the same
accuracy. However, it should be noted that unlike the original
MNIST dataset, here we have to utilise more than 30 principal
components to achieve a 90% accuracy due to the complex
nature of electromagnetic scattering data. If 20 principal
components or fewer are used, the accuracy drops signicantly.
Figure 8shows the explained variance ratio of the rst 40
principal components on a logarithmic scale.
The accuracy of the classication depends on the permit-
tivity range of the scatterers. In this work, we choose the
maximum allowed relative permittivity to be 4 for two reasons.
First, such narrow permittivity ranges, where emax
r=emin
r<10
are already being reported, for example, in [58]. Second, larger
permittivity values would impede the propagation of electro-
magnetic waves through the objects. For instance, if the
maximum allowed electrical permittivity is set to 12, the ac-
curacy of the NN implementation drops to 79%, primarily due
FIGURE 6 (left) Accuracy and (right) training time as a function of the number of receiver antennas when (a) N
training ¼15000, (b) Ntraining ¼6000,
and (c) Ntraining ¼600.
906
-
SIMSEK and MANYAM
to the strong reectance from the objects back to the trans-
mitter antennas.
Another important factor is the orientation of the scat-
terers. In this study, all the scatterers were parallel to the xaxis.
When we rotate the scatterers, we observe that the accuracy of
classication decreases to 58%. To achieve higher accuracy for
objects with arbitrary alignments, it is necessary to place
receiver antennas all around the domain of interest.
FIGURE 7 (a) Pairing two antennas by selecting only one antenna from the top antenna group and one antenna from the side antenna group, (b) grouping
4 antennas by selecting two antennas from the top and side antenna groups while keeping the interantenna distance xed, (c) accuracy versus pair or group
number.
SIMSEK and MANYAM
-
907
7
|
CONCLUSION
In this paper, we have explored the application of machine
learning and NN techniques in the context of electromagnetic
wavebased object classication. The objective of our work
was to determine whether it is feasible to classify objects based
on the electromagnetic waves scattered from them in a simple
experimental setup. We began by creating a dataset using the
MNIST data, where we transformed grayscale pixel values into
relative electrical permittivity values. A 2D electromagnetic
nite difference frequency domain simulation tool was
employed to calculate the electric and magnetic eld data
recorded by receiver antennas placed around the objects
partially. We evaluated the classication performance of
various machine learning models, including kNN, RF, GNB,
SVM, gradient boosting, and a NN. The NN architecture
demonstrated the highest accuracy, achieving an 90% classi-
cation accuracy. We also investigated the impact of the training
dataset size, number of antennas, and location of antennas on
accuracy and training time, highlighting the advantages of using
a NN, especially as the dataset size is increased. However, we
also observe that the number of samples of the dataset is more
important than the number of receiver antennas to achieve a
high accuracy to classify objects whose dimensions are close to
the wavelength of excitation and we use four or more receiver
antennas placed uniformly over one wavelength long region.
While computer vision algorithms can achieve higher accuracy
for handwritten digit recognition, our approach demonstrates
that classifying objects using electromagnetic waves is a
promising avenue. This research opens the door to exploring
the potential of electromagnetic wavebased classication for
various applications, including robotics and object recognition.
AUTHOR CONTRIBUTIONS
Ergun Simsek: Conceptualisation; Data curation; Formal
analysis; Investigation; Methodology; Project administration;
Supervision; Validation; Visualisation; Writing original draft;
Writing review & editing. Harish Reddy Manyam: Inves-
tigation; Validation.
CONFLICT OF INTEREST STATEMENT
The authors declare no conicts of interest.
DATA AVAILABILITY STATEMENT
Sample codes to produce the results presented in this work and
the dataset can be found at https://github.com/simsekergun/
CwEMW. This dataset can be used for both electromagnetic
classication and inversion problems.
ORCID
Ergun Simsek
https://orcid.org/0000-0001-9075-7071
REFERENCES
1. Bucci, O.M., Isernia, T.: Electromagnetic inverse scattering: retrievable
information and measurement strategies. Radio Sci. 32(6), 2123–2137
(1997). https://doi.org/10.1029/97rs01826
2. Wang, Y.M., Chew, W.C.: An iterative solution of the twodimensional
electromagnetic inverse scattering problem. Int. J. Imag. Syst. Technol.
1(1), 100–108 (1989). https://doi.org/10.1002/ima.1850010111
3. Chen, X.: Computational Methods for Electromagnetic Inverse Scat-
tering, vol. 244. Wiley Online Library (2018). https://doi.org/10.1002/
9781119311997
4. Egbert, G.D., Kelbert, A.: Computational recipes for electromagnetic
inverse problems. Geophys. J. Int. 189(1), 251–267 (2012). https://doi.
org/10.1111/j.1365246x.2011.05347.x
5. Abubakar, A., et al.: Inversion algorithms for largescale geophysical
electromagnetic measurements. Inverse Probl. 25(12), 123012 (2009).
https://doi.org/10.1088/02665611/25/12/123012
6. Dorn, O., et al.: A nonlinear inversion method for 3D electromagnetic
imaging using adjoint elds. Inverse Probl. 15(6), 1523–1558 (1999).
https://doi.org/10.1088/02665611/15/6/309
7. Wei, B., et al.: Threedimensional electromagnetic nonlinear inversion in
layered media by a hybrid diagonal tensor approximation: stabilized
biconjugate gradient fast Fourier transform method. Waves Random
Complex Media 17(2), 129–147 (2007). https://doi.org/10.1080/
17455030601016117
8. Abubakar, A., van den Berg, P.M.: Threedimensional nonlinear inver-
sion in crosswell electrode logging. Radio Sci. 33(4), 989–1004 (1998).
https://doi.org/10.1029/98rs00975
9. Abubakar, A., Van den Berg, P., Semenov, S.Y.: Twoand three
dimensional algorithms for microwave imaging and inverse scattering.
J. Electromagn. Waves Appl. 17(2), 209–231 (2003). https://doi.org/10.
1163/156939303322235798
10. Cui, T.J., Chew, W.C.: Novel diffraction tomographic algorithm for im-
aging twodimensional targets buried under a lossy earth. IEEE Trans.
Geosci. Rem. Sens. 38(4), 2033–2041 (2000). https://doi.org/10.1109/
36.851784
11. Cui, T.J., et al.: Inverse scattering of twodimensional dielectric objects
buried in a lossy earth using the distorted Born iterative method. IEEE
Trans. Geosci. Rem. Sens. 39(2), 339–346 (2001). https://doi.org/10.
1109/36.905242
FIGURE 8 Explained variance ratio of the rst
40 principal components in a logarithmic scale.
908
-
SIMSEK and MANYAM
12. Yu, C., Song, L.P., Liu, Q.H.: Inversion of multifrequency experimental
data for imaging complex objects by a DTA–CSI method. Inverse Probl.
21(6), S165–S178 (2005). https://doi.org/10.1088/02665611/21/6/s12
13. Li, L., et al.: DeepNIS: deep neural network for nonlinear electromag-
netic inverse scattering. IEEE Trans. Antennas Propag. 67(3), 1819–1825
(2019). https://doi.org/10.1109/tap.2018.2885437
14. Yao, H.M., Jiang, L., Sha, W.E.I.: Enhanced deep learning approach
based on the deep convolutional encoder–decoder architecture for
electromagnetic inverse scattering problems. IEEE Antennas Wireless
Propag. Lett. 19(7), 1211–1215 (2020). https://doi.org/10.1109/lawp.
2020.2995455
15. Xu, K., et al.: Deep learningbased inversion methods for solving inverse
scattering problems with phaseless data. IEEE Trans. Antennas Propag.
68(11), 7457–7470 (2020). https://doi.org/10.1109/tap.2020.2998171
16. Simsek, E.: Machine learning exercises on 1D electromagnetic inversion.
IEEE Trans. Antennas Propag. 69(10), 6797–6805 (2021). https://doi.
org/10.1109/tap.2021.3069519
17. Sanghvi, Y., Kalepu, Y., Khankhoje, U.K.: Embedding deep learning in
inverse scattering problems. IEEE Trans. Comput. Imaging 6, 46–56
(2020). https://doi.org/10.1109/tci.2019.2915580
18. Xiao, L.Y., et al.: Dualmodule NMMIEM machine learning for fast
electromagnetic inversion of inhomogeneous scatterers with high con-
trasts and large electrical dimensions. IEEE Trans. Antennas Propag.
68(8), 6245–6255 (2020). https://doi.org/10.1109/tap.2020.2990222
19. Guo, R., et al.: Physics embedded deep neural network for solving full
wave inverse scattering problems. IEEE Trans. Antennas Propag. 70(8),
6148–6159 (2022). https://doi.org/10.1109/tap.2021.3102135
20. Fei, J., et al.: Fast 3D electromagnetic fullwave inversion of dielectric
anisotropic objects based on ResUNet enhanced by variational born
iterative method. IEEE Trans. Antenn. Propag. 70(8), 6229–6239 (2022).
https://doi.org/10.1109/tap.2021.3138551
21. Shan, T., et al.: Neural Born iterative method for solving inverse scat-
tering problems: 2D cases. IEEE Trans. Antennas Propag. 71(1), 818–
829 (2023). https://doi.org/10.1109/tap.2022.3217333
22. Khorashadizadeh, A., et al.: Deep injective prior for inverse scattering.
IEEE Trans. Antennas Propag. 71(11), 8894–8906 (2023). https://doi.
org/10.1109/tap.2023.3312818
23. Chen, X., et al.: A review of deep learning approaches for inverse scat-
tering problems (invited review). Prog. Electromagn. Res. 167, 67–81
(2020). https://doi.org/10.2528/pier20030705
24. Adler, J., Öktem, O.: Solving illposed inverse problems using iterative
deep neural networks. Inverse Probl. 33(12), 124007 (2017). https://doi.
org/10.1088/13616420/aa9581
25. Chen, G., et al.: Learningassisted multimodality dielectric imaging. IEEE
Trans. Antennas Propag. 68(3), 2356–2369 (2019). https://doi.org/10.
1109/tap.2019.2948565
26. Giannakis, I., Giannopoulos, A., Warren, C.: A machine learningbased
fastforward solver for ground penetrating radar with application to
fullwaveform inversion. IEEE Trans. Geosci. Rem. Sens. 57(7), 4417–
4426 (2019). https://doi.org/10.1109/tgrs.2019.2891206
27. Hu, Y., et al.: A supervised descent learning technique for solving
directional electromagnetic loggingwhiledrilling inverse problems.
IEEE Trans. Geosci. Rem. Sens. 58(11), 8013–8025 (2020). https://doi.
org/10.1109/tgrs.2020.2986000
28. Liu, B., et al.: Deep learning inversion of electrical resistivity data. IEEE
Trans. Geosci. Rem. Sens. 58(8), 5715–5728 (2020). https://doi.org/10.
1109/tgrs.2020.2969040
29. Simsek, E.: Determining optical constants of 2D materials with neural
networks from multiangle reectometry data. Mach. Learn.: Sci. Tech-
nol. 1(1), 01LT01 (2020). https://doi.org/10.1088/26322153/ab6d5f
30. Chen, X.: Subspacebased optimization method for solving inverse
scattering problems. IEEE Trans. Geosci. Rem. Sens. 48(1), 42–49
(2009). https://dx.doi.org/10.1109/TGRS.2009.2025122
31. Shao, W., Du, Y.: Microwave imaging by deep learning network: feasibility
and training method. IEEE Trans. Antennas Propag. 68(7), 5626–5635
(2020). https://doi.org/10.1109/tap.2020.2978952
32. Suzuki, K.: Overview of deep learning in medical imaging. Radiol. Phys.
Technol. 10(3), 257–273 (2017). https://doi.org/10.1007/s12194017
04065
33. Andina, D., et al.: Deep learning for computer vision: a brief review.
Comput. Intell. Neurosci. 2018, 1–13 (2018). https://doi.org/10.1155/
2018/7068349
34. Momeni, A., Fleury, R.: Electromagnetic wavebased extreme deep
learning with nonlinear timeFloquet entanglement. Nat. Commun. 13(1),
2651 (2022). https://doi.org/10.1038/s41467022302975
35. Harrison, L., et al.: Material identication using a microwave sensor array
and machine learning. Electronics 9(2), 288 (2020). https://doi.org/10.
3390/electronics9020288
36. CovarrubiasMartínez, D., et al.: Classication of plastic materials using
machinelearning algorithms and microwave resonant sensor. J. Elec-
tromagn. Waves Appl. 36(12), 1760–1775 (2022). https://doi.org/10.
1080/09205071.2022.2043192
37. Ting, T.M., Ahmad, N.S., Goh, P.: Material classication via embedded
RF antenna array and machine learning for intelligent mobile robots.
Alex. Eng. J. 106, 60–70 (2024). https://doi.org/10.1016/j.aej.2024.
06.083
38. Simsek, E.: Classication with electromagnetic waves. GitHub (2023).
https://github.com/simsekergun/CwEMW
39. Li, F., Liu, Q.H., Song, L.P.: Threedimensional reconstruction of objects
buried in layered media using Born and distorted Born iterative methods.
Geosci. Rem. Sens. Lett. IEEE 1(2), 107–111 (2004). https://doi.org/10.
1109/lgrs.2004.826562
40. Wei, Z., Chen, X.: Deeplearning schemes for fullwave nonlinear inverse
scattering problems. IEEE Trans. Geosci. Rem. Sens. 57(4), 1849–1860
(2019). https://doi.org/10.1109/tgrs.2018.2869221
41. Puzyrev, V.: Deep learning electromagnetic inversion with convolutional
neural networks. Geophys. J. Int. 218(2), 817–832 (2019). https://doi.
org/10.1093/gji/ggz204
42. LeCun, Y., Cortes, C.: MNIST Handwritten Digit Database. (2010)
43. Hughes, T.W., et al.: Forwardmode differentiation of Maxwell’s equa-
tions. ACS Photonics 6(11), 3010–3016 (2019). https://doi.org/10.1021/
acsphotonics.9b01238
44. Pearson, K.: LIII. On lines and planes of closest t to systems of points
in space. London, Edinburgh Dublin Phil. Mag. J. Sci. 2(11), 559–572
(1901). https://doi.org/10.1080/14786440109462720
45. Gisbrecht, A., Schulz, A., Hammer, B.: Parametric nonlinear dimen-
sionality reduction using kernel tSNE. Neurocomputing 147, 71–82
(2015). Advances in SelfOrganizing Maps Subtitle of the special issue:
Selected Papers from the Workshop on SelfOrganizing Maps 2012
(WSOM 2012). Available from: https://doi.org/10.1016/j.neucom.2013.
11.045
46. Van der Maaten, L., Hinton, G.: Visualizing data using tSNE. J. Mach.
Learn. Res. 9(11) (2008)
47. Pedregosa, F., et al.: Scikitlearn: machine learning in Python. J. Mach.
Learn. Res. 12, 2825–2830 (2011)
48. xgboost developers: XGBoost, An Optimized Distributed Gradient
Boosting Library. https://xgboost.readthedocs.io/ (2022)
49. Chen, T., Guestrin, C.: XGBoost: a scalable tree boosting system. In:
Proceedings of the 22nd ACM SIGKDD International Conference on
Knowledge Discovery and Data Mining, pp. 785–794. ACM, New York
(2016). KDD ’16
50. Bergstra, J., Bengio, Y.: Random search for hyperparameter optimiza-
tion. J. Mach. Learn. Res. 13(2) (2012)
51. Chollet, F., et al. Keras (2015). https://keras.io
52. Martin, A., et al.: TensorFlow: LargeScale Machine Learning on Het-
erogeneous Systems (2015). Software available from tensorow.org.
https://www.tensorow.org/
53. He, K., et al.: Delving deep into rectiers: surpassing humanlevel per-
formance on ImageNet classication. In: 2015 IEEE International
Conference on Computer Vision (ICCV), pp. 1026–1034 (2015)
54. Rumelhart, D.E., McClelland, J.L.: Learning Internal Representations by
Error Propagation, pp. 318–362 (1987)
SIMSEK and MANYAM
-
909
55. Brown, P., et al.: A statistical approach to language translation. In: Coling
Budapest 1988 Volume 1: International Conference on Computational
Linguistics (1988)
56. Kingma, D.P., Ba, J.: Adam: A Method for Stochastic Optimization.
(2014)
57. Tsoumakas, G., Katakis, I.: Multilabel classication: an overview. Int. J.
Data Warehous. Min. 3(3), 1–13 (2007). https://doi.org/10.4018/jdwm.
2007070101
58. Kaczmarek, M., Buchnev, O., Nandhakumar, I.: Ferroelectric nano-
particles in low refractive index liquid crystals for strong electrooptic
response. Appl. Phys. Lett. 92(10), 103307 (2008). https://doi.org/10.
1063/1.2884186
How to cite this article: Simsek, E., Manyam, H.R.:
Classication with electromagnetic waves. IET Microw.
Antennas Propag. 18(12), 898–910 (2024). https://doi.
org/10.1049/mia2.12522
910
-
SIMSEK and MANYAM
Article
Full-text available
In electromagnetic inverse scattering, the goal is to reconstruct object permittivity using scattered waves. While deep learning has shown promise as an alternative to iterative solvers, it is primarily used in supervised frameworks which are sensitive to distribution drift of the scattered fields, common in practice. Moreover, these methods typically provide a single estimate of the permittivity pattern, which may be inadequate or misleading due to noise and the ill-posedness of the problem. In this paper, we propose a data-driven framework for inverse scattering based on deep generative models. Our approach learns a low-dimensional manifold as a regularizer for recovering target permittivities. Unlike supervised methods that necessitate both scattered fields and target permittivities, our method only requires the target permittivities for training; it can then be used with any experimental setup. We also introduce a Bayesian framework for approximating the posterior distribution of the target permittivity, enabling multiple estimates and uncertainty quantification. Extensive experiments with synthetic and experimental data demonstrate that our framework outperforms traditional iterative solvers, particularly for strong scatterers, while achieving comparable reconstruction quality to state-of-the-art supervised learning methods like the U-Net.
Article
Full-text available
In this paper, we propose the neural Born iterative method (NeuralBIM) for solving 2D inverse scattering problems (ISPs) by drawing on the scheme of physics-informed supervised residual learning (PhiSRL) to emulate the computing process of the traditional Born iterative method (TBIM). NeuralBIM employs independent convolutional neural networks (CNNs) to learn the alternate update rules of two different candidate solutions regarding the residuals. Two different schemes are presented in this paper, including the supervised and unsupervised learning schemes. With the data set generated by the method of moments (MoM), supervised NeuralBIM are trained with the knowledge of total fields and contrasts. Unsupervised NeuralBIM is guided by the physics-embedded objective function founding on the governing equations of ISPs, which results in no requirement of total fields and contrasts for training. Numerical and experimental results further validate the efficacy of NeuralBIM.
Article
Full-text available
Wave-based analog signal processing holds the promise of extremely fast, on-the-fly, power-efficient data processing, occurring as a wave propagates through an artificially engineered medium. Yet, due to the fundamentally weak non-linearities of traditional electromagnetic materials, such analog processors have been so far largely confined to simple linear projections such as image edge detection or matrix multiplications. Complex neuromorphic computing tasks, which inherently require strong non-linearities, have so far remained out-of-reach of wave-based solutions, with a few attempts that implemented non-linearities on the digital front, or used weak and inflexible non-linear sensors, restraining the learning performance. Here, we tackle this issue by demonstrating the relevance of time-Floquet physics to induce a strong non-linear entanglement between signal inputs at different frequencies, enabling a power-efficient and versatile wave platform for analog extreme deep learning involving a single, uniformly modulated dielectric layer and a scattering medium. We prove the efficiency of the method for extreme learning machines and reservoir computing to solve a range of challenging learning tasks, from forecasting chaotic time series to the simultaneous classification of distinct datasets. Our results open the way for optical wave-based machine learning with high energy efficiency, speed and scalability. Wave-based analog signal processing has been challenging for complex nonlinear operations such as data forecasting or classification. The authors propose here an analog neuromorphic platform for optical wave-based machine learning characterized by energy efficiency, speed and scalability.
Article
Full-text available
The identification of different plastic materials in pellet form using a microwave negative-order-resonance sensor along with the evaluation of classification algorithms from machine learning is presented in this paper. Operation of the sensor is within the unlicensed ISM 2.5 GHz band, and identification of the materials is based on the measured resonant parameters from the sensor. Several classifiers are used to process the resonant parameters having uncertainty factors involved in pellet measurements (air gaps, pellet positions, dimensions and shapes), and performance comparison between the algorithms is carried out in terms of accuracy in the classification. Moreover, the presented measurement method is proposed as a fast, non-destructive, and low-power consumption way to identify plastic raw materials using a low-profile circuit having a high potential of being used in industrial processes.
Article
Full-text available
In this work, we design an iterative deep neural network to solve full-wave inverse scattering problems in the 2D case. Forward modeling neural networks that predict the scattered field are embedded in an inversion neural network. In an iterative manner, the inversion network predicts the model update from the residual between the simulated data and the observed data. The proposed inversion network can achieve super-resolution reconstruction, meanwhile keeping the simulated data of reconstructed models well consistent with the observed data. We validate this method with both synthetic and experimental data inversion. Results show that the inversion network can predict models with high accuracy, efficiency, and good generalization ability. By combining deep learning and physical simulation together, the proposed method provides a way for real-time imaging with high reliability and accuracy.
Article
Full-text available
Without phase information of the measured field data, the phaseless data inverse scattering problems (PD-ISPs) counter more serious nonlinearity and ill-posedness compared to full data ISPs (FD-ISPs). In this paper, we propose a learning-based inversion approach in the frame of the U-net convolutional neural network (CNN) to quantitatively image unknown scatterers located in homogeneous background from amplitude-only measured total field (also denoted phaseless data). Three training schemes with different inputs to the U-net CNN are proposed and compared, i.e., the direct inversions (DIS) scheme with phaseless total field data, retrieval dominant induced currents by Levenberg-Marquardt (LM) method (PD-DICs) and phaseless data with contrast source inversion scheme (PD-CSI). We also demonstrate the setup of training data and compare the performance of the three schemes using both numerical and experimental tests. It is found that the proposed PD-CSI and PD-DICs perform better in terms of accuracy, generalization ability and robustness compared to DIS. PD-CSI has the strongest capability to tackle with PD-ISPs, which outperforms the PDDICs and DIS.
Article
In this paper, a novel artificial neural network named Residual U-Net (ResU-Net) is proposed to directly reconstruct three-dimensional (3-D) dielectric anisotropic objects from scattered electromagnetic field data recorded at the receiver array. ResU-Net has the same framework as that of U-Net but the convolution kernels are replaced with residual kernels. Meanwhile, the squeeze-and-excitation (SE) operation is added to enable information interaction among different channels and further improve prediction accuracy. ResU-Net is trained by thousands of 3-D homogeneous dielectric anisotropic handwritten digits and the corresponding synthesized scattered field data. In the online prediction, ResU-Net can invert for multiple anisotropic model parameters of homogeneous 3-D objects instantaneously. For an inhomogeneous object or multiple homogeneous objects, ResU-Net provides good initial profiles which are fed into the following variational Born iterative method (VBIM) full-wave inversion solver. In addition, the VBIM is implemented in a restricted domain instead of the whole 3-D inversion domain to save computational cost. Numerical experiments show that, compared with the traditional iterative solver such as VBIM, the proposed ResU-Net or the hybrid method not only can achieve higher reconstruction accuracy but also accomplish the multiparametric 3-D inversion in a much faster way.
Article
This work aims to enhance our fundamental understanding of how the measurement setup used to generate training and testing datasets affects the accuracy of the machine learning algorithms that attempt solving electromagnetic inversion problems solely from data. A systematic study is carried out on a one-dimensional semi-inverse electromagnetic problem, which is estimating the electrical permittivity values of a planarly layered medium with fixed layer thicknesses assuming different receiver-transmitter antenna combinations in terms of location and numbers. Accuracy of the solutions obtained with four machine learning methods including neural-networks is compared with a physics-based solver deploying the Nelder-Mead simplex method to achieve the inversion iteratively. Numerical results show that (i) deep-learning outperforms the other machine learning techniques implemented in this study, (ii) increasing number of antennas and placing them as close as possible to the domain of interest increase inversion accuracy, (iii) for neural networks, training datasets created on random grids lead to a more efficient learning than the training datasets created on uniform grids, and (iv) multi-frequency training and testing with a few antennas can achieve more accurate inversion than single-frequency setups deploying several antennas.
Article
In recent years, deep learning (DL) is becoming an increasingly important tool for solving inverse scattering problems (ISPs). This paper reviews methods, promises, and pitfalls of deep learning as applied to ISPs. More specifically, we review several state-of-the-art methods of solving ISPs with DL, and we also offer some insights on how to combine neural networks with the knowledge of the underlying physics as well as traditional non-learning techniques. Despite the successes, DL also has its own challenges and limitations in solving ISPs. These fundamental questions are discussed, and possible suitable future research directions and countermeasures will be suggested.