ArticlePDF Available

De-Aliasing and Accelerated Sparse Magnetic Resonance Image Reconstruction Using Fully Dense CNN with Attention Gates

Authors:

Abstract and Figures

When sparsely sampled data are used to accelerate magnetic resonance imaging (MRI), conventional reconstruction approaches produce significant artifacts that obscure the content of the image. To remove aliasing artifacts, we propose an advanced convolutional neural network (CNN) called fully dense attention CNN (FDA-CNN). We updated the Unet model with the fully dense connectivity and attention mechanism for MRI reconstruction. The main benefit of FDA-CNN is that an attention gate in each decoder layer increases the learning process by focusing on the relevant image features and provides a better generalization of the network by reducing irrelevant activations. Moreover, densely interconnected convolutional layers reuse the feature maps and prevent the vanishing gradient problem. Additionally, we also implement a new, proficient under-sampling pattern in the phase direction that takes low and high frequencies from the k-space both randomly and non-randomly. The performance of FDA-CNN was evaluated quantitatively and qualitatively with three different sub-sampling masks and datasets. Compared with five current deep learning-based and two compressed sensing MRI reconstruction techniques, the proposed method performed better as it reconstructed smoother and brighter images. Furthermore, FDA-CNN improved the mean PSNR by 2 dB, SSIM by 0.35, and VIFP by 0.37 compared with Unet for the acceleration factor of 5.
Content may be subject to copyright.
Bioengineering 2023, 10, 22. https://doi.org/10.3390/bioengineering10010022 www.mdpi.com/journal/bioengineering
Article
De-Aliasing and Accelerated Sparse Magnetic Resonance
Image Reconstruction Using Fully Dense CNN with
Attention Gates
Md. Biddut Hossain 1, Ki-Chul Kwon 1, Shariar Md Imtiaz 1, Oh-Seung Nam 1, Seok-Hee Jeon 2 and Nam Kim 1,*
1 School of Information and Communication Engineering, Chungbuk National University,
Cheongju-si 28644, Chungcheongbuk-do, Republic of Korea
2 Department of Electronics Engineering, Incheon National University, 119 Academy-ro, Yeonsu-gu,
Incheon 22012, Gyeonggi-do, Republic of Korea
* Correspondence: namkim@chungbuk.ac.kr; Tel.: +82-043-261-2482
Abstract: When sparsely sampled data are used to accelerate magnetic resonance imaging (MRI),
conventional reconstruction approaches produce significant artifacts that obscure the content of the
image. To remove aliasing artifacts, we propose an advanced convolutional neural network (CNN)
called fully dense attention CNN (FDA-CNN). We updated the Unet model with the fully dense
connectivity and attention mechanism for MRI reconstruction. The main benefit of FDA-CNN is
that an attention gate in each decoder layer increases the learning process by focusing on the rele-
vant image features and provides a better generalization of the network by reducing irrelevant ac-
tivations. Moreover, densely interconnected convolutional layers reuse the feature maps and pre-
vent the vanishing gradient problem. Additionally, we also implement a new, proficient under-
sampling pattern in the phase direction that takes low and high frequencies from the k-space both
randomly and non-randomly. The performance of FDA-CNN was evaluated quantitatively and
qualitatively with three different sub-sampling masks and datasets. Compared with five current
deep learning-based and two compressed sensing MRI reconstruction techniques, the proposed
method performed better as it reconstructed smoother and brighter images. Furthermore, FDA-
CNN improved the mean PSNR by 2 dB, SSIM by 0.35, and VIFP by 0.37 compared with Unet for
the acceleration factor of 5.
Keywords: aliasing artifacts; attention gate; deep learning; fully dense network; MRI reconstruction
1. Introduction
Magnetic resonance imaging (MRI) is a sophisticated and noninvasive medical im-
aging technique that has full control over data capture to visualize the anatomy and func-
tions of the human brain and body [1]. It plays a significant role in smart healthcare sys-
tems alongside medical research by providing high-quality reconstructed images without
exposure to harmful radiation [2]. However, the long image acquisition time (more than
30 min per patient) [3] is a disadvantage compared with X-ray, computed tomography
(CT), and photoacoustic tomography (PAT) imaging modalities. In MRI, frequencies are
acquired in the k-space rather than directly in the image space. Full k-space data are re-
quired for high-quality reconstructed images, but this prolongs the acquisition process.
This slow process causes the patient discomfort and generates motion artifacts due to pa-
tient movement. A common way to speed up MRI acquisition is by taking fewer scans
and reconstructing the image using a partially recorded k-space. However, this creates
blurring and aliasing artifacts in the images according to the Nyquist–Shannon principle
[4]. Hence, an effective acceleration technique is essential for MRI reconstruction. Parallel
imaging (PI) [5,6] applies coil sensitivity maps to speed up reconstruction, but the setup
Citation: Hossain, M.B.; Kwon,
K.-C.; Imtiaz, S.M.; Nam, O.-S.;
Jeon, S.
-H.; Kim, N. De-Aliasing and
Accelerated Sparse Magnetic
Resonance Image Reconstruction
Using Fully Dense CNN with
Attention Gates. Bioengineering 2023,
10, 22. https://doi.org/10.3390/
bioengineering10010022
Academic Editor: Or Perlman and
Efrat Shimron
Received: 25 November 2022
Revised: 19 December 2022
Accepted: 19 December 2022
Published: 22 December 2022
Copyright: © 2022 by the authors. Li-
censee MDPI, Basel, Switzerland.
This article is an open access article
distributed under the terms and con-
ditions of the Creative Commons At-
tribution (CC BY) license (https://cre-
ativecommons.org/licenses/by/4.0/).
Bioengineering 2023, 10, 22 2 of 18
process is complex and expensive. Compressed sensing (CS) MRI [7,8] uses sparse data to
reconstruct high-resolution images from randomly picked k-space data. However, the
sparsity of data produces aliasing artifacts, and the computing time required is quite high
for their iteration process.
Deep learning (DL) has been successfully adopted in medical image processing [9–
12] and deals with the obstacles of iterative CS reconstruction [13,14]. Efficient reconstruc-
tion can be performed without repeatedly processing after properly training a DL model.
Several post-processing strategies [15–18] based on generative adversarial networks
(GANs) have been used to improve image quality in the image domain, whereas some
approaches [19–22] estimate the unknown missing k-space frequencies in the sensor do-
main. Cross-domain techniques [23–29] operate in both the sensor and image domains.
Recently, iterative unrolled optimization methods [30–34] have been used for sparse MRI
reconstruction to improve the learning of image features. However, when the k-space is
uniformly under-sampled, DL-based iterative methods cannot properly eliminate aliasing
artifacts.
The first multilayer perceptron [35] is applied to reduce aliasing artifacts in parallel
MRI. The multi-scale Unet structure [36] has been suggested to handle phase image re-
construction and scattered artifacts. The convolutional neural network (CNN)-based two-
dimensional (2D) phase-contrast MRI reconstruction method [37] encodes low-frequency
regions in the phase direction, while high-frequency regions store image edges. The GAN-
based de-aliasing method [38] combines adversarial and innovative content loss but cal-
culates fast Fourier transforms (FFTs) of magnitude images instead of raw MRI data. Re-
fineGAN [39] measures only the cyclic loss of the training dataset using an autoencoder.
Based on deep residual learning (DLR), the denoising technique [40] identifies and then
subtracts noise from noisy images. After modifying this DLR-based network, Ouchi et al.
[41] proposed a model to generate alias-free images by subtracting the predicted alias
components from them.
Here, we adopt a post-processing strategy and propose an efficient fully dense atten-
tion convolutional neural network (FDA-CNN) architecture to remove aliasing and ghost
artifacts from sparsely reconstructed 2D MRI images. The FDA-CNN method generates
aliasing-free images instead of prediction artifacts, which is a key difference from previ-
ous MRI de-aliasing techniques. Guan et al. [42] proposed fully dense Unet (FD-Unet) for
reducing artifacts from PAT images. We use the modified FD-Unet architecture for MRI
reconstruction by adding an attention gate (AG) to each decoder layer. FDA-CNN consists
of an AG [43] in each layer of the decoding path and dense connectivity (DC) [44] in both
the encoding and decoding layers. DC solves the vanishing gradient issue and improves
information flow, whereas the AG gives more attention to essential features that make the
network more generalized. Supplementarily, we implement a hybrid mixed center and
periphery under-sampling (MCP-US) mask in the phase [ reconstructs complex image tex-
tures and contents better than the two-dimensional (2D) and one-dimensional (1D) Gauss-
ian under-sampling patterns. The results of FDA-CNN were evaluated using three sub-
sampling masks using only 20% of the data. The simulation results were compared to both
the single- and multidomain networks along with traditional CS methods on three widely
used k-space datasets.
2. Methodology
The data obtained from an MRI scanner are known as the Fourier space or k-space
and can be represented as follows:
y
=
FT (1)
where T is the artifact-free image, F is the Fourier transform, and y represents the fully
sampled k-space data. The reconstruction of image x from the k-space can be performed
by applying the inverse Fourier transform.
Bioengineering 2023, 10, 22 3 of 18
x
=
F–1
y
(2)
The reconstruction of an image with n pixels requires at least n frequencies. However,
it takes more scan time to capture n frequencies, because the number of frequencies ac-
quired is linearly related to the length of the MRI scan. Acquiring only 1/5 of the frequen-
cies can reduce both the scan time and cost by a factor of 5, but aliasing artifacts are intro-
duced. Under-sampled k-space y is represented as
y
=
U
y
=
U
FT (3)
Figure 1 exhibits a reconstructed artifact xaliased image from the under-sampled k-
space. Here, U is a sub-sampling matrix filled with 0 s (black area) and 1 s (white line),
and expresses the element-wise multiplication.
Figure 1. Reconstructed artifact image from sparsely sampled k-space.
Conventional CS methods recover images by solving the following optimization
problem:
|
y
–y
|2
2
T
min + λG
(
T
)
(4)
where ||.||2 is a generic data consistency term that assures the solution of the original
image (T) with every observation of y and G denotes the regularization term, where λ
expresses the regularization parameter. The challenge of traditional CS is that the regu-
larizer, G(T), must be manually encoded to represent the reconstructed MR images.
2.1. Deep Learning Framework
A DL-based method is applied to adjust the under-sampled image and reconstruct
an artifact-free image that is close to the actual image. This is called a supervised learning
strategy, which aims to find an appropriate reconstruction function that matches the given
image with the expected output. This can be expressed with the following equation:
R
=
{(NN)L(F–1(y
))} (5)
where R is the reconstructed image, NN is a DL-based neural network, and L measures
the loss between the original and reconstructed images.
As shown in Figure 2, the fully sampled k-space (y) is sub-sampled by means of ele-
ment-wise multiplication with the under-sampling mask (U). First, an image with aliasing
and ghost artifacts x is reconstructed using the inverse fast Fourier transform (IFFT) from
sub-sampled k-space y. Then, these artifacts are removed by FDA-CNN by decreasing the
loss (L) between the predicted image (R) and the target image (T). FDA-CNN attempts to
recover an image F-1(y) that is close to the target image using sub-sampled sensor data y
as input. Gradient descent is used to optimize the parameters of the loss function.
Bioengineering 2023, 10, 22 4 of 18
Figure 2. Flowchart of the proposed de-aliasing technique.
2.2. Proposed FDA-CNN Architecture
The proposed MRI de-aliasing network architecture is presented in this section. We
designed an improved CNN model based on the Unet [45] architecture by combining
modified dense connectivity and AGs. We applied batch normalization [46] to accelerate
training compared with earlier Unet implementations.
The network structure shown in Figure 3 has two main sections: the down- and up-
sampling parts, and the skip connection part with the AGs. The down-sampling section
consists of convolution, dense block, and max-pooling layers, whereas up-sampling con-
sists of upscaling (2 × 2 deconvolution), dense block, and convolutional layers. Firstly, a
1-channel 3 × 3 convolutional operation is applied on a 256 × 256 input image with rectified
linear unit (ReLU) [47] activation function. Then, five consecutive dense blocks (DB) are
used. The first DB starts with 32 channels and gradually increases by 64, 128, 256, and 512.
Every DB consists of a series of 1 × 1 and 3 × 3 convolutions with padding 1, batch nor-
malization, and ReLU. Initially, hyperparameters km and fm are specified by the user;
for our method, we initialized k1
=
8 and f1 = 64. Then, km and fm are changed by km=
2m-1× k1 and fm= 2m-1× f1, respectively. The concatenation of the inputs and outputs from
every layer of the DB generates its final output. Except for the last DB of the encoding
section, a max-pooling operation is executed after each DB, which halves the size of the
input at each level and doubles the number of feature maps.
The decoding or up-sampling section restores the size of the feature maps and main-
tains a form symmetric to the encoding section. This symmetry enables the reuse of fea-
tures by concatenating feature maps at the same level and reduces the loss of information
caused by the encoding/decoding process. Before concatenation, the features of encoding
and decoding layers go through the AG to focus on target features from different spatial
information. Every layer of the decoding section executes a 1 × 1 convolution with pad-
ding 1 and ReLU before going into a DB. A 1-channel 1 × 1 convolution is executed before
generating the final output.
Bioengineering 2023, 10, 22 5 of 18
Figure 3. Proposed FDA-CNN architecture.
Moreover, dense connectivity generates deeper networks. For comparison, Unet has
23 layers, while FDA-CNN has 97 convolutional and deconvolutional levels. The vanish-
ing gradient problem arises because the gradient information must flow through different
layers and may disappear before it arrives at the succeeding layers. Dense connectivity
adds more links to enable the effective backpropagation of gradient information. This less-
ens the vanishing gradient issue and makes it easier to train the network.
2.3. Dense Block
Densely connected networks [48] maximize the capability of the network by reusing
features. The input of the succeeding layers is more varied and more effective when fea-
ture maps from various layers are combined. In our method, a dense block with a growth
rate, km, is used to learn different feature maps, fm, for each spatial level, m. Initially,
hyperparameters km and fm are specified by the user. Then, km and fm are changed by
km= 2m-1× k1 and fm= 2m-1× f1, respectively, at each spatial level to preserve computational
efficiency and ensure that each dense block has the same number of convolutional layers.
A total of nine dense blocks with four layers are used in the FDA-CNN approach.
As shown in Figure 4, the Lth layer of the dense block has an initial input with
F + × (L-1) feature maps and output with km feature maps, where F is the total number
of feature maps in the dense block’s initial input. Through a series of 1 × 1 and 3 × 3 con-
volutions with batch normalization plus ReLU activation function, features are learned.
Due to the increased computational complexity of the 3 × 3 convolution, the input dimen-
sion is decreased to F feature maps by applying a 1 × 1 convolution, which increases con-
vergence speed. Then, using a 3 × 3 convolution, km attribute maps are developed from
the compacted data. The concatenation of the inputs and outputs from every layer of the
dense block generates the dense block’s final output.
Figure 4. Last dense block with four layers of encoding part, where k5 = 128 and f5 = 512.
Bioengineering 2023, 10, 22 6 of 18
2.4. Attention Gate
Models trained with AGs [49] intuitively learn to emphasize prominent features that
are helpful for a particular task while suppressing irrelevant regions in an input image.
With no additional computational work, AGs may be quickly added to common CNNs,
such as Unet topologies, improving model sensitivity and prediction accuracy. Unet em-
ploys skip connections to merge spatial data from the up- and down-sampling paths.
Low-quality feature representation exists in the first few layers, which carries in several
redundant low-level feature extractions. By actively suppressing activations in unneces-
sary regions through the use of AGs at the skip connections, the number of redundant
features transferred is decreased. Every AG takes two inputs, g and x. The gating signal,
g, comes from the next lowest layer of the network. It has greater feature representation
because it originates from a deeper region of the network. The input features, x, come
from skipped connections. They have better spatial information because they originate
from the early stages.
We incorporate an AG with every decoding part in our fully dense Unet framework.
As shown in Figure 5, input features xi
l perform 1 × 1 × 1 convolutions with stride 2 × 2 to
lessen the size of the dimensions (H × W) by half, and gating signals gi
l+1 perform 1 × 1 ×
1 convolutions with stride 1 × 1. As a result, the spatial geometry of the modified input
features and gating signals is the same. The ReLU function activates them through ele-
ment-wise summation and maps them by Wint
T into a smaller-dimensional space for gat-
ing operations. The sigmoid function levels the vector in (0, 1), with coefficients closer to
1 denoting more pertinent features. Then, a trilinear up-sampler is used to restore the size
of attention weight matrix αi
l to correspond to the pixel density of the input features. The
output of the AG, xi
l, is generated by means of element-wise multiplication between at-
tention weight matrix αi
l and input features xi
l and then is transmitted as usual through
the residual connections.
Figure 5. Schematic diagram of an attention gate.
3. Deep Learning Implementation
3.1. Datasets and Under-Sampling Masks
We used fully sampled brain k-space data from BraTs-2020 [50], fastMRI [51,52], and
IXI [53]. We used the cross-sectional T1-weighted BraTS-2020 dataset for both the training
and testing of all the networks. On the other hand, the T1-weighted axial fastMRI and T2-
weighted coronal IXI datasets were only used for testing. Each volume possessed both the
fully acquired k-space data and the associated reconstructed images of the same size (256
× 256). As we concentrated on the correlation between the number of k-space slices and
FDA-CNN performance, no data augmentation was used in training. Images were recon-
structed sequentially from every k-space.
During training, the sub-sampled zero-filled (ZF) noisy and artifact images were
used as the network input along with the fully sampled images as target images. Our new
MCP-US pattern was used for training and compared with 2D Gaussian under-sampling
(2DG-US) and 1D Gaussian under-sampling (1DG-US) distributions. Mostly, 2D-US and
1DG-US focus on the central low frequencies of the k-space. However, low-spatial-fre-
quency data, which determine the overall contrast, brightness, and form of the image, are
located in the center of the k-space. On the other hand, high-spatial-frequency data deter-
mining the image edges and details are located in the periphery of the k-space. As the k-
Bioengineering 2023, 10, 22 7 of 18
space has a symmetric nature, our MCP-US takes both the low and high frequencies.
Among the total sampled data (S) of each k-space, MCP-US constantly samples 50% of
the center (sc) and 25% of the periphery (sp) and randomly chooses 25% of data (sr) fre-
quencies, except for center and periphery data. It can be expressed as follows:
S = sc+sp+sr
where sc=
S
2%, sp=
S
4%, sr=
S
4%
In both training and testing, we took only 20% of data of each k-space for all three sub-
sampling patterns, where white spaces were replaced with zero, as shown in Figure 6.
Figure 6. Several sampling patterns: (a) fully sampled; (b) 2D Gaussian pattern; (c) 1D Gaussian
pattern; (d) mixed center and periphery under-sampling pattern.
Among the sampled data (S), our MCP-US non-randomly takes 10% of sc from the
middle position and 5% of sp from the zero position and randomly chooses 5% of sr be-
tween the sc and sp areas. As we concentrated on the correlation between the number of
training images and FDA-CNN performance, no data augmentation was used in training.
3.2. Loss Function
The disparity between the sub-sampled aliasing image and the fully sampled alias-
ing-free image was evaluated using the loss function. The optimum objective of FDA-
CNN is to minimize the value of the loss function. Smaller values between the under-
sampled and fully sampled images ensure better reconstruction. We used the mean square
error (MSE) as the loss function to calculate pixel-wise disparity and update the network
parameters, which can be expressed as follows:
LossMSE = 1
N
(Ti
N
i = 1 –Ri)2 (6)
where N indicates the number of voxels (or pixels) in the image, and Ti and Ri represent
the target and reconstructed MR images, respectively.
3.3. Performance Evaluations Metrics
To evaluate the network performance, we summarized the findings using four pa-
rameters: structural similarity index measure (SSIM) [54], peak signal-to-noise ratio
(PSNR), normalized root mean squared error (NRMSE), and pixel visual information fi-
delity (VIFP) [55]. The SSIM is a perceptual index that utilizes the mutual dependencies
among adjacent pixels to measure the similarity of two images, such as brightness, con-
trast, and structural properties. The following expression gives the SSIM between the net-
work output (R) and the desired output (T):
SSIM
(
T, R
)
=
TµR+c1
(
TR+c1
)
µT
2R
2+c1

σT
2R
2+c1
(7)
where µT and µR represent the mean values of T and R, respectively; and σT
2 and σR
2
denote the corresponding pixel variance values. The covariance value is also shown by
σTR. To stabilize the division, c1 and c2 have the following definitions:
Bioengineering 2023, 10, 22 8 of 18
c1
=
(0.01P)2, c2
=
(0.03P)2
where P = max(T) min(T)
The PSNR calculates the ratio of the signal’s highest potential power (image intensity
throughout a volume) to its fidelity-affecting distorting noise power. This can be ex-
pressed as
PSNR(T, R) = 10log10
255
1
N
(Ti
N
i = 1 –Ri)2
(8)
The ground truth and the pixel differences in network output images are compared
by the NRMSE, which can be expressed as
NRMSE
(
T, R
)
=
1
N
(Ti
N
i = 1 –Ri)2
max
(
T
)
min
(
R
)
(9)
The human viewer’s perceptual evaluation approach, VIFP, measures image infor-
mation by computing two mutual information quantities from the reference and distorted
images. This can be defined as
VIFP
(
T,R
)
=
I
CN,j; RN,j
sN,j
j
subbands
I
CN,j; TN,j
sN,j
j
subbands
(10)
where RN,j, and TN,j represent the sub-bands of the reconstructed and target images,
respectively; SN,j defines a realization for a specific image; and CN,j expresses N ele-
ments of random field Cj that specifies the coefficient of the sub-band, j. The evaluation
result of VIFP is indicated as values between 0 and 1, similar to the SSIM.
These criteria were chosen as they are typically used to evaluate image reconstruc-
tion. Higher values of SSIM, PSNR, and VIFP indicate better results, while smaller values
of the NRMSE define better reconstructions. Moreover, the reconstruction time for each
image indicates the transformation of MRI raw data into pictures. The reconstruction time
of each method was calculated using the MCP-US pattern.
3.4. Experimental Setup
The training and testing of FDA-CNN were executed on an Intel Core i7-9800X 3.80
GHz processor and 128 GB memory with NVIDIA GeForce RTX 2080 Ti GPU running on
Windows 10 Pro 64 bit. This model was implemented in Python 3.8 with the DL open-
source libraries TensorFlow v2.4 and Keras v2.4 in the PyCharm environment. We used
the Adam optimizer with momentum values of β1 = 0.9 and β2 = 0.999 to reduce the loss
function. The starting learning rate was 1e-4, and it declined with a decay factor of 0.95
every 20 epochs. A small batch size of 8 was chosen, and 2000 epochs were executed for
the training of our network. Figure 7 illustrated the training and validation losses of the
proposed network, where the regularizing impact of extensive connectivity lessens the
possibility of overfitting the training data.
Bioengineering 2023, 10, 22 9 of 18
Figure 7. Training and validation losses of proposed FDA-CNN.
4. Result and Discussion
The performance of our network was compared to classical CS total variation (TV)
[56], wavelet [57] denoising algorithms, and DL-based state-of-the-art (image and dual
domain) reconstruction methods. Lightweight autoencoder (LAE) [58], basic Unet [46],
projection-based cascade Unet (PBCU) [59], and DRL-based MRI (DRL-net) [41] recon-
structions are image-domain DL networks. LAE uses an autoencoder framework, and
PBCU uses five cascade Unets for MRI reconstruction. DRL-net subtracts the predicted
artifacts from the under-sampled aliased images. The multidomain MRI reconstruction
strategy (Wnet) [60] uses two Unets: one for the k-domain and another for the image do-
main. SSIM, PSNR, and NRMSE were used for quantitative analysis, where VIFP was used
to evaluate the perception of the de-aliased images of a human viewer. The average re-
construction times were calculated using the MCP-US pattern. All results were generated
in the same environment.
4.1. BraTs 2020-T1 Dataset
The T1-weighted axial brain Brats-2020 dataset was used for both the training and
testing of all the networks. In an ideal dataset, the training and test data are very well
correlated, providing an opportunity to acquire, from the training data, most of the fea-
tures that are required to perform effectively during testing. The efficiency of CNNs can
be compared in this ideal situation without being affected by the data. This BraTS-2020
dataset was obtained using a clinical 3T multimodal MR scanner. Among 150 k-spaces,
100 k-spaces were used for training; 30 k-spaces were used for validation; and 20 k-spaces
were used for testing. Each k-space contained 155 axial cross-sectional T1-weighted (256
× 256) images. In the training, validation, and test sets, there was no duplication of the
same k-space.
In this experiment, the same MRI sequence was used for both training and testing,
with 25% of the training data being used for validation to increase the reliability of the
results. The learning potential of FDA-CNN to eliminate artifacts was measured by ad-
justing the hyperparameters (feature maps and growth rate) of the dense block. Compar-
ative DL methods were trained using the MCP-US pattern and evaluated on the associated
datasets with each under-sampling mask. The efficiencies of all CNNs in terms of elimi-
nating artifacts were compared using various sub-sampling masks for an acceleration fac-
tor of 5. In general, the CNN produced a better image with minimal artifacts. As seen in
Table 1, the proposed FDA-CNN produced higher average SSIM, PSNR, and VIFP with a
lower average NRMSE than traditional CS and the autoencoder and Unet-based methods.
Bioengineering 2023, 10, 22 10 of 18
Table 1. Average SSIM, PSNR, NRMSE, and VIFP values and reconstruction time of state-of-art
methods on BraTs-T1 axial brain data using several sampling patterns.
Sampling
Metrics
Zero Filling
TV
Wavelet
DRL-Net
FDA-CNN
2DG-US
SSIM 0.31 0.32 0.36 0.78 0.79 0.77 0.82 0.83 0.89
PSNR 34.11 35.90 36.13 35.16 35.31 35.29 35.55 35.84 36.22
NRMSE 0.09 0.07 0.06 0.08 0.08 0.07 0.07 0.06 0.04
VIFP 0.60 0.50 0.65 0.87 0.92 0.93 0.90 0.80 0.94
1DG-US
SSIM 0.52 0.58 0.62 0.68 0.69 0.68 0.69 0.69 0.70
PSNR 33.20 33.23 33.25 33.23 33.26 33.10 33.15 33.40 33.59
NRMSE 0.12 0.12 0.12 0.12 0.12 0.12 0.12 0.11 0.10
VIFP 0.25 0.21 0.27 0.42 0.38 0.44 0.38 0.41 0.45
MCP-US
SSIM 0.76 0.77 0.76 0.96 0.95 0.96 0.96 0.95 0.97
PSNR 36.70 36.47 36.73 38.92 38.78 38.97 38.74 38.97 40.55
NRMSE 0.05 0.06 0.06 0.03 0.04 0.03 0.04 0.03 0.02
VIFP 0.52 0.41 0.50 0.90 0.91 0.91 0.89 0.85 0.93
Reconstruction time (seconds) 0.97 0.5 0.30 0.30 0.39 0.33 0.31 0.30
The classical CS methods required almost 0.5 s and 0.97 s in reconstruction time for
each slice, and the dual-domain network required approximately 0.4 s. In contrast, the
single-domain post-processing methods required approximately 0.30 s to 0.33 s. Our pro-
posed method generated better images than Unet within the same reconstruction time of
0.30 s for each slice.
As shown in Figure 8, the under-sampled image contained noise and aliasing arti-
facts. Basic Unet improved the image quality by reducing these artifacts, although some
artifacts remained. Our method effectively removed most of the artifacts and recon-
structed the images close to the reference images from the BraTs testing dataset. The per-
formance of three under-sampling masks on BraTs testing data is shown in Figure 9,
where MCP-US generated better PSNR and SSIM than other under-sampling patterns,
except for VIFP using 2DG-US.
Figure 8. Reconstructed MRI images from the BraTs testing dataset (slice No. 140) using zero filling,
Unet, and FDA-CNN: (a) 2D Gaussian distribution; (b) 1D Gaussian distribution; (c) mixed center
and peripheral mask; (d) fully sampled ground truth image.
Bioengineering 2023, 10, 22 11 of 18
Figure 9. Comparison of three sampling patterns using FDA-CNN.
4.2. FastMRI and IXI Datasets
The second experiment used the fastMRI and IXI datasets to test the CNNs after they
had been trained on the BraTs dataset. This represents a scenario using different training
and test datasets that are not perfectly matched. For testing, fastMRI comprised 2560 T1-
weighted axial brain images from 160 k-spaces, while IXI comprised the same number of
T2-weighed coronal brain images from 10 k-spaces. BraTs and fastMRI have many simi-
larities, such as features and MR sequences. On the other hand, BraTs and IXI have dis-
tinctly dissimilar features and are not compatible due to their MR sequences. This exper-
iment was performed to assess how well the CNN performs and generalizes when the
training and testing datasets are different. The results of FDA-CNN and other methods
without fine-tuning the fastMRI dataset are shown in Table 2.
Table 2. Average SSIM, PSNR, NRMSE, and VIFP values and reconstruction time of state-of-art
methods on fastMRI-T1 axial brain data using several sampling patterns.
Sampling
Metrics
Zero Filling
TV
Wavelet
LAE
Unet
Wnet
PBCU
DRL-Net
FDA-CNN
2DG-US
SSIM 0.52 0.49 0.51 0.70
0.72 0.72 0.76 0.80 0.81
PSNR 35.87 35.65 35.89 35.16
35.28 35.13 35.51 35.40 36.23
NRMSE 0.07 0.07 0.07 0.08 0.08 0.08 0.08 0.07 0.06
VIFP 0.78 0.56 0.79 0.84 0.80
0.85 0.82 0.81 0.88
1DG-US
SSIM 0.57 0.54 0.57 0.61 0.64 0.63 0.65 0.66 0.71
PSNR 32.64 32.62 32.64 32.89 33.08 32.71 33.03 33.18 34.29
NRMSE 0.14 0.14 0.14 0.13 0.13 0.14 0.13 0.11 0.09
VIFP 0.43 0.29 0.42 0.64 0.62 0.69 0.64 0.56 0.64
MCP-US
SSIM 0.78 0.76 0.78 0.88 0.87 0.88 0.88 0.88 0.91
PSNR 36.76 36.40 36.76 37.79 37.93 37.88 37.76 37.81 40.88
NRMSE 0.06 0.06 0.06 0.05 0.04 0.04 0.05 0.04 0.02
VIFP 0.56 0.40 0.54 0.75 0.77 0.82 0.80 0.73 0.75
Reconstruction time (seconds) 1.52 1.3 0.46 0.49 0.95 0.82 0.47 0.44
With the 2DG-US mask, the CS methods slightly improved image quality, but under
the other two Cartesian samplings, these methods did not perform well. The CNNs re-
moved the artifacts and improved the image quality using all three sub-sampling patterns.
Instead of VIFP, our method produced better average PSNR, SSIM, and NRMSE than the
multidomain (Wnet) network in two cases. As shown in Figure 10, FDA-CNN effectively
removed most of the artifacts and generated a better image than Unet. The performances
of three under-sampling masks on the fastMRI dataset are shown in Figure 11. MCP-US
produced better PSNR and SSIM than other under-sampling patterns, except for VIFP us-
ing the 2D random sampling pattern.
Bioengineering 2023, 10, 22 12 of 18
Figure 10. Reconstructed MRI images from the fastMRI dataset (slice No. 01) using zero filling, Unet,
and FDA-CNN: (a) 2D Gaussian distribution; (b) 1D Gaussian distribution; (c) mixed center and
peripheral mask; (d) fully sampled ground truth image.
Figure 11. Comparison of three sampling patterns on fastMRI dataset using FDA-CNN.
The test results of FDA-CNN and other methods on the IXI dataset are shown in
Table 3. In this case, the CS methods improved some quantitative values but decreased
the VIFP values. FDA-CNN performed significantly better and yielded high-quality im-
ages by eliminating unwanted artifacts compared with other networks using all masks.
However, Wnet generated better VIFP in two Cartesian samplings than our method. The
goal of this experiment was to determine whether it is feasible to test CNNs on unknown
testing datasets to remove artifacts from anatomically accurate MR images using various
sampling patterns.
Bioengineering 2023, 10, 22 13 of 18
Table 3. Average SSIM, PSNR, NRMSE, and VIFP values and reconstruction time of state-of-the-art
methods on the IXI -T2 coronal dataset using several sampling patterns.
Sampling
Metrics
Zero Filling
TV
Wavelet
DRL-Net
FDA-CNN
2DG-US
SSIM 0.46 0.48 0.50 0.67 0.70 0.71 0.73 0.76 0.77
PSNR 33.97 33.46 33.98 34.25 34.85 34.71 35.02 34.75 35.47
NRMSE 0.15 0.07 0.07 0.10 0.08 0.08 0.08 0.07 0.06
VIFP 0.69 0.70 0.67 0.89 0.86 0.90 0.86 0.75 0.97
1DG-US
SSIM 0.50 0.52 0.56 0.64 0.66 0.66 0.68 0.71 0.79
PSNR 33.38 33.47 33.67 33.43 33.63 33.34 33.64 34.14 35.02
NRMSE 0.12 0.12 0.12 0.12 0.11 0.12 0.11 0.09 0.07
VIFP 0.35 0.19 0.32 0.58 0.57 0.60 0.55 0.56 0.51
MCP-US
SSIM 0.70 0.72 0.72 0.82 0.81 0.82 0.83 0.83 0.89
PSNR 35.74 35.40 35.40 36.36 36.58 36.44 36.52 36.65 37.70
NRMSE 0.07 0.08 0.08 0.06 0.06 0.06 0.06 0.05 0.04
VIFP 0.44 0.25 0.25 0.71 0.73 0.77 0.73 0.68 0.68
Reconstruction time (seconds) 0.9 0.91 0.33 0.33 0.43 0.33 0.32 0.32
As shown in Figure 12, FDA-CNN effectively reconstructed a better image, which
was close to the reference image from different MRI sequence data, than Unet. The per-
formances of three under-sampling patterns on the IXI dataset are shown in Figure 13.
MCP-US performed better than other under-sampling patterns, except for VIFP using
2DG-US.
Figure 12. Reconstructed MRI images from IXI dataset (slice No. 75) using zero filling, Unet, and
FDA-CNN at a sampling rate of 20%: (a) 2D Gaussian distribution; (b) 1D Gaussian distribution; (c)
mixed center and peripheral under-sampling; (d) fully sampled ground truth image.
Bioengineering 2023, 10, 22 14 of 18
Figure 13. Comparison of three sampling patterns on IXI dataset using FDA-CNN.
The average NRMSEs of the BraTs test dataset slices are displayed in Figure 14. The
NRMSEs exhibited a recognizable pattern over the middle slices. The borders of the brain con-
tain a lower number of frequencies that produce more unspecified and inconsistent images.
Figure 14. Average NRMSE variation across the slices. Edge slices have larger errors.
FDA-CNN performed better in artifact removal and image restoration than the reg-
ular Unet-based CNN and CS techniques according to the described test results. The two
3 × 3 convolutions in Unet are replaced by a dense block in FD-CNN. The input and output
of all of the convolutional layers are comparatively small, although the dense block has
eight distinct convolutional layers (four 1 × 1 and four 3 × 3). Therefore, the computational
cost of the dense block-based convolutional layer is less expensive than that of Unet. Ad-
ditionally, the regularizing impact of extensive connectivity lessens the possibility of over-
fitting the training data. The effectiveness of the CNN is heavily reliant on the accuracy of
the MRI spatial frequencies, which is a drawback of post-processing techniques such as
FDA-CNN. CNN reconstruction is likely to restore image features inaccurately if they are
heavily obscured. Some of the lower frequencies may be recoverable if the CNN is directly
employed to restore the sensor data. Furthermore, FD-CNN is more generalized than
other state-of-the-art methods, as it generated higher average SSIM and PSNR, and lower
average NRMSE on both the fastMRI and IXI datasets.
Bioengineering 2023, 10, 22 15 of 18
5. Conclusions
This article presents an efficient and effective deep learning-based method for MRI
reconstruction from a sparsely sampled k-space using a fully dense attention convolu-
tional neural network. In the proposed approach, edge information and geometry struc-
ture are restored more effectively from zero-filled MRI images. This network has the com-
petency to extract realistic features and reconstruct 2D images that are virtually similar to
the original. Dense connectivity remarkably promotes feature reuse and improves infor-
mation flow within the network. Furthermore, AGs combine lower and higher spatial in-
formation to pick up more useful features, so the model needs a smaller number of pa-
rameters than the more complex Unet. This makes the network more generalized. Alt-
hough network training requires many hours, reconstruction can be performed fast after
training. Compared with CS-based iterative existing approaches, the proposed network
needs less reconstruction time.
Compared with existing DL-based denoising and de-aliasing methods, the proposed
network shows outstanding performance with regard to quantitative and qualitative hu-
man vision indexes, and reconstruction time. Furthermore, the correlations between the
acquired image quality and several under-sampling patterns were evaluated. Future re-
search will focus on recovering unmeasured frequencies in the k-domain. Moreover, we
will implement our approach for real-time interactive temperature-based MRI.
Author Contributions: Conceptualization, M.B.H.; methodology, M.B.H.; software, M.B.H.; valida-
tion, M.B.H. and K.-C.K.; formal analysis, M.B.H., K.-C.K., and N.K.; investigation, M.B.H. and K.-
C.K.; resources, M.B.H. and S.M.I.; data curation, M.B.H.; writing—original draft preparation,
M.B.H.; writing—review and editing, M.B.H., K.-C.K., S.M.I., O.-S.N., S.-H.J., and N.K.; visualiza-
tion, M.B.H. and K.-C.K.; supervision, K.-C.K. and N.K.; project administration, K.-C.K. and N.K.;
funding acquisition, N.K. M.B.H. and K.-C.K. are co-first authors. All authors have read and agreed
to the published version of the manuscript.
Funding: This work was supported by Institute for Information and Communications Technology
Promotion (IITP), funded by the Korean government (MSIP) (No. 2021-0-00490; Development of
precision analysis and imaging technology for biological radio waves).
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The datasets are available at https://www.med.up-
enn.edu/cbica/brats2020/registration.html, https://fastmri.med.nyu.edu/ and https://brain-develop-
ment.org/ixi-dataset/. The source code of this manuscript is available at https://github.com/bid-
dut2j8/FDA-CNN (Last accessed on 20 December 2022).
Acknowledgments: We thank Shahinur Alam and Rupali Shinde for validating the proposed
method from a deep learning perspective.
Conflicts of Interest: The authors declare no conflict of interest.
Abbreviations
1D One-dimensional
1DG-US 1D Gaussian under-sampling
2D Two-dimensional
2DG-US 2D Gaussian under-sampling
AG Attention gate
CNN Convolutional neural network
CS Compressed sensing
CT Computed tomography
DB Dense block
DC Dense connectivity
DL Deep learning
DRL Deep residual learning
Bioengineering 2023, 10, 22 16 of 18
FDA-CNN
Fully dense attention CNN
FFT Fast Fourier transform
GAN Generative adversarial network
IFFT Inverse fast Fourier transform
LAE Lightweight autoencoder
MCP-US Mixed center and periphery under-sampling
MRI Magnetic resonance imaging
MSE Mean square error
NRMSE Normalized root mean squared error
PAT Photoacoustic tomography
PBCU Projection-based cascade Unet
PI Parallel imaging
PSNR Peak signal-to-noise ratio
ReLU Rectified linear unit
SSIM Structural similarity index measure
TV Total variation
VIFP Pixel visual information fidelity
References
1. Brown, R.W.; Cheng, Y.-C.N.; Haacke, E.M.; Thompson, M.R.; Venkatesan, R. Magnetic Resonance Imaging: Physical Principles and
Sequence Design, 2nd ed.; John Wiley & Sons Ltd: Hoboken, NJ, USA, 2014; ISBN 978-1-11863-395-3.
2. Cercignani, M.; Dowell, N.G.; Paul, S. Tofts Quantitative MRI of the Brain: Principles of Physical Measurement; CRC Press: Boca
Raton, FL, USA, 2018; Volume 15; ISBN 978-1-31536-357-8.
3. Muckley, M.J.; Riemenschneider, B.; Radmanesh, A.; Kim, S.; Jeong, G.; Ko, J.; Jun, Y.; Shin, H.; Hwang, D.; Mostapha, M.; et al.
Results of the 2020 fastMRI challenge for machine learning MR image reconstruction. IEEE Trans. Med. Imaging 2021, 40, 2306–2317.
4. Por, E.; van Kooten, M.; Sarkovic, V. Nyquist–Shannon Sampling Theorem; Leiden University: Leiden, The Netherlands, 2019.
5. Schoenberg, S.O.; Dietrich, O.; Reiser, M.F. Parallel Imaging in Clinical MR Applications; Medical Radiology; Springer:
Berlin/Heidelberg, Germany, 2007; ISBN 978-3-54023-102-8.
6. Deshmane, A.; Gulani, V.; Griswold, M.A.; Seiberlich, N. Parallel MR imaging. J. Magn. Reson. Imaging 2012, 36, 55–72.
https://doi.org/10.1002/jmri.23639.
7. Lustig, M.; Donoho, D.; Pauly, J.M. Sparse MRI: The application of compressed sensing for rapid MR imaging. Magn. Reson.
Med. 2007, 58, 1182–1195. https://doi.org/10.1002/mrm.21391.
8. Lustig, M.; Donoho, D. Compressed sensing MRI. IEEE Signal Process. Mag. 2008, 25, 72–82.
9. Lee, J.G.; Jun, S.; Cho, Y.W.; Lee, H.; Kim, G.B.; Seo, J.B.; Kim, N. Deep learning in medical imaging: General overview. Korean
J. Radiol. 2017, 18, 570–584. https://doi.org/10.3348/kjr.2017.18.4.570.
10. Wu, Y.; Alley, M.; Li, Z.; Datta, K.; Wen, Z.; Sandino, C.; Syed, A.; Ren, H.; Xing, L.; Lustig, M.; et al. Deep learning-based water-
fat separation from dual-echo chemical shift-encoded imaging. Bioengineering 2022, 9, 579.
https://doi.org/10.3390/bioengineering9100579.
11. Ahishakiye, E.; Van Gijzen, M.B.; Tumwiine, J.; Wario, R.; Obungoloch, J. A survey on deep learning in medical image
reconstruction. Intell. Med. 2021, 1, 118–127. https://doi.org/10.1016/j.imed.2021.03.003.
12. Amer, R.; Nassar, J.; Trabelsi, A.; Bendahan, D.; Greenspan, H.; Ben-Eliezer, N. Quantification of intra-muscular adipose infil-
tration in calf/thigh MRI using fully and weakly supervised semantic segmentation. Bioengineering 2022, 9, 315.
https://doi.org/10.3390/bioengineering9070315.
13. Lin, D.J.; Johnson, P.M.; Knoll, F.; Lui, Y.W. Artificial intelligence for MR image reconstruction: An Overview for Clinicians. J.
Magn. Reson. Imaging 2021, 53, 1015–1028.
14. Pal, A.; Rathi, Y. A review and experimental evaluation of deep learning methods for MRI reconstruction. Mach. Learn. Biomed.
Imaging 2022, 2022, 1–58. https://doi.org/10.48550/arXiv.2109.08618.
15. Yuan, Z.; Jiang, M.; Wang, Y.; Wei, B.; Li, Y.; Wang, P.; Menpes-Smith, W.; Niu, Z.; Yang, G. SARA-GAN: Self-attention and
relative average discriminator based generative adversarial networks for fast compressed sensing MRI reconstruction. Front.
Neuroinform. 2020, 14, 611666. https://doi.org/10.3389/fninf.2020.611666.
16. Wang, J.; Chen, Y.; Wu, Y.; Shi, J.; Gee, J. Enhanced generative adversarial network for 3d brain MRI super-resolution. In 2020
IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass, CO, USA, 1–5 March 2020; pp. 3616–3625.
https://doi.org/10.1109/WACV45572.2020.9093603.
17. Jiang, M.; Zhi, M.; Wei, L.; Yang, X.; Zhang, J.; Li, Y.; Wang, P.; Huang, J.; Yang, G. FA-GAN: Fused attentive generative
adversarial networks for MRI image super-resolution. Comput. Med. Imaging Graph. 2021, 92, 101969.
https://doi.org/10.1016/j.compmedimag.2021.101969.
Bioengineering 2023, 10, 22 17 of 18
18. Zhang, K.; Hu, H.; Philbrick, K.; Conte, G.M.; Sobek, J.D.; Rouzrokh, P.; Erickson, B.J. SOUP-GAN: Super-resolution MRI using
generative adversarial networks. Int. J. Res. Appl. Sci. Eng. Technol. 2021, 9, 3896–3905. https://doi.org/10.22214/ijraset.2021.37237.
19. Pineda, L.; Basu, S.; Romero, A.; Calandra, R.; Drozdzal, M. Active MR k-space sampling with reinforcement learning. Lect.
Notes Comput. Sci. 2020, 12262, 23–33. https://doi.org/10.1007/978-3-030-59713-9_3.
20. Han, Y.; Sunwoo, L.; Ye, J.C. K-Space deep learning for accelerated MRI. IEEE Trans. Med. Imaging 2020, 39, 377–386.
https://doi.org/10.1109/TMI.2019.2927101.
21. Du, T.; Zhang, H.; Li, Y.; Pickup, S.; Rosen, M.; Zhou, R.; Song, H.K.; Fan, Y. Adaptive convolutional neural networks for
accelerating magnetic resonance imaging via k-space data interpolation. Med. Image Anal. 2021, 72, 102098.
https://doi.org/10.1016/j.media.2021.102098.
22. Arefeen, Y.; Beker, O.; Cho, J.; Yu, H.; Adalsteinsson, E.; Bilgic, B. Scan-specific artifact reduction in k-space (SPARK) neural
networks synergize with physics-based reconstruction to accelerate MRI. Magn. Reson. Med. 2022, 87, 764–780.
https://doi.org/10.1002/mrm.29036.
23. Eo, T.; Jun, Y.; Kim, T.; Jang, J.; Lee, H.J.; Hwang, D. KIKI-Net: Cross-domain convolutional neural networks for reconstructing
undersampled magnetic resonance images. Magn. Reson. Med. 2018, 80, 2188–2201. https://doi.org/10.1002/mrm.27201.
24. Zou, J.; Li, C.; Jia, S.; Wu, R.; Pei, T.; Zheng, H.; Wang, S. SelfCoLearn: Self-supervised collaborative learning for accelerating
dynamic MR imaging. Bioengineering 2022, 9, 650. https://doi.org/10.3390/bioengineering9110650.
25. Souza, R.; Lebel, R.M.; Frayne, R. A hybrid, dual domain, cascade of convolutional neural networks for magnetic resonance
image reconstruction. Proc. Mach. Learn. Res. 2019, 102, 437–446.
26. Sun, L.; Wu, Y.; Shu, B.; Ding, X.; Cai, C.; Huang, Y.; Paisley, J. A Dual-domain deep lattice network for rapid MRI
reconstruction. Neurocomputing 2020, 397, 94–107. https://doi.org/10.1016/j.neucom.2020.01.063.
27. Souza, R.; Bento, M.; Nogovitsyn, N.; Chung, K.J.; Loos, W.; Lebel, R.M.; Frayne, R. Dual-domain cascade of u-nets for multi-
channel magnetic resonance image reconstruction. Magn. Reson. Imaging 2020, 71, 140–153.
https://doi.org/10.1016/j.mri.2020.06.002.
28. Wang, Z.; Jiang, H.; Du, H.; Xu, J.; Qiu, B. IKWI-Net: A cross-domain convolutional neural network for undersampled magnetic
resonance image reconstruction. Magn. Reson. Imaging 2020, 73, 1–10. https://doi.org/10.1016/j.mri.2020.06.015.
29. El-Rewaidy, H.; Fahmy, A.S.; Pashakhanloo, F.; Cai, X.; Kucukseymen, S.; Csecs, I.; Neisius, U.; Haji-Valizadeh, H.; Menze, B.;
Nezafat, R. Multi-domain convolutional neural network (MD-CNN) for radial reconstruction of dynamic cardiac MRI. Magn.
Reson. Med. 2021, 85, 1195–1208. https://doi.org/10.1002/mrm.28485.
30. Wang, Y.; Pang, Y.; Tong, C. DSMENet: Detail and structure mutually enhancing network for under-sampled MRI reconstruc-
tion. Comput. Biol. Med. 2022, 106204. https://doi.org/10.1016/j.compbiomed.2022.106204.
31. Qin, C.; Schlemper, J.; Duan, J.; Seegoolam, G.; Price, A.; Hajnal, J.; Rueckert, D. K-t NEXT: Dynamic MR image reconstruction
exploiting spatio-temporal correlations. Lect. Notes Comput. Sci. 2019, 11765, 505–513. https://doi.org/10.1007/978-3-030-32245-8_56.
32. Cheng, J.; Wang, H.; Ying, L.; Liang, D. Model learning: Primal dual networks for fast MR imaging. Lect. Notes Comput. Sci. 2019,
11766, 21–29. https://doi.org/10.1007/978-3-030-32248-9_3.
33. Bahadir, C.D.; Wang, A.Q.; Dalca, A.V.; Sabuncu, M.R. Deep-learning-based optimization of the under-sampling pattern in
MRI. IEEE Trans. Comput. Imaging 2020, 6, 1139–1152. https://doi.org/10.1109/TCI.2020.3006727.
34. Hosseini, S.A.H.; Yaman, B.; Moeller, S.; Hong, M.; Akcakaya, M. Dense recurrent neural networks for accelerated MRI: History-
cognizant unrolling of optimization algorithms. IEEE J. Sel. Top. Signal Process. 2020, 14, 1280–1291.
https://doi.org/10.1109/JSTSP.2020.3003170.
35. Kwon, K.; Kim, D.; Park, H. A parallel MR imaging method using multilayer perceptron. Med. Phys. 2017, 44, 6209–6224.
https://doi.org/10.1002/mp.12600.
36. Lee, D.; Yoo, J.; Tak, S.; Ye, J.C. Deep residual learning for accelerated MRI using magnitude and phase networks. IEEE Trans.
Biomed. Eng. 2018, 65, 1985–1995. https://doi.org/10.1109/TBME.2018.2821699.
37. Nath, R.; Callahan, S.; Singam, N.; Stoddard, M.; Amini, A.A. Accelerated phase contrast magnetic resonance imaging via deep
learning. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–
7 April 2020; pp. 834–838. https://doi.org/10.1109/ISBI45749.2020.9098508.
38. Yang, G.; Yu, S.; Dong, H.; Slabaugh, G.; Dragotti, P.L.; Ye, X.; Liu, F.; Arridge, S.; Keegan, J.; Guo, Y.; et al. DAGAN: Deep de-
aliasing generative adversarial networks for fast compressed sensing MRI reconstruction. IEEE Trans. Med. Imaging 2018, 37,
1310–1321. https://doi.org/10.1109/TMI.2017.2785879.
39. Quan, T.M.; Nguyen-Duc, T.; Jeong, W.K. Compressed sensing MRI reconstruction using a generative adversarial network with
a cyclic loss. IEEE Trans. Med. Imaging 2018, 37, 1488–1497. https://doi.org/10.1109/TMI.2018.2820120.
40. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image
denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. https://doi.org/10.1109/TIP.2017.2662206.
41. Ouchi, S.; Ito, S. Reconstruction of compressed-sensing MR imaging using deep residual learning in the image domain. Magn.
Reson. Med. Sci. 2021, 20, 190–203. https://doi.org/10.2463/mrms.mp.2019-0139.
42. Guan, S.; Khan, A.A.; Sikdar, S.; Chitnis, P.V. Fully dense unet for 2-D sparse photoacoustic tomography artifact removal. IEEE
J. Biomed. Health Inform. 2020, 24, 568–576. https://doi.org/10.1109/JBHI.2019.2912935.
Bioengineering 2023, 10, 22 18 of 18
43. Oktay, O.; Schlemper, J.; Le Folgoc, L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.;
et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999.
44. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the
2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–
2269. https://doi.org/10.1109/CVPR.2017.243.
45. Santurkar, S.; Tsipras, D.; Ilyas, A.; Madry, A. How does batch normalization help optimization? In Proceedings of the Advances
in Neural Information Processing Systems 31 (NeurIPS 2018), Montreal, QC, Canada, 3–8 December 2018; pp. 2483–2493.
46. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Lecture Notes in
Computer Science, Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, 18th International
Conference, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; Volume 9351, pp. 234–241; ISBN 978-3-31924-
573-7.
47. Agarap, A.F. Deep learning using rectified linear units (ReLU). arXiv 2018, arXiv:1803.08375.
48. Ottesen, J.A.; Caan, M.W.A.; Groote, I.R.; Bjørnerud, A. A densely interconnected network for deep learning accelerated MRI.
Magn. Reson. Mater. Phys. Biol. Med. 2022, 1–19. https://doi.org/10.1007/s10334-022-01041-3.
49. Schlemper, J.; Oktay, O.; Schaap, M.; Heinrich, M.; Kainz, B.; Glocker, B.; Rueckert, D. Attention gated networks: Learning to
leverage salient regions in medical images. Med. Image Anal. 2019, 53, 197–207. https://doi.org/10.1016/j.media.2019.01.012.
50. CBICA. Multimodal Brain Tumor Segmentation Challenge 2020: Registration/Data Request. Available online:
https://www.med.upenn.edu/cbica/brats2020/registration.html (accessed on 15 January 2022).
51. Knoll, F.; Zbontar, J.; Sriram, A.; Muckley, M.J.; Bruno, M.; Defazio, A.; Parente, M.; Geras, K.J.; Katsnelson, J.; Chandarana, H.;
et al. FastMRI: A publicly available raw k-space and dicom dataset of knee images for accelerated MR image reconstruction
using machine learning. Radiol. Artif. Intell. 2020, 2, e190007. https://doi.org/10.1148/ryai.2020190007.
52. Knoll et al. FastMRI Dataset. Available online: https://fastmri.med.nyu.edu/ (accessed on 10 June 2022).
53. IXI Dataset. Available online: https://brain-development.org/ixi-dataset/ (accessed on 20 July 2022).
54. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity.
IEEE Trans. Image Process. 2004, 13, 600–612. https://doi.org/10.1109/TIP.2003.819861.
55. Chow, L.S.; Rajagopal, H.; Paramesran, R. Correlation between subjective and objective assessment of magnetic resonance (MR)
images. Magn. Reson. Imaging 2016, 34, 820–831. https://doi.org/10.1016/j.mri.2016.03.006.
56. Bao, L.; Liu, W.; Zhu, Y.; Pu, Z.; Magnin, I.E. Sparse representation based MRI denoising with total variation. In Proceedings of
the 2008 9th International Conference on Signal Processing, Beijing, China, 26–29 October 2008; pp. 2154–2157.
https://doi.org/10.1109/ICOSP.2008.4697573.
57. Yang, X.; Fei, B. A wavelet multiscale denoising algorithm for magnetic resonance (MR) images. Meas. Sci. Technol. 2011, 22,
025803. https://doi.org/10.1088/0957-0233/22/2/025803.
58. Andrew, J.; Mhatesh, T.S.R.; Sebastin, R.D.; Sagayam, K.M.; Eunice, J.; Pomplun, M.; Dang, H. Super-resolution reconstruction
of brain magnetic resonance images via lightweight autoencoder. Inform. Med. Unlocked 2021, 26, 100713.
https://doi.org/10.1016/j.imu.2021.100713.
59. Aghabiglou, A.; Eksioglu, E.M. Projection-based cascaded u-net model for MR image reconstruction. Comput. Methods Programs
Biomed. 2021, 207, 106151. https://doi.org/10.1016/j.cmpb.2021.106151.
60. Souza, R.; Frayne, R. A hybrid frequency-domain/image-domain deep network for magnetic resonance image reconstruction.
In Proceedings of the 2019 32nd SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Rio de Janeiro, Brazil,
28–30 October 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 257–264.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual au-
thor(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.
... In recent years, deep learning has become extensively used in the healthcare field, for the diagnosis of COVID-19 infections, cancers, skin conditions, and Alzheimer's disease; remote monitoring; predicting cardiac issues; tracking blood sugar levels; and face mask detection [1][2][3][4][5]. The skin is the largest human organ; annual skin-related treatment costs exceed $75 billion in the USA [6]. ...
... Adegun et.al and Shinde et.al proposed CNN network for melanoma detection with encoder-decode and transfer learning approaches respectively as binary classification [2][3][4][5][6][7][8]. A Monkeypox diagnosis study based on a skin lesion dermatologist with 89% sensitivity was experimented by Stanford University [4]. ...
Article
Full-text available
Subjective visual examination by human dermatologists is associated with inter-observer variability and error. To address this problem, we present a method to accurately diagnose the 23 most common skin conditions, using an adaptive GrabCut approach with the EfficientNetB3 model, for accurate segmentation and classification, respectively. Using a custom loss function, this strategy is combined with data-level preprocessing employing algorithm-level approaches. The unbalanced Dermnet dataset, which includes 19,500 images of skin lesions representing the 23 most common skin conditions, was corrected by downsampling the major classes and enhancing the minor classes. The custom loss function and ADG significantly enhanced model accuracy by accurately segmenting regions of interest in the images, retaining the most relevant diagnostic information. The MSE, PSNR and Jacquard index support the best segmentation result with values 32.94, 70.23, 0.71 respectively. The results demonstrated very high accuracy, with top-5 and top-1 accuracy rates of 95% and 80.02%, respectively. The diagnoses of acne and nail fungi were exceptionally good, with precision rates exceeding 0.80 and 0.856, respectively. Our Dermnet-trained model is the most accurate of all state-of-the-art models published to date.
... The effectiveness of our proposed RA-CNN was compared with direct mapping a single-and multi-domain networks. Unet [36], the de-aliasing generative adversarial n work (DAGAN) [13], RefineGAN [14], the projection-based cascade Unet (PBCU) [5 and the fully dense attention (FDA)-CNN [52] are single-domain techniques, while DC-CNN [34], KIKI-net [29], W-net [35], the hybrid cascade [30], and the dual-encod Unet [53] are multi-domain approaches. The GAN-based DAGAN applies a residual U architecture for the generator and combined adversarial and innovative content loss RefineGAN measures cyclic loss using a residual Wasserstein GAN [54]. ...
... The effectiveness of our proposed RA-CNN was compared with direct mapping and single-and multi-domain networks. Unet [36], the de-aliasing generative adversarial network (DAGAN) [13], RefineGAN [14], the projection-based cascade Unet (PBCU) [51], and the fully dense attention (FDA)-CNN [52] are single-domain techniques, while the DC-CNN [34], KIKI-net [29], W-net [35], the hybrid cascade [30], and the dual-encoder Unet [53] are multi-domain approaches. The GAN-based DAGAN applies a residual Unet architecture for the generator and combined adversarial and innovative content losses. ...
Article
Full-text available
We propose a dual-domain deep learning technique for accelerating compressed sensing magnetic resonance image reconstruction. An advanced convolutional neural network with residual connectivity and an attention mechanism was developed for frequency and image domains. First, the sensor domain subnetwork estimates the unmeasured frequencies of k-space to reduce aliasing artifacts. Second, the image domain subnetwork performs a pixel-wise operation to remove blur and noisy artifacts. The skip connections efficiently concatenate the feature maps to alleviate the vanishing gradient problem. An attention gate in each decoder layer enhances network generalizability and speeds up image reconstruction by eliminating irrelevant activations. The proposed technique reconstructs real-valued clinical images from sparsely sampled k-spaces that are identical to the reference images. The performance of this novel approach was compared with state-of-the-art direct mapping, single-domain, and multi-domain methods. With acceleration factors (AFs) of 4 and 5, our method improved the mean peak signal-to-noise ratio (PSNR) to 8.67 and 9.23, respectively, compared with the single-domain Unet model; similarly, our approach increased the average PSNR to 3.72 and 4.61, respectively, compared with the multi-domain W-net. Remarkably, using an AF of 6, it enhanced the PSNR by 9.87 ± 1.55 and 6.60 ± 0.38 compared with Unet and W-net, respectively.
... There are native neural network implementations as well as neural networks formed by unfolding iterative algorithms. CNN based neural network is proposed in MD-Recon-Net [19], CSNET+ [20], CNN With Attention [21]. OPINE-NET [22], ADMM-CSNET [23], ISTA-NET+ [24] and AMP-NET [25] are some examples of neural network based unfolded iterative algorithms. ...
Article
Full-text available
Adaptive Block-Based Compressed Sensing (ABCS) enables optimization of image and video sensing platforms with limited resources, using novel algorithms for efficient reconstruction and real-time operations. Taking number of measurements adaptively based on information contained in each block of the image, results in better quality of recovered image, and is paving ways for general purpose use of compressed sensing in various applications. Some of the major challenges in ABCS are, complexity introduced at encoder end for adaptive rate allocation and choosing or learning a measurement matrix for each allocation rate. As part of this paper, we introduce a novel adaptive measurement allocation technique based on edge information with in the block. The algorithm is further improved by giving special considerations for edges present at the block boundary. DCT (Discrete Cosine Transform) dictionaries are used for measurement and recovery. Using simple encoding for measurement allocation greatly reduces the complexity of encoder network in this scheme. Recovery is done by simple IDCT (Inverse Discrete Cosine Transform) in case of DCT dictionaries. This approach demonstrates high effectiveness without the need for computationally expensive GPU-based training. To assess the network’s generalizability, we conducted tests using both natural and medical images. Remarkably, the method exhibited consistent accuracy across various measurement rates in both scenarios. The recovery time for compressively sensed images, whether they are natural or medical, is real-time, with an average duration of around 50 -65 milliseconds. The algorithm is also used in conjunction with Content Aware Scalable deep compressed sensing Network (CASNET) to get learned matrix for measurements on encoder side and pretrained model for reconstruction on decoder side. Proposed method not only converges in constant time across blocks in contrary to the rate allocation method of CASNET, but also outperforms the recovery quality in lower measurement rates. Extensive experimental results shows that proposed algorithm out performs other state of the art algorithms recovery.
... This results in reduced errors and more reliable diagnostic and treatment results. Fig. 3 illustrates the attention gate module, which enables the model to "ignore" less relevant features in the image [26]. This was done by assigning higher weights to the most important features and ignoring less relevant features. ...
Conference Paper
Diabetic Foot Ulcers (DFU) require effective management through early diagnosis and appropriate treatment. Image segmentation is very important in the management of DFU to identify the ulcer area. The popularity of U -Net architecture in the field of medical image segmentation still has several shortcomings in managing extensive contextual information and the existence of activation functions that are susceptible to saturation. The contribution of this research is to modify U-Net by applying fuzzy activation functions, attention gates, and hyperparameter optimization using genetic algorithms to improve DFU image segmentation performance. Testing on public datasets shows significant improvements in evaluation metrics such as dice coefficient, IoU, precision, recall, and 11 score. These results show that the research carried out can improve the diagnosis and treatment of diabetes patients through more accurate and stable medical image segmentation. Evaluation metrics show that after optimization with a genetic algorithm, the model achieves significant performance improvements compared to previous results. The test results also show that the model tested on the DFU2022 dataset has higher performance in terms of dice coefficient, IoU, precision and recall. Future research could consider using different data sets and additional approaches, as well as further evaluation to improve model performance.
... Magnetic resonance imaging (MRI) is used to diagnose many conditions, and there is no radiation exposure [1]. MRI image resolution and the signal-to-noise ratio have increased as the technology progressed, but longer scan times increase susceptibility to motion artifacts (MAs) [2]. ...
Article
Full-text available
We present a deep learning-based method that corrects motion artifacts and thus accelerates data acquisition and reconstruction of magnetic resonance images. The novel model, the Motion Artifact Correction by Swin Network (MACS-Net), uses a Swin transformer layer as the fundamental block and the Unet architecture as the neural network backbone. We employ a hierarchical transformer with shifted windows to extract multiscale contextual features during encoding. A new dual upsampling technique is employed to enhance the spatial resolutions of feature maps in the Swin transformer-based decoder layer. A raw magnetic resonance imaging dataset is used for network training and testing; the data contain various motion artifacts with ground truth images of the same subjects. The results were compared to six state-of-the-art MRI image motion correction methods using two types of motions. When motions were brief (within 5 s), the method reduced the average normalized root mean square error (NRMSE) from 45.25% to 17.51%, increased the mean structural similarity index measure (SSIM) from 79.43% to 91.72%, and increased the peak signal-to-noise ratio (PSNR) from 18.24 to 26.57 dB. Similarly, when motions were extended from 5 to 10 s, our approach decreased the average NRMSE from 60.30% to 21.04%, improved the mean SSIM from 33.86% to 90.33%, and increased the PSNR from 15.64 to 24.99 dB. The anatomical structures of the corrected images and the motion-free brain data were similar.
... Cartesian and non-Cartesian (random, Poisson-disc) are some common k-space sampling strategies [87]. A new sample pattern that combines the random and non-random frequencies of the phase direction was proposed by Hossain et al. [88]. Additionally, they developed an enhanced fully dense NN, which employs attention gates to eliminate redundant features. ...
Article
Full-text available
Citation: Hossain, M.B.; Shinde, R.K.; Oh, S.; Kwon, K.-C.; Kim, N. A Systematic Review and Identification of the Challenges of Deep Learning Techniques for Undersampled Magnetic Resonance Image Reconstruction. Sensors 2024, 24, 753. Abstract: Deep learning (DL) in magnetic resonance imaging (MRI) shows excellent performance in image reconstruction from undersampled k-space data. Artifact-free and high-quality MRI reconstruction is essential for ensuring accurate diagnosis, supporting clinical decision-making, enhancing patient safety, facilitating efficient workflows, and contributing to the validity of research studies and clinical trials. Recently, deep learning has demonstrated several advantages over conventional MRI reconstruction methods. Conventional methods rely on manual feature engineering to capture complex patterns and are usually computationally demanding due to their iterative nature. Conversely, DL methods use neural networks with hundreds of thousands of parameters and automatically learn relevant features and representations directly from the data. Nevertheless, there are some limitations to DL-based techniques concerning MRI reconstruction tasks, such as the need for large, labeled datasets, the possibility of overfitting, and the complexity of model training. Researchers are striving to develop DL models that are more efficient, adaptable, and capable of providing valuable information for medical practitioners. We provide a comprehensive overview of the current developments and clinical uses by focusing on state-of-the-art DL architectures and tools used in MRI reconstruction. This study has three objectives. Our main objective is to describe how various DL designs have changed over time and talk about cutting-edge tactics, including their advantages and disadvantages. Hence, data pre-and post-processing approaches are assessed using publicly available MRI datasets and source codes. Secondly, this work aims to provide an extensive overview of the ongoing research on transformers and deep convolutional neural networks for rapid MRI reconstruction. Thirdly, we discuss several network training strategies, like supervised, unsupervised, transfer learning, and federated learning for rapid and efficient MRI reconstruction. Consequently, this article provides significant resources for future improvement of MRI data pre-processing and fast image reconstruction.
... Using sparsely collected information to speed up MRI causes substantial artefacts that make it difficult to see the image's actual content. Fully dense attention CNN (FDA-CNN), a cutting-edge convolutional neural network (CNN), was suggested by Biddut et al. [16] to eliminate aliasing artefacts. Incorporating fully dense connection and an attention procedure for MRI restoration, researchers upgraded the Unet framework. ...
Article
Full-text available
Advancements in data capture techniques in the field of Magnetic Resonance Imaging (MRI) offer faster retrieval of critical medical imagery. Even with these advances, reconstruction techniques are generally slow and visually poor, making it difficult to include compression sensors. To address these issues, this work proposes a novel hybrid GAN-DRN architecture-based method for MRI reconstruction. This approach greatly improves texture, boundary characteristics, and picture fidelity over previous methods by combining Generative Adversarial Networks (GANs) with Deep Residual Networks (DRNs). One important innovation is the GAN's all-encompassing learning mechanism, which modifies the generator's behaviour to protect the network against corrupted input. In addition, the discriminator assesses forecast validity thoroughly at the same time. With this special technique, intrinsic features in the original photo are skillfully extracted and managed, producing excellent results that adhere to predetermined quality criteria. The Hybrid GAN-DRN technique's effectiveness is demonstrated by experimental findings, which use Python software to achieve an astounding 0.99 SSIM (Structural Similarities Index) and an amazing 50.3 peak signal-to-noise ratio. This achievement is a significant advancement in MRI reconstructions and has the potential to completely transform the medical imaging industry. In the future, efforts will be directed towards improving real-time MRI reconstruction, going multi-modal MRI fusion, confirming clinical effectiveness via trials, and investigating robustness, intuitive interfaces, transferable learning, and explanatory techniques to improve clinical interpretive practices and adoption.
... Cartesian and non-Cartesian (random, Poisson-disc) are some common k-space sampling strategies [87]. A new sample pattern that combines the random and non-random frequencies of the phase direction was proposed by Hossain et al. [88]. Additionally, they developed an enhanced fully dense NN, which employs attention gates to eliminate redundant features. ...
Article
Full-text available
Citation: Hossain, M.B.; Shinde, R.K.; Oh, S.; Kwon, K.-C.; Kim, N. A Systematic Review and Identification of the Challenges of Deep Learning Techniques for Undersampled Magnetic Resonance Image Reconstruction. Sensors 2024, 24, 753. Abstract: Deep learning (DL) in magnetic resonance imaging (MRI) shows excellent performance in image reconstruction from undersampled k-space data. Artifact-free and high-quality MRI reconstruction is essential for ensuring accurate diagnosis, supporting clinical decision-making, enhancing patient safety, facilitating efficient workflows, and contributing to the validity of research studies and clinical trials. Recently, deep learning has demonstrated several advantages over conventional MRI reconstruction methods. Conventional methods rely on manual feature engineering to capture complex patterns and are usually computationally demanding due to their iterative nature. Conversely, DL methods use neural networks with hundreds of thousands of parameters and automatically learn relevant features and representations directly from the data. Nevertheless, there are some limitations to DL-based techniques concerning MRI reconstruction tasks, such as the need for large, labeled datasets, the possibility of overfitting, and the complexity of model training. Researchers are striving to develop DL models that are more efficient, adaptable, and capable of providing valuable information for medical practitioners. We provide a comprehensive overview of the current developments and clinical uses by focusing on state-of-the-art DL architectures and tools used in MRI reconstruction. This study has three objectives. Our main objective is to describe how various DL designs have changed over time and talk about cutting-edge tactics, including their advantages and disadvantages. Hence, data pre-and post-processing approaches are assessed using publicly available MRI datasets and source codes. Secondly, this work aims to provide an extensive overview of the ongoing research on transformers and deep convolutional neural networks for rapid MRI reconstruction. Thirdly, we discuss several network training strategies, like supervised, unsupervised, transfer learning, and federated learning for rapid and efficient MRI reconstruction. Consequently, this article provides significant resources for future improvement of MRI data pre-processing and fast image reconstruction.
... Although attention modules can introduce computational demands, memory-efficient self-attention modules have been developed to address this limitation, making the integration of attention mechanisms more efficient in MRI reconstruction [88,89]. Overall, the attention module efficiently guides the network's attention to relevant image content, contributing to MRI reconstruction [90][91][92][93][94][95][96]. Table 6 summarizes various AM-based MRI reconstruction models and their key contributions and limitations. ...
Article
Full-text available
Magnetic Resonance Imaging (MRI) is an essential medical imaging modality that provides excellent soft-tissue contrast and high-resolution images of the human body, allowing us to understand detailed information on morphology, structural integrity, and physiologic processes. However, MRI exams usually require lengthy acquisition times. Methods such as parallel MRI and Compressive Sensing (CS) have significantly reduced the MRI acquisition time by acquiring less data through undersampling k-space. The state-of-the-art of fast MRI has recently been redefined by integrating Deep Learning (DL) models with these undersampled approaches. This Systematic Literature Review (SLR) comprehensively analyzes deep MRI reconstruction models, emphasizing the key elements of recently proposed methods and highlighting their strengths and weaknesses. This SLR involves searching and selecting relevant studies from various databases, including Web of Science and Scopus, followed by a rigorous screening and data extraction process using the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines. It focuses on various techniques, such as residual learning, image representation using encoders and decoders, data-consistency layers, unrolled networks, learned activations, attention modules, plug-and-play priors, diffusion models, and Bayesian methods. This SLR also discusses the use of loss functions and training with adversarial networks to enhance deep MRI reconstruction methods. Moreover, we explore various MRI reconstruction applications, including non-Cartesian reconstruction, super-resolution, dynamic MRI, joint learning of reconstruction with coil sensitivity and sampling, quantitative mapping, and MR fingerprinting. This paper also addresses research questions, provides insights for future directions, and emphasizes robust generalization and artifact handling. Therefore, this SLR serves as a valuable resource for advancing fast MRI, guiding research and development efforts of MRI reconstruction for better image quality and faster data acquisition.
... Retrospective experiments with 2D and 3D MRI data indicate that the PROJECTOR-generated trajectories can exploit the full possible range of gradients and slew rates well and produce sharp images. In another work, Hossain et al. [46] propose a new sampling pattern for 2D MRI, which combines the random and non-random sampling of the phase-encoding direction. The authors also introduce an advanced fully dense attention convolutional neural network (FDA-CNN), which reduces the number of redundant features using attention gates. ...
Article
Full-text available
Over the last decade, artificial intelligence (AI) has made an enormous impact on a wide range of fields, including science, engineering, informatics, finance, and transportation [...]
Article
Full-text available
Lately, deep learning technology has been extensively investigated for accelerating dynamic magnetic resonance (MR) imaging, with encouraging progresses achieved. However, without fully sampled reference data for training, the current approaches may have limited abilities in recovering fine details or structures. To address this challenge, this paper proposes a self-supervised collaborative learning framework (SelfCoLearn) for accurate dynamic MR image reconstruction from undersampled k-space data directly. The proposed SelfCoLearn is equipped with three important components, namely, dual-network collaborative learning, reunderampling data augmentation and a special-designed co-training loss. The framework is flexible and can be integrated into various model-based iterative un-rolled networks. The proposed method has been evaluated on an in vivo dataset and was compared to four state-of-the-art methods. The results show that the proposed method possesses strong capabilities in capturing essential and inherent representations for direct reconstructions from the undersampled k-space data and thus enables high-quality and fast dynamic MR imaging.
Article
Full-text available
Conventional water–fat separation approaches suffer long computational times and are prone to water/fat swaps. To solve these problems, we propose a deep learning-based dual-echo water–fat separation method. With IRB approval, raw data from 68 pediatric clinically indicated dual echo scans were analyzed, corresponding to 19382 contrast-enhanced images. A densely connected hierarchical convolutional network was constructed, in which dual-echo images and corresponding echo times were used as input and water/fat images obtained using the projected power method were regarded as references. Models were trained and tested using knee images with 8-fold cross validation and validated on out-of-distribution data from the ankle, foot, and arm. Using the proposed method, the average computational time for a volumetric dataset with ~400 slices was reduced from 10 min to under one minute. High fidelity was achieved (correlation coefficient of 0.9969, l1 error of 0.0381, SSIM of 0.9740, pSNR of 58.6876) and water/fat swaps were mitigated. I is of particular interest that metal artifacts were substantially reduced, even when the training set contained no images with metallic implants. Using the models trained with only contrast-enhanced images, water/fat images were predicted from non-contrast-enhanced images with high fidelity. The proposed water–fat separation method has been demonstrated to be fast, robust, and has the added capability to compensate for metal artifacts.
Article
Full-text available
Objective To improve accelerated MRI reconstruction through a densely connected cascading deep learning reconstruction framework. Materials and methods A cascading deep learning reconstruction framework (reference model) was modified by applying three architectural modifications: input-level dense connections between cascade inputs and outputs, an improved deep learning sub-network, and long-range skip-connections between subsequent deep learning networks. An ablation study was performed, where five model configurations were trained on the NYU fastMRI neuro dataset with an end-to-end scheme conjunct on four- and eightfold acceleration. The trained models were evaluated by comparing their respective structural similarity index measure (SSIM), normalized mean square error (NMSE), and peak signal to noise ratio (PSNR). Results The proposed densely interconnected residual cascading network (DIRCN), utilizing all three suggested modifications achieved a SSIM improvement of 8% and 11%, a NMSE improvement of 14% and 23%, and a PSNR improvement of 2% and 3% for four- and eightfold acceleration, respectively. In an ablation study, the individual architectural modifications all contributed to this improvement for both acceleration factors, by improving the SSIM, NMSE, and PSNR with approximately 2–4%, 4–9%, and 0.5–1%, respectively. Conclusion The proposed architectural modifications allow for simple adjustments on an already existing cascading framework to further improve the resulting reconstructions.
Article
Full-text available
Purpose: Infiltration of fat into lower limb muscles is one of the key markers for the severity of muscle pathologies. The level of fat infiltration varies in its severity across and within patients, and it is traditionally estimated using visual radiologic inspection. Precise quantification of the severity and spatial distribution of this pathological process requires accurate segmentation of lower limb anatomy into muscle and fat. Methods: Quantitative magnetic resonance imaging (qMRI) of the calf and thigh muscles is one of the most effective techniques for estimating pathological accumulation of intra-muscular adipose tissue (IMAT) in muscular dystrophies. In this work, we present a new deep learning (DL) network tool for automated and robust segmentation of lower limb anatomy that is based on the quantification of MRI’s transverse (T2) relaxation time. The network was used to segment calf and thigh anatomies into viable muscle areas and IMAT using a weakly supervised learning process. A new disease biomarker was calculated, reflecting the level of abnormal fat infiltration and disease state. A biomarker was then applied on two patient populations suffering from dysferlinopathy and Charcot–Marie–Tooth (CMT) diseases. Results: Comparison of manual vs. automated segmentation of muscle anatomy, viable muscle areas, and intermuscular adipose tissue (IMAT) produced high Dice similarity coefficients (DSCs) of 96.4%, 91.7%, and 93.3%, respectively. Linear regression between the biomarker value calculated based on the ground truth segmentation and based on automatic segmentation produced high correlation coefficients of 97.7% and 95.9% for the dysferlinopathy and CMT patients, respectively. Conclusions: Using a combination of qMRI and DL-based segmentation, we present a new quantitative biomarker of disease severity. This biomarker is automatically calculated and, most importantly, provides a spatially global indication for the state of the disease across the entire thigh or calf.
Article
Full-text available
Following the success of deep learning in a wide range of applications, neural network-based machine-learning techniques have received significant interest for accelerating magnetic resonance imaging (MRI) acquisition and reconstruction strategies. A number of ideas inspired by deep learning techniques for computer vision and image processing have been successfully applied to nonlinear image reconstruction in the spirit of compressed sensing for accelerated MRI. Given the rapidly growing nature of the field, it is imperative to consolidate and summarize the large number of deep learning methods that have been reported in the literature, to obtain a better understanding of the field in general. This article provides an overview of the recent developments in neural-network based approaches that have been proposed specifically for improving parallel imaging. A general background and introduction to parallel MRI is also given from a classical view of k-space based reconstruction methods. Image domain based techniques that introduce improved regularizers are covered along with k-space based methods which focus on better interpolation strategies using neural networks. While the field is rapidly evolving with plenty of papers published each year, in this review, we attempt to cover broad categories of methods that have shown good performance on publicly available data sets. Limitations and open problems are also discussed and recent efforts for producing open data sets and benchmarks for the community are examined.
Article
Full-text available
Purpose To develop a scan‐specific model that estimates and corrects k‐space errors made when reconstructing accelerated MRI data. Methods Scan‐specific artifact reduction in k‐space (SPARK) trains a convolutional‐neural‐network to estimate and correct k‐space errors made by an input reconstruction technique by back‐propagating from the mean‐squared‐error loss between an auto‐calibration signal (ACS) and the input technique’s reconstructed ACS. First, SPARK is applied to generalized autocalibrating partially parallel acquisitions (GRAPPA) and demonstrates improved robustness over other scan‐specific models, such as robust artificial‐neural‐networks for k‐space interpolation (RAKI) and residual‐RAKI. Subsequent experiments demonstrate that SPARK synergizes with residual‐RAKI to improve reconstruction performance. SPARK also improves reconstruction quality when applied to advanced acquisition and reconstruction techniques like 2D virtual coil (VC‐) GRAPPA, 2D LORAKS, 3D GRAPPA without an integrated ACS region, and 2D/3D wave‐encoded imaging. Results SPARK yields SSIM improvement and 1.5 – 2× root mean squared error (RMSE) reduction when applied to GRAPPA and improves robustness to ACS size for various acceleration rates in comparison to other scan‐specific techniques. When applied to advanced reconstruction techniques such as residual‐RAKI, 2D VC‐GRAPPA and LORAKS, SPARK achieves up to 20% RMSE improvement. SPARK with 3D GRAPPA also improves RMSE performance by ~2×, SSIM performance, and perceived image quality without a fully sampled ACS region. Finally, SPARK synergizes with non‐Cartesian, 2D and 3D wave‐encoding imaging by reducing RMSE between 20% and 25% and providing qualitative improvements. Conclusion SPARK synergizes with physics‐based acquisition and reconstruction techniques to improve accelerated MRI by training scan‐specific models to estimate and correct reconstruction errors in k‐space.
Article
Full-text available
Magnetic Resonance Imaging (MRI) is useful to provide detailed anatomical information such as images of tissues and organs within the body that are vital for quantitative image analysis. However, typically the MR images acquired lacks adequate resolution because of the constraints such as patients’ comfort and long sampling duration. Processing the low resolution MRI may lead to an incorrect diagnosis. Therefore, there is a need for super resolution techniques to obtain high resolution MRI images. Single image super resolution (SR) is one of the popular techniques to enhance image quality. Reconstruction based SR technique is a category of single image SR that can reconstruct the low resolution MRI images to high resolution images. Inspired by the advanced deep learning based SR techniques, in this paper we propose an autoencoder based MRI image super resolution technique that performs reconstruction of the high resolution MRI images from low resolution MRI images. Experimental results on synthetic and real brain MRI images show that our autoencoder based SR technique surpasses other state-of-the-art techniques in terms of peak signal-to-noise ratio (PSNR), structural similarity (SSIM), Information Fidelity Criterion (IFC), and computational time.
Article
Reconstructing zero-filled MR images (ZF) from partial k-space by convolutional neural networks (CNN) is an important way to accelerate MRI. However, due to the lack of attention to different components in ZF, it is challenging to learn the mapping from ZF to targets effectively. To ameliorate this issue, we propose a Detail and Structure Mutually Enhancing Network (DSMENet), which benefits from the complementary of the Structure Reconstruction UNet (SRUN) and the Detail Feature Refinement Module (DFRM). The SRUN learns structure-dominated information at multiple scales. And the DRFM enriches detail-dominated information from coarse to fine. The bidirectional alternate connections then exchange information between them. Moreover, the Detail Representation Construction Module (DRCM) extracts valuable initial detail representation for DFRM. And the Detail Guided Fusion Module (DGFM) facilitates the deep fusion of these complementary information. With the help of them, various components in ZF can be applied with discriminative attentions and mutually enhanced. In addition, the performance can be further improved by the Deep Enhanced Restoration (DER), a strategy based on recursion and constrain. Extensive experiments on fastMRI and CC-359 datasets demonstrate that DSMENet has robustness in terms of various body parts, under-sampling rates, and masks. Furthermore, DSMENet can achieve promising performance on qualitative and quantitative results, especially the competitive NMSE of 0.0268, PSNE of 33.7, and SSIM of 0.7808 on fastMRI 4 × single-coil knee leaderboard.
Article
2018 Curran Associates Inc.All rights reserved. Batch Normalization (BatchNorm) is a widely adopted technique that enables faster and more stable training of deep neural networks (DNNs). Despite its pervasiveness, the exact reasons for BatchNorm's effectiveness are still poorly understood. The popular belief is that this effectiveness stems from controlling the change of the layers' input distributions during training to reduce the so-called “internal covariate shift”. In this work, we demonstrate that such distributional stability of layer inputs has little to do with the success of BatchNorm. Instead, we uncover a more fundamental impact of BatchNorm on the training process: it makes the optimization landscape significantly smoother. This smoothness induces a more predictive and stable behavior of the gradients, allowing for faster training.