Next Article in Journal
Metallic Microneedles for Transdermal Drug Delivery: Applications, Fabrication Techniques and the Effect of Geometrical Characteristics
Next Article in Special Issue
Automated MRI Field of View Prescription from Region of Interest Prediction by Intra-Stack Attention Neural Network
Previous Article in Journal
Current Role of Intra-Articular Injections of Platelet-Rich Plasma in Adhesive Capsulitis of Shoulder: A Systematic Review
Previous Article in Special Issue
Wave-Encoded Model-Based Deep Learning for Highly Accelerated Imaging with Joint Reconstruction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

De-Aliasing and Accelerated Sparse Magnetic Resonance Image Reconstruction Using Fully Dense CNN with Attention Gates

1
School of Information and Communication Engineering, Chungbuk National University, Cheongju-si 28644, Chungcheongbuk-do, Republic of Korea
2
Department of Electronics Engineering, Incheon National University, 119 Academy-ro, Yeonsu-gu, Incheon 22012, Gyeonggi-do, Republic of Korea
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(1), 22; https://doi.org/10.3390/bioengineering10010022
Submission received: 25 November 2022 / Revised: 19 December 2022 / Accepted: 19 December 2022 / Published: 22 December 2022
(This article belongs to the Special Issue AI in MRI: Frontiers and Applications)

Abstract

:
When sparsely sampled data are used to accelerate magnetic resonance imaging (MRI), conventional reconstruction approaches produce significant artifacts that obscure the content of the image. To remove aliasing artifacts, we propose an advanced convolutional neural network (CNN) called fully dense attention CNN (FDA-CNN). We updated the Unet model with the fully dense connectivity and attention mechanism for MRI reconstruction. The main benefit of FDA-CNN is that an attention gate in each decoder layer increases the learning process by focusing on the relevant image features and provides a better generalization of the network by reducing irrelevant activations. Moreover, densely interconnected convolutional layers reuse the feature maps and prevent the vanishing gradient problem. Additionally, we also implement a new, proficient under-sampling pattern in the phase direction that takes low and high frequencies from the k-space both randomly and non-randomly. The performance of FDA-CNN was evaluated quantitatively and qualitatively with three different sub-sampling masks and datasets. Compared with five current deep learning-based and two compressed sensing MRI reconstruction techniques, the proposed method performed better as it reconstructed smoother and brighter images. Furthermore, FDA-CNN improved the mean PSNR by 2 dB, SSIM by 0.35, and VIFP by 0.37 compared with Unet for the acceleration factor of 5.

1. Introduction

Magnetic resonance imaging (MRI) is a sophisticated and noninvasive medical imaging technique that has full control over data capture to visualize the anatomy and functions of the human brain and body [1]. It plays a significant role in smart healthcare systems alongside medical research by providing high-quality reconstructed images without exposure to harmful radiation [2]. However, the long image acquisition time (more than 30 min per patient) [3] is a disadvantage compared with X-ray, computed tomography (CT), and photoacoustic tomography (PAT) imaging modalities. In MRI, frequencies are acquired in the k-space rather than directly in the image space. Full k-space data are required for high-quality reconstructed images, but this prolongs the acquisition process. This slow process causes the patient discomfort and generates motion artifacts due to patient movement. A common way to speed up MRI acquisition is by taking fewer scans and reconstructing the image using a partially recorded k-space. However, this creates blurring and aliasing artifacts in the images according to the Nyquist–Shannon principle [4]. Hence, an effective acceleration technique is essential for MRI reconstruction. Parallel imaging (PI) [5,6] applies coil sensitivity maps to speed up reconstruction, but the setup process is complex and expensive. Compressed sensing (CS) MRI [7,8] uses sparse data to reconstruct high-resolution images from randomly picked k-space data. However, the sparsity of data produces aliasing artifacts, and the computing time required is quite high for their iteration process.
Deep learning (DL) has been successfully adopted in medical image processing [9,10,11,12] and deals with the obstacles of iterative CS reconstruction [13,14]. Efficient reconstruction can be performed without repeatedly processing after properly training a DL model. Several post-processing strategies [15,16,17,18] based on generative adversarial networks (GANs) have been used to improve image quality in the image domain, whereas some approaches [19,20,21,22] estimate the unknown missing k-space frequencies in the sensor domain. Cross-domain techniques [23,24,25,26,27,28,29] operate in both the sensor and image domains. Recently, iterative unrolled optimization methods [30,31,32,33,34] have been used for sparse MRI reconstruction to improve the learning of image features. However, when the k-space is uniformly under-sampled, DL-based iterative methods cannot properly eliminate aliasing artifacts.
The first multilayer perceptron [35] is applied to reduce aliasing artifacts in parallel MRI. The multi-scale Unet structure [36] has been suggested to handle phase image reconstruction and scattered artifacts. The convolutional neural network (CNN)-based two-dimensional (2D) phase-contrast MRI reconstruction method [37] encodes low-frequency regions in the phase direction, while high-frequency regions store image edges. The GAN-based de-aliasing method [38] combines adversarial and innovative content loss but calculates fast Fourier transforms (FFTs) of magnitude images instead of raw MRI data. RefineGAN [39] measures only the cyclic loss of the training dataset using an autoencoder. Based on deep residual learning (DLR), the denoising technique [40] identifies and then subtracts noise from noisy images. After modifying this DLR-based network, Ouchi et al. [41] proposed a model to generate alias-free images by subtracting the predicted alias components from them.
Here, we adopt a post-processing strategy and propose an efficient fully dense attention convolutional neural network (FDA-CNN) architecture to remove aliasing and ghost artifacts from sparsely reconstructed 2D MRI images. The FDA-CNN method generates aliasing-free images instead of prediction artifacts, which is a key difference from previous MRI de-aliasing techniques. Guan et al. [42] proposed fully dense Unet (FD-Unet) for reducing artifacts from PAT images. We use the modified FD-Unet architecture for MRI reconstruction by adding an attention gate (AG) to each decoder layer. FDA-CNN consists of an AG [43] in each layer of the decoding path and dense connectivity (DC) [44] in both the encoding and decoding layers. DC solves the vanishing gradient issue and improves information flow, whereas the AG gives more attention to essential features that make the network more generalized. Supplementarily, we implement a hybrid mixed center and periphery under-sampling (MCP-US) mask in the phase [ reconstructs complex image textures and contents better than the two-dimensional (2D) and one-dimensional (1D) Gaussian under-sampling patterns. The results of FDA-CNN were evaluated using three sub-sampling masks using only 20% of the data. The simulation results were compared to both the single- and multidomain networks along with traditional CS methods on three widely used k-space datasets.

2. Methodology

The data obtained from an MRI scanner are known as the Fourier space or k-space and can be represented as follows:
y | = | FT
where T is the artifact-free image, F is the Fourier transform, and y represents the fully sampled k-space data. The reconstruction of image x ^ from the k-space can be performed by applying the inverse Fourier transform.
x ^ | = | F 1 y
The reconstruction of an image with n pixels requires at least n frequencies. However, it takes more scan time to capture n frequencies, because the number of frequencies acquired is linearly related to the length of the MRI scan. Acquiring only 1/5 of the frequencies can reduce both the scan time and cost by a factor of 5, but aliasing artifacts are introduced. Under-sampled k-space y ^ is represented as
y ^ | = | U     y | = | U     FT
Figure 1 exhibits a reconstructed artifact x aliased image from the under-sampled k-space. Here, U is a sub-sampling matrix filled with 0 s (black area) and 1 s (white line), and expresses the element-wise multiplication.
Conventional CS methods recover images by solving the following optimization problem:
  T min | y ^ y | 2 2 + λ G T
where | | . | | 2   is a generic data consistency term that assures the solution of the original image (T) with every observation of y and G denotes the regularization term, where λ expresses the regularization parameter. The challenge of traditional CS is that the regularizer, G(T), must be manually encoded to represent the reconstructed MR images.

2.1. Deep Learning Framework

A DL-based method is applied to adjust the under-sampled image and reconstruct an artifact-free image that is close to the actual image. This is called a supervised learning strategy, which aims to find an appropriate reconstruction function that matches the given image with the expected output. This can be expressed with the following equation:
R | = | { ( NN ) L ( F 1 ( y ^ ) ) }
where R is the reconstructed image, NN is a DL-based neural network, and L measures the loss between the original and reconstructed images.
As shown in Figure 2, the fully sampled k-space (y) is sub-sampled by means of element-wise multiplication with the under-sampling mask (U). First, an image with aliasing and ghost artifacts x ^ is reconstructed using the inverse fast Fourier transform (IFFT) from sub-sampled k-space y ^ . Then, these artifacts are removed by FDA-CNN by decreasing the loss (L) between the predicted image (R) and the target image (T). FDA-CNN attempts to recover an image F 1 ( y ^ ) that is close to the target image using sub-sampled sensor data y ^ as input. Gradient descent is used to optimize the parameters of the loss function.

2.2. Proposed FDA-CNN Architecture

The proposed MRI de-aliasing network architecture is presented in this section. We designed an improved CNN model based on the Unet [45] architecture by combining modified dense connectivity and AGs. We applied batch normalization [46] to accelerate training compared with earlier Unet implementations.
The network structure shown in Figure 3 has two main sections: the down- and up-sampling parts, and the skip connection part with the AGs. The down-sampling section consists of convolution, dense block, and max-pooling layers, whereas up-sampling consists of upscaling (2 × 2 deconvolution), dense block, and convolutional layers. Firstly, a 1-channel 3 × 3 convolutional operation is applied on a 256 × 256 input image with rectified linear unit (ReLU) [47] activation function. Then, five consecutive dense blocks (DB) are used. The first DB starts with 32 channels and gradually increases by 64, 128, 256, and 512. Every DB consists of a series of 1 × 1 and 3 × 3 convolutions with padding 1, batch normalization, and ReLU. Initially, hyperparameters k m and f m are specified by the user; for our method, we initialized k 1 = 8 and f 1 = 64. Then, k m and f m are changed by k m = 2 m 1 ×   k 1 and f m = 2 m 1 ×   f 1 , respectively. The concatenation of the inputs and outputs from every layer of the DB generates its final output. Except for the last DB of the encoding section, a max-pooling operation is executed after each DB, which halves the size of the input at each level and doubles the number of feature maps.
The decoding or up-sampling section restores the size of the feature maps and maintains a form symmetric to the encoding section. This symmetry enables the reuse of features by concatenating feature maps at the same level and reduces the loss of information caused by the encoding/decoding process. Before concatenation, the features of encoding and decoding layers go through the AG to focus on target features from different spatial information. Every layer of the decoding section executes a 1 × 1 convolution with padding 1 and ReLU before going into a DB. A 1-channel 1 × 1 convolution is executed before generating the final output.
Moreover, dense connectivity generates deeper networks. For comparison, Unet has 23 layers, while FDA-CNN has 97 convolutional and deconvolutional levels. The vanishing gradient problem arises because the gradient information must flow through different layers and may disappear before it arrives at the succeeding layers. Dense connectivity adds more links to enable the effective backpropagation of gradient information. This lessens the vanishing gradient issue and makes it easier to train the network.

2.3. Dense Block

Densely connected networks [48] maximize the capability of the network by reusing features. The input of the succeeding layers is more varied and more effective when feature maps from various layers are combined. In our method, a dense block with a growth rate, k m , is used to learn different feature maps, f m , for each spatial level, m. Initially, hyperparameters k m and f m are specified by the user. Then, k m and f m are changed by k m = 2 m 1 ×   k 1 and f m = 2 m 1 ×   f 1 , respectively, at each spatial level to preserve computational efficiency and ensure that each dense block has the same number of convolutional layers. A total of nine dense blocks with four layers are used in the FDA-CNN approach.
As shown in Figure 4, the L th layer of the dense block has an initial input with F + ×   ( L 1 ) feature maps and output with k m feature maps, where F is the total number of feature maps in the dense block’s initial input. Through a series of 1 × 1 and 3 × 3 convolutions with batch normalization plus ReLU activation function, features are learned. Due to the increased computational complexity of the 3 × 3 convolution, the input dimension is decreased to F feature maps by applying a 1 × 1 convolution, which increases convergence speed. Then, using a 3 × 3 convolution, k m attribute maps are developed from the compacted data. The concatenation of the inputs and outputs from every layer of the dense block generates the dense block’s final output.

2.4. Attention Gate

Models trained with AGs [49] intuitively learn to emphasize prominent features that are helpful for a particular task while suppressing irrelevant regions in an input image. With no additional computational work, AGs may be quickly added to common CNNs, such as Unet topologies, improving model sensitivity and prediction accuracy. Unet employs skip connections to merge spatial data from the up- and down-sampling paths. Low-quality feature representation exists in the first few layers, which carries in several redundant low-level feature extractions. By actively suppressing activations in unnecessary regions through the use of AGs at the skip connections, the number of redundant features transferred is decreased. Every AG takes two inputs, g and x. The gating signal, g, comes from the next lowest layer of the network. It has greater feature representation because it originates from a deeper region of the network. The input features, x, come from skipped connections. They have better spatial information because they originate from the early stages.
We incorporate an AG with every decoding part in our fully dense Unet framework. As shown in Figure 5, input features x i l perform 1 × 1 × 1 convolutions with stride 2 × 2 to lessen the size of the dimensions (H × W) by half, and gating signals g i l + 1 perform 1 × 1 × 1 convolutions with stride 1 × 1. As a result, the spatial geometry of the modified input features and gating signals is the same. The ReLU function activates them through element-wise summation and maps them by W int T into a smaller-dimensional space for gating operations. The sigmoid function levels the vector in (0, 1), with coefficients closer to 1 denoting more pertinent features. Then, a trilinear up-sampler is used to restore the size of attention weight matrix α i l to correspond to the pixel density of the input features. The output of the AG, x ^ i l , is generated by means of element-wise multiplication between attention weight matrix α i l and input features x i l and then is transmitted as usual through the residual connections.

3. Deep Learning Implementation

3.1. Datasets and Under-Sampling Masks

We used fully sampled brain k-space data from BraTs-2020 [50], fastMRI [51,52], and IXI [53]. We used the cross-sectional T1-weighted BraTS-2020 dataset for both the training and testing of all the networks. On the other hand, the T1-weighted axial fastMRI and T2-weighted coronal IXI datasets were only used for testing. Each volume possessed both the fully acquired k-space data and the associated reconstructed images of the same size (256 × 256). As we concentrated on the correlation between the number of k-space slices and FDA-CNN performance, no data augmentation was used in training. Images were reconstructed sequentially from every k-space.
During training, the sub-sampled zero-filled (ZF) noisy and artifact images were used as the network input along with the fully sampled images as target images. Our new MCP-US pattern was used for training and compared with 2D Gaussian under-sampling (2DG-US) and 1D Gaussian under-sampling (1DG-US) distributions. Mostly, 2D-US and 1DG-US focus on the central low frequencies of the k-space. However, low-spatial-frequency data, which determine the overall contrast, brightness, and form of the image, are located in the center of the k-space. On the other hand, high-spatial-frequency data determining the image edges and details are located in the periphery of the k-space. As the k-space has a symmetric nature, our MCP-US takes both the low and high frequencies. Among the total sampled data S of each k-space, MCP-US constantly samples 50% of the center ( s c ) and 25% of the periphery ( s p ) and randomly chooses 25% of data ( s r ) frequencies, except for center and periphery data. It can be expressed as follows:
S = s c + s p + s r
where   s c = S 2 % ,   s p = S 4 % ,   s r = S 4 %
In both training and testing, we took only 20% of data of each k-space for all three sub-sampling patterns, where white spaces were replaced with zero, as shown in Figure 6.
Among the sampled data S , our MCP-US non-randomly takes 10% of s c from the middle position and 5% of s p from the zero position and randomly chooses 5% of   s r between the s c and s p areas. As we concentrated on the correlation between the number of training images and FDA-CNN performance, no data augmentation was used in training.

3.2. Loss Function

The disparity between the sub-sampled aliasing image and the fully sampled aliasing-free image was evaluated using the loss function. The optimum objective of FDA-CNN is to minimize the value of the loss function. Smaller values between the under-sampled and fully sampled images ensure better reconstruction. We used the mean square error (MSE) as the loss function to calculate pixel-wise disparity and update the network parameters, which can be expressed as follows:
Loss MSE | = | 1 N i | = | 1 N ( T i R i ) 2
where N indicates the number of voxels (or pixels) in the image, and T i and R i represent the target and reconstructed MR images, respectively.

3.3. Performance Evaluations Metrics

To evaluate the network performance, we summarized the findings using four parameters: structural similarity index measure (SSIM) [54], peak signal-to-noise ratio (PSNR), normalized root mean squared error (NRMSE), and pixel visual information fidelity (VIFP) [55]. The SSIM is a perceptual index that utilizes the mutual dependencies among adjacent pixels to measure the similarity of two images, such as brightness, contrast, and structural properties. The following expression gives the SSIM between the network output (R) and the desired output (T):
SSIM   T ,   R | = | 2 μ T μ R + c 1 2 σ TR + c 1 μ T 2 + μ R 2 + c 1 σ T 2 + σ R 2 + c 1  
where μ T and μ R represent the mean values of T and R, respectively; and σ T 2 and σ R 2 denote the corresponding pixel variance values. The covariance value is also shown by σ TR . To stabilize the division, c 1 and c 2 have the following definitions:
c 1 | = | ( 0 . 01 P ) 2 ,   c 2 | = | ( 0 . 03 P ) 2
where P | = | max T min T
The PSNR calculates the ratio of the signal’s highest potential power (image intensity throughout a volume) to its fidelity-affecting distorting noise power. This can be expressed as
PSNR T ,   R | = | 10 log 10 255 1 N i | = | 1 N ( T i R i ) 2  
The ground truth and the pixel differences in network output images are compared by the NRMSE, which can be expressed as
NRMSE T ,   R | = | 1 N i | = | 1 N ( T i R i ) 2 max T min R
The human viewer’s perceptual evaluation approach, VIFP, measures image information by computing two mutual information quantities from the reference and distorted images. This can be defined as
VIFP T , R | = | j subbands I C N , j ;   R N , j | s N , j j subbands I C N , j ;   T N , j | s N , j
where R N , j , and T N , j   represent the sub-bands of the reconstructed and target images, respectively; S N , j defines a realization for a specific image; and C N , j   expresses N elements of random field C j that specifies the coefficient of the sub-band, j. The evaluation result of VIFP is indicated as values between 0 and 1, similar to the SSIM.
These criteria were chosen as they are typically used to evaluate image reconstruction. Higher values of SSIM, PSNR, and VIFP indicate better results, while smaller values of the NRMSE define better reconstructions. Moreover, the reconstruction time for each image indicates the transformation of MRI raw data into pictures. The reconstruction time of each method was calculated using the MCP-US pattern.

3.4. Experimental Setup

The training and testing of FDA-CNN were executed on an Intel Core i7-9800X 3.80 GHz processor and 128 GB memory with NVIDIA GeForce RTX 2080 Ti GPU running on Windows 10 Pro 64 bit. This model was implemented in Python 3.8 with the DL open-source libraries TensorFlow v2.4 and Keras v2.4 in the PyCharm environment. We used the Adam optimizer with momentum values of β1 = 0.9 and β2 = 0.999 to reduce the loss function. The starting learning rate was 1   ×   10 4 , and it declined with a decay factor of 0.95 every 20 epochs. A small batch size of 8 was chosen, and 2000 epochs were executed for the training of our network. Figure 7 illustrated the training and validation losses of the proposed network, where the regularizing impact of extensive connectivity lessens the possibility of overfitting the training data.

4. Result and Discussion

The performance of our network was compared to classical CS total variation (TV) [56], wavelet [57] denoising algorithms, and DL-based state-of-the-art (image and dual domain) reconstruction methods. Lightweight autoencoder (LAE) [58], basic Unet [46], projection-based cascade Unet (PBCU) [59], and DRL-based MRI (DRL-net) [41] reconstructions are image-domain DL networks. LAE uses an autoencoder framework, and PBCU uses five cascade Unets for MRI reconstruction. DRL-net subtracts the predicted artifacts from the under-sampled aliased images. The multidomain MRI reconstruction strategy (Wnet) [60] uses two Unets: one for the k-domain and another for the image domain. SSIM, PSNR, and NRMSE were used for quantitative analysis, where VIFP was used to evaluate the perception of the de-aliased images of a human viewer. The average reconstruction times were calculated using the MCP-US pattern. All results were generated in the same environment.

4.1. BraTs 2020-T1 Dataset

The T1-weighted axial brain Brats-2020 dataset was used for both the training and testing of all the networks. In an ideal dataset, the training and test data are very well correlated, providing an opportunity to acquire, from the training data, most of the features that are required to perform effectively during testing. The efficiency of CNNs can be compared in this ideal situation without being affected by the data. This BraTS-2020 dataset was obtained using a clinical 3T multimodal MR scanner. Among 150 k-spaces, 100 k-spaces were used for training; 30 k-spaces were used for validation; and 20 k-spaces were used for testing. Each k-space contained 155 axial cross-sectional T1-weighted (256 × 256) images. In the training, validation, and test sets, there was no duplication of the same k-space.
In this experiment, the same MRI sequence was used for both training and testing, with 25% of the training data being used for validation to increase the reliability of the results. The learning potential of FDA-CNN to eliminate artifacts was measured by adjusting the hyperparameters (feature maps and growth rate) of the dense block. Comparative DL methods were trained using the MCP-US pattern and evaluated on the associated datasets with each under-sampling mask. The efficiencies of all CNNs in terms of eliminating artifacts were compared using various sub-sampling masks for an acceleration factor of 5. In general, the CNN produced a better image with minimal artifacts. As seen in Table 1, the proposed FDA-CNN produced higher average SSIM, PSNR, and VIFP with a lower average NRMSE than traditional CS and the autoencoder and Unet-based methods.
The classical CS methods required almost 0.5 s and 0.97 s in reconstruction time for each slice, and the dual-domain network required approximately 0.4 s. In contrast, the single-domain post-processing methods required approximately 0.30 s to 0.33 s. Our proposed method generated better images than Unet within the same reconstruction time of 0.30 s for each slice.
As shown in Figure 8, the under-sampled image contained noise and aliasing artifacts. Basic Unet improved the image quality by reducing these artifacts, although some artifacts remained. Our method effectively removed most of the artifacts and reconstructed the images close to the reference images from the BraTs testing dataset. The performance of three under-sampling masks on BraTs testing data is shown in Figure 9, where MCP-US generated better PSNR and SSIM than other under-sampling patterns, except for VIFP using 2DG-US.

4.2. FastMRI and IXI Datasets

The second experiment used the fastMRI and IXI datasets to test the CNNs after they had been trained on the BraTs dataset. This represents a scenario using different training and test datasets that are not perfectly matched. For testing, fastMRI comprised 2560 T1-weighted axial brain images from 160 k-spaces, while IXI comprised the same number of T2-weighed coronal brain images from 10 k-spaces. BraTs and fastMRI have many similarities, such as features and MR sequences. On the other hand, BraTs and IXI have distinctly dissimilar features and are not compatible due to their MR sequences. This experiment was performed to assess how well the CNN performs and generalizes when the training and testing datasets are different. The results of FDA-CNN and other methods without fine-tuning the fastMRI dataset are shown in Table 2.
With the 2DG-US mask, the CS methods slightly improved image quality, but under the other two Cartesian samplings, these methods did not perform well. The CNNs removed the artifacts and improved the image quality using all three sub-sampling patterns. Instead of VIFP, our method produced better average PSNR, SSIM, and NRMSE than the multidomain (Wnet) network in two cases. As shown in Figure 10, FDA-CNN effectively removed most of the artifacts and generated a better image than Unet. The performances of three under-sampling masks on the fastMRI dataset are shown in Figure 11. MCP-US produced better PSNR and SSIM than other under-sampling patterns, except for VIFP using the 2D random sampling pattern.
The test results of FDA-CNN and other methods on the IXI dataset are shown in Table 3. In this case, the CS methods improved some quantitative values but decreased the VIFP values. FDA-CNN performed significantly better and yielded high-quality images by eliminating unwanted artifacts compared with other networks using all masks. However, Wnet generated better VIFP in two Cartesian samplings than our method. The goal of this experiment was to determine whether it is feasible to test CNNs on unknown testing datasets to remove artifacts from anatomically accurate MR images using various sampling patterns.
As shown in Figure 12, FDA-CNN effectively reconstructed a better image, which was close to the reference image from different MRI sequence data, than Unet. The performances of three under-sampling patterns on the IXI dataset are shown in Figure 13. MCP-US performed better than other under-sampling patterns, except for VIFP using 2DG-US.
The average NRMSEs of the BraTs test dataset slices are displayed in Figure 14. The NRMSEs exhibited a recognizable pattern over the middle slices. The borders of the brain contain a lower number of frequencies that produce more unspecified and inconsistent images.
FDA-CNN performed better in artifact removal and image restoration than the regular Unet-based CNN and CS techniques according to the described test results. The two 3 × 3 convolutions in Unet are replaced by a dense block in FD-CNN. The input and output of all of the convolutional layers are comparatively small, although the dense block has eight distinct convolutional layers (four 1 × 1 and four 3 × 3). Therefore, the computational cost of the dense block-based convolutional layer is less expensive than that of Unet. Additionally, the regularizing impact of extensive connectivity lessens the possibility of overfitting the training data. The effectiveness of the CNN is heavily reliant on the accuracy of the MRI spatial frequencies, which is a drawback of post-processing techniques such as FDA-CNN. CNN reconstruction is likely to restore image features inaccurately if they are heavily obscured. Some of the lower frequencies may be recoverable if the CNN is directly employed to restore the sensor data. Furthermore, FD-CNN is more generalized than other state-of-the-art methods, as it generated higher average SSIM and PSNR, and lower average NRMSE on both the fastMRI and IXI datasets.

5. Conclusions

This article presents an efficient and effective deep learning-based method for MRI reconstruction from a sparsely sampled k-space using a fully dense attention convolutional neural network. In the proposed approach, edge information and geometry structure are restored more effectively from zero-filled MRI images. This network has the competency to extract realistic features and reconstruct 2D images that are virtually similar to the original. Dense connectivity remarkably promotes feature reuse and improves information flow within the network. Furthermore, AGs combine lower and higher spatial information to pick up more useful features, so the model needs a smaller number of parameters than the more complex Unet. This makes the network more generalized. Although network training requires many hours, reconstruction can be performed fast after training. Compared with CS-based iterative existing approaches, the proposed network needs less reconstruction time.
Compared with existing DL-based denoising and de-aliasing methods, the proposed network shows outstanding performance with regard to quantitative and qualitative human vision indexes, and reconstruction time. Furthermore, the correlations between the acquired image quality and several under-sampling patterns were evaluated. Future research will focus on recovering unmeasured frequencies in the k-domain. Moreover, we will implement our approach for real-time interactive temperature-based MRI.

Author Contributions

Conceptualization, M.B.H.; methodology, M.B.H.; software, M.B.H.; validation, M.B.H. and K.-C.K.; formal analysis, M.B.H., K.-C.K. and N.K.; investigation, M.B.H. and K.-C.K.; resources, M.B.H. and S.M.I.; data curation, M.B.H.; writing—original draft preparation, M.B.H.; writing—review and editing, M.B.H., K.-C.K., S.M.I., O.-S.N., S.-H.J. and N.K.; visualization, M.B.H. and K.-C.K.; supervision, K.-C.K. and N.K.; project administration, K.-C.K. and N.K.; funding acquisition, N.K., M.B.H. and K.-C.K. are co-first authors. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Institute for Information and Communications Technology Promotion (IITP), funded by the Korean government (MSIP) (No. 2021-0-00490; Development of precision analysis and imaging technology for biological radio waves).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets are available at https://www.med.upenn.edu/cbica/brats2020/registration.html, https://fastmri.med.nyu.edu/ and https://brain-development.org/ixi-dataset/. The source code of this manuscript is available at https://github.com/biddut2j8/FDA-CNN (Last accessed on 20 December 2022).

Acknowledgments

We thank Shahinur Alam and Rupali Shinde for validating the proposed method from a deep learning perspective.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

1DOne-dimensional
1DG-US1D Gaussian under-sampling
2DTwo-dimensional
2DG-US 2D Gaussian under-sampling
AGAttention gate
CNNConvolutional neural network
CSCompressed sensing
CTComputed tomography
DBDense block
DCDense connectivity
DLDeep learning
DRLDeep residual learning
FDA-CNNFully dense attention CNN
FFTFast Fourier transform
GANGenerative adversarial network
IFFTInverse fast Fourier transform
LAELightweight autoencoder
MCP-USMixed center and periphery under-sampling
MRIMagnetic resonance imaging
MSEMean square error
NRMSENormalized root mean squared error
PATPhotoacoustic tomography
PBCUProjection-based cascade Unet
PIParallel imaging
PSNRPeak signal-to-noise ratio
ReLURectified linear unit
SSIMStructural similarity index measure
TVTotal variation
VIFPPixel visual information fidelity

References

  1. Brown, R.W.; Cheng, Y.-C.N.; Haacke, E.M.; Thompson, M.R.; Venkatesan, R. Magnetic Resonance Imaging: Physical Principles and Sequence Design, 2nd ed.; John Wiley & Sons Ltd: Hoboken, NJ, USA, 2014; ISBN 978-1-11863-395-3. [Google Scholar]
  2. Cercignani, M.; Dowell, N.G.; Paul, S. Tofts Quantitative MRI of the Brain: Principles of Physical Measurement; CRC Press: Boca Raton, FL, USA, 2018; Volume 15, ISBN 978-1-31536-357-8. [Google Scholar]
  3. Muckley, M.J.; Riemenschneider, B.; Radmanesh, A.; Kim, S.; Jeong, G.; Ko, J.; Jun, Y.; Shin, H.; Hwang, D.; Mostapha, M.; et al. Results of the 2020 fastMRI challenge for machine learning MR image reconstruction. IEEE Trans. Med. Imaging 2021, 40, 2306–2317. [Google Scholar] [CrossRef] [PubMed]
  4. Por, E.; van Kooten, M.; Sarkovic, V. Nyquist–Shannon Sampling Theorem; Leiden University: Leiden, The Netherlands, 2019. [Google Scholar]
  5. Schoenberg, S.O.; Dietrich, O.; Reiser, M.F. Parallel Imaging in Clinical MR Applications; Medical Radiology; Springer: Berlin/Heidelberg, Germany, 2007; ISBN 978-3-54023-102-8. [Google Scholar]
  6. Deshmane, A.; Gulani, V.; Griswold, M.A.; Seiberlich, N. Parallel MR imaging. J. Magn. Reson. Imaging 2012, 36, 55–72. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Lustig, M.; Donoho, D.; Pauly, J.M. Sparse MRI: The application of compressed sensing for rapid MR imaging. Magn. Reson. Med. 2007, 58, 1182–1195. [Google Scholar] [CrossRef] [PubMed]
  8. Lustig, M.; Donoho, D. Compressed sensing MRI. IEEE Signal Process. Mag. 2008, 25, 72–82. [Google Scholar] [CrossRef]
  9. Lee, J.G.; Jun, S.; Cho, Y.W.; Lee, H.; Kim, G.B.; Seo, J.B.; Kim, N. Deep learning in medical imaging: General overview. Korean J. Radiol. 2017, 18, 570–584. [Google Scholar] [CrossRef] [Green Version]
  10. Wu, Y.; Alley, M.; Li, Z.; Datta, K.; Wen, Z.; Sandino, C.; Syed, A.; Ren, H.; Xing, L.; Lustig, M.; et al. Deep learning-based water-fat separation from dual-echo chemical shift-encoded imaging. Bioengineering 2022, 9, 579. [Google Scholar] [CrossRef]
  11. Ahishakiye, E.; Van Gijzen, M.B.; Tumwiine, J.; Wario, R.; Obungoloch, J. A survey on deep learning in medical image reconstruction. Intell. Med. 2021, 1, 118–127. [Google Scholar] [CrossRef]
  12. Amer, R.; Nassar, J.; Trabelsi, A.; Bendahan, D.; Greenspan, H.; Ben-Eliezer, N. Quantification of intra-muscular adipose infiltration in calf/thigh MRI using fully and weakly supervised semantic segmentation. Bioengineering 2022, 9, 315. [Google Scholar] [CrossRef]
  13. Lin, D.J.; Johnson, P.M.; Knoll, F.; Lui, Y.W. Artificial intelligence for MR image reconstruction: An Overview for Clinicians. J. Magn. Reson. Imaging 2021, 53, 1015–1028. [Google Scholar] [CrossRef]
  14. Pal, A.; Rathi, Y. A review and experimental evaluation of deep learning methods for MRI reconstruction. Mach. Learn. Biomed. Imaging 2022, 2022, 1–58. [Google Scholar] [CrossRef]
  15. Yuan, Z.; Jiang, M.; Wang, Y.; Wei, B.; Li, Y.; Wang, P.; Menpes-Smith, W.; Niu, Z.; Yang, G. SARA-GAN: Self-attention and relative average discriminator based generative adversarial networks for fast compressed sensing MRI reconstruction. Front. Neuroinform. 2020, 14, 611666. [Google Scholar] [CrossRef]
  16. Wang, J.; Chen, Y.; Wu, Y.; Shi, J.; Gee, J. Enhanced generative adversarial network for 3d brain MRI super-resolution. In Proceedings of the 2020 IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass, CO, USA, 1–5 March 2020; pp. 3616–3625. [Google Scholar] [CrossRef]
  17. Jiang, M.; Zhi, M.; Wei, L.; Yang, X.; Zhang, J.; Li, Y.; Wang, P.; Huang, J.; Yang, G. FA-GAN: Fused attentive generative adversarial networks for MRI image super-resolution. Comput. Med. Imaging Graph. 2021, 92, 101969. [Google Scholar] [CrossRef]
  18. Zhang, K.; Hu, H.; Philbrick, K.; Conte, G.M.; Sobek, J.D.; Rouzrokh, P.; Erickson, B.J. SOUP-GAN: Super-resolution MRI using generative adversarial networks. Int. J. Res. Appl. Sci. Eng. Technol. 2021, 9, 3896–3905. [Google Scholar] [CrossRef]
  19. Pineda, L.; Basu, S.; Romero, A.; Calandra, R.; Drozdzal, M. Active MR k-space sampling with reinforcement learning. Lect. Notes Comput. Sci. 2020, 12262, 23–33. [Google Scholar] [CrossRef]
  20. Han, Y.; Sunwoo, L.; Ye, J.C. K-Space deep learning for accelerated MRI. IEEE Trans. Med. Imaging 2020, 39, 377–386. [Google Scholar] [CrossRef] [Green Version]
  21. Du, T.; Zhang, H.; Li, Y.; Pickup, S.; Rosen, M.; Zhou, R.; Song, H.K.; Fan, Y. Adaptive convolutional neural networks for accelerating magnetic resonance imaging via k-space data interpolation. Med. Image Anal. 2021, 72, 102098. [Google Scholar] [CrossRef]
  22. Arefeen, Y.; Beker, O.; Cho, J.; Yu, H.; Adalsteinsson, E.; Bilgic, B. Scan-specific artifact reduction in k-space (SPARK) neural networks synergize with physics-based reconstruction to accelerate MRI. Magn. Reson. Med. 2022, 87, 764–780. [Google Scholar] [CrossRef]
  23. Eo, T.; Jun, Y.; Kim, T.; Jang, J.; Lee, H.J.; Hwang, D. KIKI-Net: Cross-domain convolutional neural networks for reconstructing undersampled magnetic resonance images. Magn. Reson. Med. 2018, 80, 2188–2201. [Google Scholar] [CrossRef]
  24. Zou, J.; Li, C.; Jia, S.; Wu, R.; Pei, T.; Zheng, H.; Wang, S. SelfCoLearn: Self-supervised collaborative learning for accelerating dynamic MR imaging. Bioengineering 2022, 9, 650. [Google Scholar] [CrossRef]
  25. Souza, R.; Lebel, R.M.; Frayne, R. A hybrid, dual domain, cascade of convolutional neural networks for magnetic resonance image reconstruction. Proc. Mach. Learn. Res. 2019, 102, 437–446. [Google Scholar]
  26. Sun, L.; Wu, Y.; Shu, B.; Ding, X.; Cai, C.; Huang, Y.; Paisley, J. A Dual-domain deep lattice network for rapid MRI reconstruction. Neurocomputing 2020, 397, 94–107. [Google Scholar] [CrossRef]
  27. Souza, R.; Bento, M.; Nogovitsyn, N.; Chung, K.J.; Loos, W.; Lebel, R.M.; Frayne, R. Dual-domain cascade of u-nets for multi-channel magnetic resonance image reconstruction. Magn. Reson. Imaging 2020, 71, 140–153. [Google Scholar] [CrossRef]
  28. Wang, Z.; Jiang, H.; Du, H.; Xu, J.; Qiu, B. IKWI-Net: A cross-domain convolutional neural network for undersampled magnetic resonance image reconstruction. Magn. Reson. Imaging 2020, 73, 1–10. [Google Scholar] [CrossRef] [PubMed]
  29. El-Rewaidy, H.; Fahmy, A.S.; Pashakhanloo, F.; Cai, X.; Kucukseymen, S.; Csecs, I.; Neisius, U.; Haji-Valizadeh, H.; Menze, B.; Nezafat, R. Multi-domain convolutional neural network (MD-CNN) for radial reconstruction of dynamic cardiac MRI. Magn. Reson. Med. 2021, 85, 1195–1208. [Google Scholar] [CrossRef] [PubMed]
  30. Wang, Y.; Pang, Y.; Tong, C. DSMENet: Detail and structure mutually enhancing network for under-sampled MRI reconstruction. Comput. Biol. Med. 2022, 106204. [Google Scholar] [CrossRef]
  31. Qin, C.; Schlemper, J.; Duan, J.; Seegoolam, G.; Price, A.; Hajnal, J.; Rueckert, D. K-t NEXT: Dynamic MR image reconstruction exploiting spatio-temporal correlations. Lect. Notes Comput. Sci. 2019, 11765, 505–513. [Google Scholar] [CrossRef] [Green Version]
  32. Cheng, J.; Wang, H.; Ying, L.; Liang, D. Model learning: Primal dual networks for fast MR imaging. Lect. Notes Comput. Sci. 2019, 11766, 21–29. [Google Scholar] [CrossRef] [Green Version]
  33. Bahadir, C.D.; Wang, A.Q.; Dalca, A.V.; Sabuncu, M.R. Deep-learning-based optimization of the under-sampling pattern in MRI. IEEE Trans. Comput. Imaging 2020, 6, 1139–1152. [Google Scholar] [CrossRef]
  34. Hosseini, S.A.H.; Yaman, B.; Moeller, S.; Hong, M.; Akcakaya, M. Dense recurrent neural networks for accelerated MRI: History-cognizant unrolling of optimization algorithms. IEEE J. Sel. Top. Signal Process. 2020, 14, 1280–1291. [Google Scholar] [CrossRef]
  35. Kwon, K.; Kim, D.; Park, H. A parallel MR imaging method using multilayer perceptron. Med. Phys. 2017, 44, 6209–6224. [Google Scholar] [CrossRef]
  36. Lee, D.; Yoo, J.; Tak, S.; Ye, J.C. Deep residual learning for accelerated MRI using magnitude and phase networks. IEEE Trans. Biomed. Eng. 2018, 65, 1985–1995. [Google Scholar] [CrossRef] [Green Version]
  37. Nath, R.; Callahan, S.; Singam, N.; Stoddard, M.; Amini, A.A. Accelerated phase contrast magnetic resonance imaging via deep learning. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; pp. 834–838. [Google Scholar] [CrossRef]
  38. Yang, G.; Yu, S.; Dong, H.; Slabaugh, G.; Dragotti, P.L.; Ye, X.; Liu, F.; Arridge, S.; Keegan, J.; Guo, Y.; et al. DAGAN: Deep de-aliasing generative adversarial networks for fast compressed sensing MRI reconstruction. IEEE Trans. Med. Imaging 2018, 37, 1310–1321. [Google Scholar] [CrossRef] [Green Version]
  39. Quan, T.M.; Nguyen-Duc, T.; Jeong, W.K. Compressed sensing MRI reconstruction using a generative adversarial network with a cyclic loss. IEEE Trans. Med. Imaging 2018, 37, 1488–1497. [Google Scholar] [CrossRef] [Green Version]
  40. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [Green Version]
  41. Ouchi, S.; Ito, S. Reconstruction of compressed-sensing MR imaging using deep residual learning in the image domain. Magn. Reson. Med. Sci. 2021, 20, 190–203. [Google Scholar] [CrossRef]
  42. Guan, S.; Khan, A.A.; Sikdar, S.; Chitnis, P.V. Fully dense unet for 2-D sparse photoacoustic tomography artifact removal. IEEE J. Biomed. Health Inform. 2020, 24, 568–576. [Google Scholar] [CrossRef] [Green Version]
  43. Oktay, O.; Schlemper, J.; Le Folgoc, L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  44. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef] [Green Version]
  45. Santurkar, S.; Tsipras, D.; Ilyas, A.; Madry, A. How does batch normalization help optimization? In Proceedings of the Advances in Neural Information Processing Systems 31 (NeurIPS 2018), Montreal, QC, Canada, 3–8 December 2018; pp. 2483–2493. [Google Scholar]
  46. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Lecture Notes in Computer Science, Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. ISBN 978-3-31924-573-7. [Google Scholar]
  47. Agarap, A.F. Deep learning using rectified linear units (ReLU). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  48. Ottesen, J.A.; Caan, M.W.A.; Groote, I.R.; Bjørnerud, A. A densely interconnected network for deep learning accelerated MRI. Magn. Reson. Mater. Phys. Biol. Med. 2022, 1–19. [Google Scholar] [CrossRef]
  49. Schlemper, J.; Oktay, O.; Schaap, M.; Heinrich, M.; Kainz, B.; Glocker, B.; Rueckert, D. Attention gated networks: Learning to leverage salient regions in medical images. Med. Image Anal. 2019, 53, 197–207. [Google Scholar] [CrossRef]
  50. CBICA. Multimodal Brain Tumor Segmentation Challenge 2020: Registration/Data Request. Available online: https://www.med.upenn.edu/cbica/brats2020/registration.html (accessed on 15 January 2022).
  51. Knoll, F.; Zbontar, J.; Sriram, A.; Muckley, M.J.; Bruno, M.; Defazio, A.; Parente, M.; Geras, K.J.; Katsnelson, J.; Chandarana, H.; et al. FastMRI: A publicly available raw k-space and dicom dataset of knee images for accelerated MR image reconstruction using machine learning. Radiol. Artif. Intell. 2020, 2, e190007. [Google Scholar] [CrossRef] [PubMed]
  52. Knoll, F.; Sodickson, D.K.; Recht, M.P.; Lui, Y.W. FastMRI Dataset. Available online: https://fastmri.med.nyu.edu/ (accessed on 10 June 2022).
  53. IXI Dataset. Available online: https://brain-development.org/ixi-dataset/ (accessed on 20 July 2022).
  54. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Chow, L.S.; Rajagopal, H.; Paramesran, R. Correlation between subjective and objective assessment of magnetic resonance (MR) images. Magn. Reson. Imaging 2016, 34, 820–831. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Bao, L.; Liu, W.; Zhu, Y.; Pu, Z.; Magnin, I.E. Sparse representation based MRI denoising with total variation. In Proceedings of the 2008 9th International Conference on Signal Processing, Beijing, China, 26–29 October 2008; pp. 2154–2157. [Google Scholar] [CrossRef]
  57. Yang, X.; Fei, B. A wavelet multiscale denoising algorithm for magnetic resonance (MR) images. Meas. Sci. Technol. 2011, 22, 25803. [Google Scholar] [CrossRef] [Green Version]
  58. Andrew, J.; Mhatesh, T.S.R.; Sebastin, R.D.; Sagayam, K.M.; Eunice, J.; Pomplun, M.; Dang, H. Super-resolution reconstruction of brain magnetic resonance images via lightweight autoencoder. Inform. Med. Unlocked 2021, 26, 100713. [Google Scholar] [CrossRef]
  59. Aghabiglou, A.; Eksioglu, E.M. Projection-based cascaded u-net model for MR image reconstruction. Comput. Methods Programs Biomed. 2021, 207, 106151. [Google Scholar] [CrossRef]
  60. Souza, R.; Frayne, R. A hybrid frequency-domain/image-domain deep network for magnetic resonance image reconstruction. In Proceedings of the 2019 32nd SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Rio de Janeiro, Brazil, 28–30 October 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 257–264. [Google Scholar]
Figure 1. Reconstructed artifact image from sparsely sampled k-space.
Figure 1. Reconstructed artifact image from sparsely sampled k-space.
Bioengineering 10 00022 g001
Figure 2. Flowchart of the proposed de-aliasing technique.
Figure 2. Flowchart of the proposed de-aliasing technique.
Bioengineering 10 00022 g002
Figure 3. Proposed FDA-CNN architecture.
Figure 3. Proposed FDA-CNN architecture.
Bioengineering 10 00022 g003
Figure 4. Last dense block with four layers of encoding part, where k 5 = 128 and f 5 = 512.
Figure 4. Last dense block with four layers of encoding part, where k 5 = 128 and f 5 = 512.
Bioengineering 10 00022 g004
Figure 5. Schematic diagram of an attention gate.
Figure 5. Schematic diagram of an attention gate.
Bioengineering 10 00022 g005
Figure 6. Several sampling patterns: (a) fully sampled; (b) 2D Gaussian pattern; (c) 1D Gaussian pattern; (d) mixed center and periphery under-sampling pattern.
Figure 6. Several sampling patterns: (a) fully sampled; (b) 2D Gaussian pattern; (c) 1D Gaussian pattern; (d) mixed center and periphery under-sampling pattern.
Bioengineering 10 00022 g006
Figure 7. Training and validation losses of proposed FDA-CNN.
Figure 7. Training and validation losses of proposed FDA-CNN.
Bioengineering 10 00022 g007
Figure 8. Reconstructed MRI images from the BraTs testing dataset (slice No. 140) using zero filling, Unet, and FDA-CNN: (a) 2D Gaussian distribution; (b) 1D Gaussian distribution; (c) mixed center and peripheral mask; (d) fully sampled ground truth image.
Figure 8. Reconstructed MRI images from the BraTs testing dataset (slice No. 140) using zero filling, Unet, and FDA-CNN: (a) 2D Gaussian distribution; (b) 1D Gaussian distribution; (c) mixed center and peripheral mask; (d) fully sampled ground truth image.
Bioengineering 10 00022 g008
Figure 9. Comparison of three sampling patterns using FDA-CNN.
Figure 9. Comparison of three sampling patterns using FDA-CNN.
Bioengineering 10 00022 g009
Figure 10. Reconstructed MRI images from the fastMRI dataset (slice No. 01) using zero filling, Unet, and FDA-CNN: (a) 2D Gaussian distribution; (b) 1D Gaussian distribution; (c) mixed center and peripheral mask; (d) fully sampled ground truth image.
Figure 10. Reconstructed MRI images from the fastMRI dataset (slice No. 01) using zero filling, Unet, and FDA-CNN: (a) 2D Gaussian distribution; (b) 1D Gaussian distribution; (c) mixed center and peripheral mask; (d) fully sampled ground truth image.
Bioengineering 10 00022 g010
Figure 11. Comparison of three sampling patterns on fastMRI dataset using FDA-CNN.
Figure 11. Comparison of three sampling patterns on fastMRI dataset using FDA-CNN.
Bioengineering 10 00022 g011
Figure 12. Reconstructed MRI images from IXI dataset (slice No. 75) using zero filling, Unet, and FDA-CNN at a sampling rate of 20%: (a) 2D Gaussian distribution; (b) 1D Gaussian distribution; (c) mixed center and peripheral under-sampling; (d) fully sampled ground truth image.
Figure 12. Reconstructed MRI images from IXI dataset (slice No. 75) using zero filling, Unet, and FDA-CNN at a sampling rate of 20%: (a) 2D Gaussian distribution; (b) 1D Gaussian distribution; (c) mixed center and peripheral under-sampling; (d) fully sampled ground truth image.
Bioengineering 10 00022 g012
Figure 13. Comparison of three sampling patterns on IXI dataset using FDA-CNN.
Figure 13. Comparison of three sampling patterns on IXI dataset using FDA-CNN.
Bioengineering 10 00022 g013
Figure 14. Average NRMSE variation across the slices. Edge slices have larger errors.
Figure 14. Average NRMSE variation across the slices. Edge slices have larger errors.
Bioengineering 10 00022 g014
Table 1. Average SSIM, PSNR, NRMSE, and VIFP values and reconstruction time of state-of-art methods on BraTs-T1 axial brain data using several sampling patterns.
Table 1. Average SSIM, PSNR, NRMSE, and VIFP values and reconstruction time of state-of-art methods on BraTs-T1 axial brain data using several sampling patterns.
SamplingMetricsZero FillingTVWaveletLAEUnetWnetPBCUDRL-NetFDA-CNN
2DG-USSSIM0.310.320.360.780.790.770.820.830.89
PSNR34.1135.9036.1335.1635.3135.2935.5535.8436.22
NRMSE0.090.070.060.080.080.070.070.060.04
VIFP0.600.500.650.870.920.930.900.800.94
1DG-USSSIM0.520.580.620.680.690.680.690.690.70
PSNR33.2033.2333.2533.2333.2633.1033.1533.4033.59
NRMSE0.120.120.120.120.120.120.120.110.10
VIFP0.250.210.270.420.380.440.380.410.45
MCP-USSSIM0.760.770.760.960.950.960.960.950.97
PSNR36.7036.4736.7338.9238.7838.9738.7438.9740.55
NRMSE0.050.060.060.030.040.030.040.030.02
VIFP0.520.410.500.900.910.910.890.850.93
Reconstruction time (seconds)0.970.50.300.300.390.330.310.30
Table 2. Average SSIM, PSNR, NRMSE, and VIFP values and reconstruction time of state-of-art methods on fastMRI-T1 axial brain data using several sampling patterns.
Table 2. Average SSIM, PSNR, NRMSE, and VIFP values and reconstruction time of state-of-art methods on fastMRI-T1 axial brain data using several sampling patterns.
SamplingMetricsZero FillingTVWaveletLAEUnetWnetPBCUDRL-NetFDA-CNN
2DG-USSSIM0.520.490.510.700.720.720.760.800.81
PSNR35.8735.6535.8935.1635.2835.1335.5135.4036.23
NRMSE0.070.070.070.080.080.080.080.070.06
VIFP0.780.560.790.840.800.850.820.810.88
1DG-USSSIM0.570.540.570.610.640.630.650.660.71
PSNR32.6432.6232.6432.8933.0832.7133.0333.1834.29
NRMSE0.140.140.140.130.130.140.130.110.09
VIFP0.430.290.420.640.620.690.640.560.64
MCP-USSSIM0.780.760.780.880.870.880.880.880.91
PSNR36.7636.4036.7637.7937.9337.8837.7637.8140.88
NRMSE0.060.060.060.050.040.040.050.040.02
VIFP0.560.400.540.750.770.820.800.730.75
Reconstruction time (seconds)1.521.30.460.490.950.820.470.44
Table 3. Average SSIM, PSNR, NRMSE, and VIFP values and reconstruction time of state-of-the-art methods on the IXI -T2 coronal dataset using several sampling patterns.
Table 3. Average SSIM, PSNR, NRMSE, and VIFP values and reconstruction time of state-of-the-art methods on the IXI -T2 coronal dataset using several sampling patterns.
SamplingMetricsZero FillingTVWaveletLAEUnetWnetPBCUDRL-NetFDA-CNN
2DG-USSSIM0.460.480.500.670.700.710.730.760.77
PSNR33.9733.4633.9834.2534.8534.7135.0234.7535.47
NRMSE0.150.070.070.100.080.080.080.070.06
VIFP0.690.700.670.890.860.900.860.750.97
1DG-USSSIM0.500.520.560.640.660.660.680.710.79
PSNR33.3833.4733.6733.4333.6333.3433.6434.1435.02
NRMSE0.120.120.120.120.110.120.110.090.07
VIFP0.350.190.320.580.570.600.550.560.51
MCP-USSSIM0.700.720.720.820.810.820.830.830.89
PSNR35.7435.4035.4036.3636.5836.4436.5236.6537.70
NRMSE0.070.080.080.060.060.060.060.050.04
VIFP0.440.250.250.710.730.770.730.680.68
Reconstruction time (seconds)0.90.910.330.330.430.330.320.32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hossain, M.B.; Kwon, K.-C.; Imtiaz, S.M.; Nam, O.-S.; Jeon, S.-H.; Kim, N. De-Aliasing and Accelerated Sparse Magnetic Resonance Image Reconstruction Using Fully Dense CNN with Attention Gates. Bioengineering 2023, 10, 22. https://doi.org/10.3390/bioengineering10010022

AMA Style

Hossain MB, Kwon K-C, Imtiaz SM, Nam O-S, Jeon S-H, Kim N. De-Aliasing and Accelerated Sparse Magnetic Resonance Image Reconstruction Using Fully Dense CNN with Attention Gates. Bioengineering. 2023; 10(1):22. https://doi.org/10.3390/bioengineering10010022

Chicago/Turabian Style

Hossain, Md. Biddut, Ki-Chul Kwon, Shariar Md Imtiaz, Oh-Seung Nam, Seok-Hee Jeon, and Nam Kim. 2023. "De-Aliasing and Accelerated Sparse Magnetic Resonance Image Reconstruction Using Fully Dense CNN with Attention Gates" Bioengineering 10, no. 1: 22. https://doi.org/10.3390/bioengineering10010022

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop