0% found this document useful (0 votes)
17 views9 pages

Fnins 16 887634

Uploaded by

i221435
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
17 views9 pages

Fnins 16 887634

Uploaded by

i221435
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 9

ORIGINAL RESEARCH

published: 07 June 2022


doi: 10.3389/fnins.2022.887634

FetalGAN: Automated Segmentation


of Fetal Functional Brain MRI Using
Deep Generative Adversarial
Learning and Multi-Scale 3D U-Net
Josepheen De Asis-Cruz 1 , Dhineshvikram Krishnamurthy 1 , Chris Jose 2 , Kevin M. Cook 1
and Catherine Limperopoulos 1*
1
Developing Brain Institute, Department of Diagnostic Radiology, Children’s National Hospital, Washington, DC,
United States, 2 Department of Computer Science, University of Maryland, College Park, MD, United States

An important step in the preprocessing of resting state functional magnetic resonance


images (rs-fMRI) is the separation of brain from non-brain voxels. Widely used imaging
tools such as FSL’s BET2 and AFNI’s 3dSkullStrip accomplish this task effectively in
children and adults. In fetal functional brain imaging, however, the presence of maternal
tissue around the brain coupled with the non-standard position of the fetal head limit
Edited by:
Weihao Zheng, the usefulness of these tools. Accurate brain masks are thus generated manually,
Lanzhou University, China a time-consuming and tedious process that slows down preprocessing of fetal rs-
Reviewed by: fMRI. Recently, deep learning-based segmentation models such as convolutional neural
Jian Shen,
networks (CNNs) have been increasingly used for automated segmentation of medical
Beijing Institute of Technology, China
Shijie Zhao, images, including the fetal brain. Here, we propose a computationally efficient end-to-
Northwestern Polytechnical University, end generative adversarial neural network (GAN) for segmenting the fetal brain. This
China
method, which we call FetalGAN, yielded whole brain masks that closely approximated
*Correspondence:
Catherine Limperopoulos the manually labeled ground truth. FetalGAN performed better than 3D U-Net model
climpero@childrensnational.org and BET2: FetalGAN, Dice score = 0.973 ± 0.013, precision = 0.977 ± 0.015;
3D U-Net, Dice score = 0.954 ± 0.054, precision = 0.967 ± 0.037; BET2, Dice
Specialty section:
This article was submitted to score = 0.856 ± 0.084, precision = 0.758 ± 0.113. FetalGAN was also faster than 3D
Brain Imaging Methods, U-Net and the manual method (7.35 s vs. 10.25 s vs. ∼5 min/volume). To the best of
a section of the journal
Frontiers in Neuroscience
our knowledge, this is the first successful implementation of 3D CNN with GAN on fetal
Received: 01 March 2022
fMRI brain images and represents a significant advance in fully automating processing
Accepted: 16 May 2022 of rs-MRI images.
Published: 07 June 2022
Keywords: fetal rs-fMRI, resting state, segmentation, deep learning, generative adversarial networks (GANs), 3D
Citation: U-Net, fetal brain
De Asis-Cruz J, Krishnamurthy D,
Jose C, Cook KM and
Limperopoulos C (2022) FetalGAN:
Automated Segmentation of Fetal
INTRODUCTION
Functional Brain MRI Using Deep
Generative Adversarial Learning
Resting state functional MRI (rs-fMRI) is an emergent technique for interrogating in-vivo fetal
and Multi-Scale 3D U-Net. brain function. A critical step in preparing rs-fMRI images for analyses is separating brain from
Front. Neurosci. 16:887634. non-brain voxels. In most cases, fetal brain masks are generated manually, as imaging tools that
doi: 10.3389/fnins.2022.887634 are effectively used for adult whole brain segmentation do not accurately extract the fetal brain.

Frontiers in Neuroscience | www.frontiersin.org 1 June 2022 | Volume 16 | Article 887634


De Asis-Cruz et al. FetalGAN: Automated Fetal EPI Brain Segmentation

This suboptimal performance likely arises from the presence of effectively modeling any distribution of data (Gonog and
surrounding maternal tissue, non-standard orientation of the Zhou, 2019). The generative network creates new examples
fetal head, and reduced gray/white matter contrast in the fetal of the data while the discriminator simultaneously evaluates
brain. While manual segmentation of the fetal brain provides these exemplars in a cyclic fashion effectively giving rise to
reasonable brain masks, the process is time consuming and a network that self-optimizes its error rate and converges on
operator dependent. Automated processes have the potential to a model with high accuracy. Specifically, adversarial losses
increase efficiency of pipelines and reproducibility of results. enforced by the discriminator network consider higher-order
A growing body of literature has demonstrated that deep potentials, as opposed to the pairwise correlations utilized by
learning-based segmentation outperforms traditional approaches voxel-wise loss functions, such as softmax. This adversarial loss
including multi-atlas registration techniques (Huo et al., 2019; serves as an adaptively learned similarity measure between the
Khalili et al., 2019; Dolz et al., 2020; Zhao et al., 2022). predicted segmentation label maps and the annotated ground
Deep convolutional neural networks (CNN) such as U-Net truth that improves localization accuracy while enforcing spatial
have achieved remarkable success for anatomical medical image contiguity at low contrast regions, including image boundaries.
segmentation and have been shown to be versatile and effective Various end-to-end adversarial neural networks (e.g., SegAN)
(Ronneberger et al., 2015; Yang et al., 2018; Zhao et al., 2018; have been proposed as stable and effective frameworks for
Son et al., 2020). Recently, 2D U-Net has been successfully automatic segmentation (SegAN) of organs such as the brain,
applied to fetal resting state functional MRI data (Rutherford chest, and abdomen, among others (Frid-Adar et al., 2018;
et al., 2021), a crucial step in automating preprocessing of Giacomello et al., 2020; Xun et al., 2021; Zhu et al., 2021).
fetal rs-fMRI. However, there are several limitations in using Furthermore, a recent study by Chen et al. (2022) showed
CNN-based approaches for segmentation (Ronneberger et al., that a GAN-based paradigm improved the robustness and
2015; Xue et al., 2018; Li et al., 2019; Rutherford et al., generalizability of deep learning models like graph neural
2021). Although U-Nets can use skip connections to combine networks (GNNs). Using their model on multi-modal MRI
both low- and high-level features, there is no guarantee of data, they identified autism spectrum disorders (ASD) with
spatial consistency in the final segmentation map, especially higher accuracy (74.7%) compared to other state-ot-the-art deep
at the boundaries (Isola et al., 2017; Yang et al., 2018; learning methods.
Zhao et al., 2018; Dhinagar et al., 2021). To address this Motivated by SegAN, here, we propose FetalGAN, a GAN
limitation, methods that consider spatial correlations among based end-to-end architecture for the automated segmentation
neighboring pixels such as conditional random field and other of fetal rs-fMRI brain images. FetalGAN addresses the previously
graph cut techniques are used as post-processing refinement described drawbacks of deep CNNs and may be better suited for
(Pereira et al., 2016b; Nancy, 2019; Son et al., 2020). Utilizing low-contrast fetal rs-fMRI. We hypothesized that FetalGAN will
pair-wise potentials, however, may cause serious boundary produce whole brain labels that closely approximate the manually
leakage, especially in low-contrast regions (Vijayanarasimhan created ground truth and will outperform deep CNN-based
and Grauman, 2010). To prevent leakage and the lack of models (i.e., 3D U-Net) and the commonly used BET2 algorithm.
spatial consistency, methods such as patch-based networks for
training CNNs and multi-scale, multi-path CNNs with different
input resolutions/network architectures have been used (Pereira MATERIALS AND METHODS
et al., 2016a; Havaei et al., 2017; Kamnitsas et al., 2017;
Chattopadhay et al., 2018; Xiao et al., 2020; Ghimire et al., Data
2021; Zhang et al., 2021; Zhu et al., 2021). However, patch- We initially evaluated 75 rs-fMRI scans. Out of the 75 datasets,
based training is computationally costly. Moreover, finding the four were excluded from further analyses: three had image
optimal patch size that achieves superior localization accuracy dimensions (x, y, or z) that exceeded the chosen patch size of
is challenging. Generally, traditional CNNs have a tradeoff 32 × 32 × 32, and one had incomplete demographic data. The
between achieving good localization performance/higher level final sample consisted of 71 datasets from 64 healthy fetuses.
of semantics (i.e., correctly classifying each voxel’s label) and Pregnant women were recruited as part of a larger study
crisper, more well-defined boundaries. This is a potential investigating brain development in healthy and high-risk fetuses.
disadvantage specifically when applied to brain segmentation of All participants had normal ultrasonograms/echocardiograms
fetal rs-fMRI, which often have low-contrast boundaries, varied and structurally normal brains on MRI. Fetal exclusion
voxel intensities, and features at different scales/orientations criteria included: dysmorphic features by antenatal
(Ronneberger et al., 2015; Xue et al., 2018; Dolz et al., 2020; ultrasound, chromosomal abnormalities by amniocentesis,
Rutherford et al., 2021). evidence of congenital infections, presentation after
Recently, generative adversarial networks (GANs) have been 28 weeks gestational age, and multiple gestation. Maternal
shown to be a robust approach for automated medical image exclusion criteria included: pregnant women with known
segmentation and to yield better, stable performance compared to psychiatric/metabolic/genetic disorders, complicated/multiple
state-of-the-art CNN-based models (Isola et al., 2017; Xue et al., pregnancies, alcohol and/or tobacco use, maternal medications,
2018; Xun et al., 2021). Using two competing neural networks— and contraindications to MRI.
a generator and a discriminator—GANs create exemplar images Data were collected using a 1.5T GE MRI scanner (GE
that are difficult to distinguish from real (i.e., training) images, Healthcare, Milwaukee, WI) with an 8-channel receiver

Frontiers in Neuroscience | www.frontiersin.org 2 June 2022 | Volume 16 | Article 887634


De Asis-Cruz et al. FetalGAN: Automated Fetal EPI Brain Segmentation

coil. Anatomical single-shot fast spin-echo anatomical T2- in the decoding branch, the downsampling layers resized by
weighted images were collected with the following parameters: a factor of 2 and used a kernel size of 2 × 2 × 2. The
TR = 1,100 ms, TE = 160 ms, flip angle = 90◦ , and slice discriminator network’s structure was like the deconvolution
thickness = 2 mm. Resting-state echo planar images (EPI) images block of the generator. Receiving both the ground truth and
were collected with the following parameters: TR = 3,000 ms, predicted label map, the discriminator extracted hierarchical
TE = 60 ms, voxel size = 2.578 mm × 2.578 mm × 3 mm, flip features to quantify differences between these two input images.
angle = 90◦ , field of view = 33 cm, matrix size = 128 × 128, Please see Supplementary Material for a summary of generator
and scan duration = 7 min (140 volumes). On average, and discriminator parameters.
5:21 min (107 volumes) of resting-state data was available SegAN learns a loss function that penalizes structural
after preprocessing. differences between the discriminator network output and target
(Xue et al., 2018). Rather than computing discriminator loss for
Preprocessing the entire network, we computed loss at each discriminator layer.
Fetal resting state data were preprocessed up to the point of The multi-scale loss function L was defined by Xue et al. (2018) as,
brain segmentation using AFNI, unless specified otherwise (Cox,
1996). Briefly, as previously described here (De Asis-Cruz et al., N
1 X
2021), fetal EPI images were slice time corrected, trimmed by min max L (θG , θD ) = lmae (fD (xn · G (xn )) , fD (xn · yn ))
θG θD N
removing the first four volumes to stabilize magnetic gradients, n=1
manually oriented to radiologic orientation using landmark (1)
based rigid registration (IRTK1 ), despiked, and then corrected for
bias-field inhomogeneities (N4BiasFieldCorrection) (Tustison where x is the training image; y its corresponding ground
et al., 2010). At this point, the oriented EPI images were truth; N is the number of training images; lmae is the
ready for motion correction. For this step, we used a two- mean absolute error (MAE) or L1 distance; xn · G(xn ) is the
pass registration approach optimized to correct for the high- probabilistic map generated by the generator network; xn ·
motion typically observed in fetuses and newborns (Joshi yn is the input image masked by its corresponding ground
et al., 2011; Scheinost et al., 2018). This method required truth; and fD (x) represent the hierarchical features extracted
two inputs: a reference volume and its mask. For each from image x by the discriminator network. Using a multi-
resting state (RS) dataset, a reference volume was defined scale loss function to quantify training error, the network
using AFNI’s 3dToutcount; this identifies the volume with sequentially learned both global and local features and encoded
the lowest fraction of outlier voxels based on signal intensity long and short-range spatial relationships between voxels. As
trend. A brain mask was then manually drawn (JDC) for training progressed, the generator network was able to produce
each reference brain volume using ITK-SNAP (Yushkevich probabilistic predictions that more closely approximated the
et al., 2006). The goal was to automatically create this whole expert-annotated, ground truth.
brain mask and provide it as input to the motion correction
algorithm. The selected reference volume and the manual brain 3D U-Net Architecture
mask were utilized as inputs for training the model. During 3D U-Net, patch-based architecture was also performed
testing, the reference image was segmented using three different (Figure 1, top). The network consisted of both an expanding and
approaches: FSL Brain Extraction Tool v2 (BET2) (Smith, 2002), contracting path. Here, the contracting path was supplemented
3D U-Net (Çiçek et al., 2016), and FetalGAN. Segmentation with successive layers where the standard pooling operators
outputs were compared to the manually created mask using were replaced with upsampling operators to enhance image
the following metrics: Dice index, Jaccard score, sensitivity, resolution. The high-resolution feature from the contracting
specificity, and precision. We also reported the computation time path was then concatenated with the upsampled features
for each method. from the expanding path for localization of the fetal brain.
The expanding and contracting paths had four convolutional
SegAN Architecture blocks, each with two Conv3D layers, BatchNormalization, and
We used the GAN framework to automatically segment the fetal the PReLU activation function. In each convolutional block,
brain from rs-fMRI scans. The algorithm consisted of two neural the number of feature maps was doubled per layer (96 initial
networks: the generator (segmenter) based on 3D U-Net, and feature maps and 364 feature maps generated after the last
the discriminator (critic) based on a fully convolutional decoder block); a kernel size of 3 and 2 was used for the expanding
network (Xue et al., 2018). and contracting paths, respectively. At the junction of the
The generator network received a 3D patch as an input and contracting/expanding path, the layers were regularized using
consisted of eight residual convolutional blocks with the leaky dropout with a rate of 15%. In the expanding path, a MaxPooling
rectified linear unit (ReLU) activation, batch normalization, and (downsampling) layer with stride 2 followed each convolution
maxpooling layers (Figure 1, top; see 3D U-Net Architecture block to encode the input 3D patches into feature representations
for details). In the encoding branch, the upsampling layers at different levels. Deconvolution layers (upsampling) were
had a kernel size of 3 × 3 × 3 with stride 2 × 2 × 2; used intermittently throughout the contracting path to increase
the density of the sparse feature maps of the expanding path
1
https://github.com/BioMedIA/IRTK using a transpose convolution with multiple trainable filters.

Frontiers in Neuroscience | www.frontiersin.org 3 June 2022 | Volume 16 | Article 887634


De Asis-Cruz et al. FetalGAN: Automated Fetal EPI Brain Segmentation

FIGURE 1 | Architecture of proposed FetalGAN network.

TABLE 1 | Comparison of FetalGAN, 3D U-Net, and BET2.

FetalGAN 3D U-Net BET2

Mean ± SD Mean ± SD p* Mean ± SD p*

Dice 0.973 ± 0.013 0.954 ± 0.054 9.260 × 10−4 0.856 ± 0.084 1.124 × 10−18
Jaccard 0.948 ± 0.024 0.916 ± 0.082 1.993 × 10−4 0.756 ± 0.113 4.910 × 10−23
Precision 0.977 ± 0.015 0.967 ± 0.037 0.043 0.758 ± 0.113 6.685 × 10−26
Sensitivity 0.971 ± 0.021 0.945 ± 0.077 0.002 0.996 ± 0.011 1.493 × × 10−17
Specificity 0.994 ± 0.005 0.992 ± 0.010 0.239 0.915 ± 0.051 3.703 × 10−21
Time/patch (s) 0.05 0.08 -
Time/vol (s) 7.35 10.25 4.40

*FetalGAN compared to 3D U-Net and BET2, asterisk (*) indicates significant difference between method and FetalGAN using paired t-test.

The successive downsampling and upsampling feature maps The final segmentation results were derived by averaging
were concatenated to localize and learn representations after together the probability maps of each overlapped patch. The
each convolution. model training and validation are performed on NVIDIA
V100 multi-GPU. After prediction, isolated and/or misidentified
Training Specifications voxels were corrected, and internal holes were filled using
The SegAN was trained using a multi-scale loss function, the morphological operations available in the openCV library
U-Net model using binary cross-entropy loss. For both, weights (Bradski, 2000).
were determined using an Adam optimizer (Kingma and Ba,
2014). The models were trained on 71 reference brain volumes
and their corresponding manually drawn masks. We used RESULTS
k-fold cross validation, where k = 5, for evaluating the model’s
performance. Each 3D MR scan and its respective normalized We evaluated 71 datasets from 64 healthy fetuses between 25
mask was split into patches of size 32 × 32 × 32 with stride and 39.43 gestational weeks (mean GA ± SD: 33.28 ± 3.79; see
2 × 2 × 2 and fed into the model for training. Given a Supplementary Material for age distribution). The average scan
test EPI image, we extracted the overlapped patches with size interval for the seven fetuses with two scans is 7.63 ± 2.48 weeks.
32 × 32 × 32, and fed them to the trained network to obtain the The proposed SegAN method was more time efficient than 3D
final probability label map. U-Net, requiring, on average, 7.35 s to segment a single volume

Frontiers in Neuroscience | www.frontiersin.org 4 June 2022 | Volume 16 | Article 887634


De Asis-Cruz et al. FetalGAN: Automated Fetal EPI Brain Segmentation

FIGURE 2 | Representative whole brain masks from manual segmentation, BET2, 3D U-Net, and FetalGAN. Manual corrections were done using ITK-SNAP.
FetalGAN produced the most accurate segmentation relative to the ground truth with an average Dice score of 0.942 ± 0.095. (A) 25 4/7 weeks, (B) 29 2/7 weeks,
(C) 34 4/7 weeks, and (D) 38 6/7 weeks.

compared to 10.25 s for the latter (Table 1). BET2 was the fastest 3D U-Net and higher than BET2. Visual inspection of
algorithm, needing only 4.40 s to extract the brain. representative scans (Figure 2) showed FetalGAN outputs
Whole brain segmentation outputs of FetalGAN, FSL’s more closely resembled the ground truth. The 3D reconstructed
BET2, and 3D U-Net were compared to the manually surface of the segmentations by SegAN and U-Net is
segmented brains. The proposed method’s Dice score, smoother when compared to the outputs of the manual
Jaccard index, precision, and specificity were significantly and BET segmentation.
higher than 3d U-Net and BET2 (paired t-test p < 0.05; FetalGAN and 3D U-Net performance showed stability
see Table 1). FetalGAN’s specificity was comparable to across GAs (Figure 3). The Dice and Jaccard scores for these

Frontiers in Neuroscience | www.frontiersin.org 5 June 2022 | Volume 16 | Article 887634


De Asis-Cruz et al. FetalGAN: Automated Fetal EPI Brain Segmentation

FIGURE 3 | Performance scores fo FetalGAN, 3D U-Net and BET2 methods across gestation ages: (A) Dice coefficient, (B) Jaccard Score, (C) Precision,
(D) Sensitivity, and (E) Specificity.

two models were uncorrelated with age (Pearson r = –0.114, To the best of our knowledge, the proposed method is
p = 0.230 and r = –0.1410, p = 0.241, FetalGAN and 3D U-Net, the first successful application of the SegAN framework for
respectively; see Supplementary Table 1). FetalGAN specificity segmentation of the fetal EPI brain and only the second
decreased with increasing GA. Despite this decrease, specificity automated tool for accurately separating the fetal brain from
remained high (range: 0.9723–0.9993) and was comparable to surrounding maternal tissue (Rutherford et al., 2021). We
3D U-Net and significantly better than BET2. Unlike the deep speculate that the modifications applied to the conventional
learning models, BET2 Dice coefficients and Jaccard indices were GAN framework accounted for the superior performance of
positively correlated with age (r = 0.558, p = 4.228 × 10−7 FetalGAN over the 3D U-Net model. Previous, conventional
and r = 0.564, p = 2.985 × 10−7 , respectively). Precision GAN approaches have been reported to be unstable at times
also positively scaled with increasing GA for both BET2 due to failures during training, such as vanishing gradients
(r = 0.568, p = 2.396 × 10−7 ) and 3D U-Net (r = 0.317, and non-convergence (Isola et al., 2017; Xue et al., 2018). In
p = 0.007). other words, the adversarial loss, which classifies the image
based on a scalar output by the discriminator, was unable to
propagate sufficient gradients to improve the performance of
DISCUSSION the generator network (i.e., insufficient information passed on to
the generator). FetalGAN utilized a multi-scale, weighted feature
We successfully implemented FetalGAN, a SegAN-based model, loss function, which effectively quantified minute differences
to accurately extract the fetal EPI brain from the maternal between the generated and ground truth segmentation across
compartment in a sample of 71 normative fetal rs-fMRI multiple layers of the network. This enabled both the generator
datasets. The whole brain mask generated by FetalGAN closely and discriminator networks to learn hierarchical features
approximated manually segmented images. The proposed model that captured relationships between voxels, especially in low
produced outputs superior to labels derived from 3D U-Net contrast regions around the boundary between the fetal
and FSL’s BET2. FetalGAN masks were also generated at a brain and maternal tissue. Altogether, these permitted the
faster rate than U-Net and with only a minimal increase training process of FetalGAN to be end-to-end and stable.
in preprocessing time compared to BET2. In addition, the Moreover, FetalGAN performed faster than the comparable
proposed method produced consistently accurate segmentation 3D U-Net implementation because the number of trainable
across gestational ages. These findings suggest that FetalGAN parameters in the generator network was less than a 3D
is a robust, fast, and reliable approach to segmenting fetal rs- U-Net model. FetalGAN also outperformed BET2, likely because
fMRI images. the boundary between fetal brain and non-brain voxels was

Frontiers in Neuroscience | www.frontiersin.org 6 June 2022 | Volume 16 | Article 887634


De Asis-Cruz et al. FetalGAN: Automated Fetal EPI Brain Segmentation

low-contrast and BET2 relied on intensity differences between would support the generalizability of FetalGAN. Lastly,
tissues to accurately estimate the boundary of the brain the paper demonstrated FetalGAN’s superior performance,
(Smith, 2002). but further studies that integrate brain extraction with
One recent study successfully implemented 2D U-Net to other preprocessing steps to yield a fully automated
automatically segment the fetal EPI brain (Rutherford et al., pipeline are needed.
2021). Trained on 855 images from 129 subjects, their With mounting evidence supporting the fetal origins of
model yielded slightly lower performance metrics compared many prevalent adult disorders including mental illness (Barker
to FetalGAN (2D U-Net: Dice score = 0.94 ± 0.069, Jaccard et al., 2009; Al-Haddad et al., 2019), there has been increased
index = 0.89 ± 0.069 vs. FetalGAN: Dice score = 0.973 ± 0.013, interest in investigating fetal functional brain development
Jaccard index = 0.948 ± 0.024). In the 2D U-Net model, in vivo using MRI. FetalGAN, an implementation of SegAN
images were segmented in their original space; in contrast, for fetal rs-fMRI brain, offers a fast, automated, unbiased,
FetalGAN was applied to oriented images. During development and accurate alternative to currently available fetal EPI
of our pipeline, we observed that orienting images prior brain extraction techniques. Further improvements that focus
to brain extraction allowed more options in subsequent on increasing computational efficiency, extracting the brain
preprocessing steps, thus we repositioned the brains prior in the original space, and integrating FetalGAN into a
to segmentation. Another critical difference between the two fully automated fetal rs-fMRI pipeline, among others, are
models is that FetalGAN was trained using 3D patches, thus currently underway. It is our hope that this technique
it can leverage spatial information across three dimensions would help facilitate in utero investigations of emerging
(i.e., interslice relationships) whereas 2D convolutional kernels functional connectivity.
obtain context only across the width and height of a slice.
Moreover, with 3D U-Net, warping or normalization was
not required. While we did not directly compare 2D and DATA AVAILABILITY STATEMENT
3D U-Net models, previous studies have demonstrated the
advantage of 3D over 2D CNNs (Nemoto et al., 2020; The datasets presented will be provided upon reasonable
Woo and Lee, 2021). request to the corresponding author. Requests
FetalGAN aims to provide an automated alternative to to access the datasets should be directed to CL,
manual segmentation of fetal rs-fMRI data. FetalGAN addresses climpero@childrensnational.org.
drawbacks inherent to manual processes. First, since the
process is automated, outputs are replicable. Second, the need
for highly skilled operators is eliminated. Lastly, relative to ETHICS STATEMENT
manual segmentation, the time required to segment a brain
volume is markedly reduced. Taken together, these three main The studies involving human participants were reviewed and
areas of improvement are a critical step toward increasing approved by the Institutional Review Board at Children’s
rigor and reproducibility in fetal neuroimaging. While this National Hospital. The patients/participants provided their
is but one of the first steps in fetal rs-fMRI preprocessing, written informed consent to participate in this study.
we believe that our proposed method will contribute to
the field’s broader and overarching goal of creating fully
automated pipelines such as what’s currently available for older AUTHOR CONTRIBUTIONS
children and adults with SPM,2 AFNI (Cox, 1996), or FSL
(Jenkinson et al., 2012) (or pipelines that combine these such JD-C: conceptualization, methodology, validation, formal
as fMRIPrep3 and CPAC,4 among others). The widespread analysis, investigation, data curation, writing—original
availability of these tools to the larger scientific community has draft, review, and editing, and supervision. DK and CJ:
been instrumental in advancing our understanding of human conceptualization, methodology, validation, formal analysis,
health and disease. investigation, data curation, writing—original draft, review,
Our work has several limitations. First, we used fewer and editing, and visualization. KC: investigation and writing—
training data sets for fetal EPI brain segmentation compared original draft, review, and editing. CL: conceptualization,
to a previous study (Rutherford et al., 2021). With the methodology, resources, writing—review and editing,
smaller sample size, however, we achieved comparable supervision, and funding acquisition. All authors contributed to
performance. Moreover, it should be noted that our inputs the article and approved the submitted version.
are 3D rather than 2D, thus the information that is fed
into the learning model is likely comparable. Second,
we used data from a single site. Additional studies that FUNDING
test the model on data collected from other institutions
This study was funded by grant R01 HL116585–01 from the
2
https://www.fil.ion.ucl.ac.uk/spm/software/spm12/ National Heart, Lung, and Blood Institute, National Institutes of
3
https://fmriprep.org/en/stable/index.html Health (CL) and grant MOP-81116 from the Canadian Institute
4
https://fcp-indi.github.io/docs/latest/user/quick of Health Research (CL).

Frontiers in Neuroscience | www.frontiersin.org 7 June 2022 | Volume 16 | Article 887634


De Asis-Cruz et al. FetalGAN: Automated Fetal EPI Brain Segmentation

ACKNOWLEDGMENTS SUPPLEMENTARY MATERIAL


We thank the pregnant women who participated in this study. The Supplementary Material for this article can be found
We are grateful to our study team for supporting recruitment, online at: https://www.frontiersin.org/articles/10.3389/fnins.
enrollment, and for performing the fetal MRI studies. 2022.887634/full#supplementary-material

REFERENCES Isola, P., Zhu, J.-Y., Zhou, T., and Efros, A. A. (2017). “Image-to-image translation
with conditional adversarial networks,” in Proceeding of the 2017 IEEE
Al-Haddad, B. J. S., Jacobsson, B., Chabra, S., Modzelewska, D., Olson, E. M., Conference on Computer Vision and Pattern Recognition (CVPR) (IEEE), doi:
Bernier, R., et al. (2019). Long-term risk of neuropsychiatric disease after 10.1109/cvpr.2017.632
exposure to infection in utero. JAMA Psychiatry 76, 594–602. doi: 10.1001/ Jenkinson, M., Beckmann, C. F., Behrens, T. E. J., Woolrich, M. W., and Smith,
jamapsychiatry.2019.0029 S. M. (2012). FSL. Neuroimage 62, 782–790. doi: 10.1016/j.neuroimage.2011.09.
Barker, D. J. P., Osmond, C., Kajantie, E., and Eriksson, J. G. (2009). Growth 015
and chronic disease: findings in the helsinki birth cohort. Ann. Hum. Biol. 36, Joshi, A., Scheinost, D., Okuda, H., Belhachemi, D., Murphy, I., Staib, L. H., et al.
445–458. doi: 10.1080/03014460902980295 (2011). Unified framework for development, deployment and robust testing of
Bradski, G. (2000). The openCV library. Dr. Dobb’s J. Soft. Tools Profess. Progr. 25, neuroimaging algorithms. Neuroinformatics 9, 69–84. doi: 10.1007/s12021-010-
120–123. 9092-8
Chattopadhay, A., Sarkar, A., Howlader, P., and Balasubramanian, V. N. Kamnitsas, K., Ledig, C., Newcombe, V. F. J., Simpson, J. P., Kane, A. D., Menon,
(2018). “Grad-CAM++: generalized gradient-based visual explanations for deep D. K., et al. (2017). Efficient multi-scale 3D CNN with fully connected CRF for
convolutional networks,” in Proceeding of the 2018 IEEE Winter Conference on accurate brain lesion segmentation. Med. Image Anal. 36, 61–78. doi: 10.1016/j.
Applications of Computer Vision (WACV) (IEEE), doi: 10.1109/wacv.2018.00097 media.2016.10.004
Chen, Y., Yan, J., Jiang, M., Zhang, T., Zhao, Z., Zhao, W., et al. (2022). Khalili, N., Turk, E., Benders, M. J. N. L., Moeskops, P., Claessens, N. H. P., de
Adversarial learning based node-edge graph attention networks for autism Heus, R., et al. (2019). Automatic extraction of the intracranial volume in fetal
spectrum disorder identification. Trans. Neural Netw. Learn. Syst. 2022:4755. and neonatal MR scans using convolutional neural networks. NeuroImage Clin.
doi: 10.1109/TNNLS.2022.3154755 24:102061. doi: 10.1016/j.nicl.2019.102061
Çiçek, Ö, Abdulkadir, A., Lienkamp, S. S., Brox, T., and Ronneberger, O. (2016). Kingma, D. P., and Ba, J. (2014). Adam: a method for stochastic optimization. arXiv
“3D U-net: learning dense volumetric segmentation from sparse annotation,” in [preprint] doi: 10.48550/arXiv.1412.6980
Medical Image Computing and Computer-Assisted Intervention – MICCAI 2016 Li, J., Gao, M., and D’Agostino, R. (2019). Evaluating classification accuracy for
Lecture Notes in Computer Science, eds S. Ourselin, L. Joskowicz, M. R. Sabuncu, modern learning approaches. Stat. Med. 38, 2477–2503. doi: 10.1002/sim.8103
G. Unal, and W. Wells (Cham: Springer International Publishing), 424–432. Nancy, W. A. (2019). Brain tumor segmentation using normalized graph cuts. Int.
Cox, R. W. (1996). AFNI: software for analysis and visualization of functional J. Eng. Adv. Technol. 8, 3159–3161. doi: 10.35940/ijeat.F9258.088619
magnetic resonance neuroimages. Comput. Biomed. Res. 29, 162–173. doi: 10. Nemoto, T., Futakami, N., Yagi, M., Kumabe, A., Takeda, A., Kunieda, E., et al.
1006/cbmr.1996.0014 (2020). Efficacy evaluation of 2D, 3D U-Net semantic segmentation and atlas-
De Asis-Cruz, J., Andersen, N., Kapse, K., Khrisnamurthy, D., Quistorff, J., based segmentation of normal lungs excluding the trachea and main bronchi.
Lopez, C., et al. (2021). Global network organization of the fetal functional J. Radiat. Res. 61, 257–264. doi: 10.1093/jrr/rrz086
connectome. Cereb. Cortex 2021:410. doi: 10.1093/cercor/bhaa410 Pereira, S., Pinto, A., Oliveira, J., Mendrik, A. M., Correia, J. H., and Silva, C. A.
Dhinagar, N. J., Thomopoulos, S. I., Owens-Walton, C., Stripelis, D., Ambite, J. L., (2016b). Automatic brain tissue segmentation in MR images using random
Ver Steeg, G., et al. (2021). 3D convolutional neural networks for classification forests and conditional random fields. J. Neurosci. Methods 270, 111–123. doi:
of Alzheimer’s and parkinson’s disease with T1-weighted brain MRI. bioRxiv 10.1016/j.jneumeth.2016.06.017
[Preprint] doi: 10.1101/2021.07.26.453903 Pereira, S., Pinto, A., Alves, V., and Silva, C. A. (2016a). Brain tumor segmentation
Dolz, J., Desrosiers, C., Wang, L., Yuan, J., Shen, D., and Ben Ayed, I. (2020). Deep using convolutional neural networks in MRI images. IEEE Trans. Med. Imaging
CNN ensembles and suggestive annotations for infant brain MRI segmentation. 35, 1240–1251. doi: 10.1109/TMI.2016.2538465
Comput. Med. Imaging Graph. 79:101660. doi: 10.1016/j.compmedimag.2019. Ronneberger, O., Fischer, P., and Brox, T. (2015). “U-net: convolutional networks
101660 for biomedical image segmentation,” in Medical Image Computing and
Frid-Adar, M., Diamant, I., Klang, E., Amitai, M., Goldberger, J., and Greenspan, Computer-Assisted Intervention – MICCAI 2015, eds N. Navab, J. Hornegger,
H. (2018). GAN-based synthetic medical image augmentation for increased W. Wells, and A. Frangi (Cham: Springer International Publishing), 234–241.
CNN performance in liver lesion classification. Neurocomputing 321, 321–331. Rutherford, S., Sturmfels, P., Angstadt, M., Hect, J., Wiens, J., van den Heuvel, M. I.,
doi: 10.1016/j.neucom.2018.09.013 et al. (2021). Automated brain masking of fetal functional MRI with open data.
Ghimire, K., Chen, Q., and Feng, X. (2021). “Patch-based 3D UNet for head Neuroinformatics 2021:528. doi: 10.1007/s12021-021-09528-5
and neck tumor segmentation with an ensemble of conventional and dilated Scheinost, D., Onofrey, J. A., Kwon, S. H., Cross, S. N., Sze, G., Ment, L. R., et al.
convolutions,” in Head and Neck Tumor Segmentation Lecture Notes in (2018). “A fetal fMRI specific motion correction algorithm using 2nd order
Computer Science, (Cham: Springer International Publishing), 78–84.doi: 10. edge features,” in Proceeding of the 2018 IEEE 15th International Symposium
1007/978-3-030-67194-5_9 on Biomedical Imaging (ISBI 2018), (IEEE), doi: 10.1109/isbi.2018.8363807
Giacomello, E., Loiacono, D., and Mainardi, L. (2020). “Brain MRI tumor Smith, S. M. (2002). Fast robust automated brain extraction. Hum. Brain Mapp. 17,
segmentation with adversarial networks,” in Proceeding of the 2020 International 143–155. doi: 10.1002/hbm.10062
Joint Conference on Neural Networks (IJCNN) (IEEE), doi: 10.1109/ijcnn48605. Son, H. H., Phuong, P. C., van Walsum, T., and Ha, L. M. (2020). Liver
2020.9207220 segmentation on a variety of computed tomography (CT) images based on
Gonog, L., and Zhou, Y. (2019). “A review: generative adversarial networks,” in convolutional neural networks combined with connected components. VNU J.
Proceeding of the 2019 14th IEEE Conference on Industrial Electronics and Sci. Comput. Sci. Commun. Eng. 36:241. doi: 10.25073/2588-1086/vnucsce.241
Applications (ICIEA) (IEEE), doi: 10.1109/iciea.2019.8833686 Tustison, N. J., Avants, B. B., Cook, P. A., Zheng, Y., Egan, A., Yushkevich, P. A.,
Havaei, M., Davy, A., Warde-Farley, D., Biard, A., Courville, A., Bengio, Y., et al. et al. (2010). N4ITK: improved N3 bias correction. IEEE Trans. Med. Imaging
(2017). Brain tumor segmentation with deep neural networks. Med. Image Anal. 29, 1310–1320. doi: 10.1109/TMI.2010.2046908
35, 18–31. doi: 10.1016/j.media.2016.05.004 Vijayanarasimhan, S., and Grauman, K. (2010). “Top-down pairwise potentials for
Huo, Y., Xu, Z., Xiong, Y., Aboud, K., Parvathaneni, P., Bao, S., et al. (2019). piecing together multi-class segmentation puzzles,” in Proceeding of the 2010
3D whole brain segmentation using spatially localized atlas network tiles. IEEE Computer Society Conference on Computer Vision and Pattern Recognition
Neuroimage 194, 105–119. doi: 10.1016/j.neuroimage.2019.03.041 - Workshops, (IEEE), doi: 10.1109/cvprw.2010.5543728

Frontiers in Neuroscience | www.frontiersin.org 8 June 2022 | Volume 16 | Article 887634


De Asis-Cruz et al. FetalGAN: Automated Fetal EPI Brain Segmentation

Woo, B., and Lee, M. (2021). “Comparison of tissue segmentation performance Zhao, X., Wu, Y., Song, G., Li, Z., Zhang, Y., and Fan, Y. (2018). “3D brain tumor
between 2D U-Net and 3D U-net on brain MR images,” in Proceeding segmentation through integrating multiple 2D FCNNs,” in Brainlesion: Glioma,
of the 2021 International Conference on Electronics, Information, Multiple Sclerosis, Stroke and Traumatic Brain Injuries Lecture notes in computer
and Communication (ICEIC), (IEEE), doi: 10.1109/iceic51217.2021.93 science, eds A. Crimi, S. Bakas, H. Kuijf, B. Menze, and M. Reyes (Cham:
69797 Springer International Publishing), 191–203.
Xiao, Z., Liu, B., Geng, L., Zhang, F., and Liu, Y. (2020). Segmentation of lung Zhu, C., Mei, K., Peng, T., Luo, Y., Liu, J., Wang, Y., et al. (2021). Multi-
nodules using improved 3D-UNet neural network. Symmetry (Basel) 12:1787. level colonoscopy malignant tissue detection with adversarial CAC-UNet.
doi: 10.3390/sym12111787 Neurocomputing 438, 165–183. doi: 10.1016/j.neucom.2020.04.154
Xue, Y., Xu, T., Zhang, H., Long, L. R., and Huang, X. (2018). SegAN:
adversarial network with multi-scale L1 loss for medical image segmentation. Conflict of Interest: The authors declare that the research was conducted in the
Neuroinformatics 16, 383–392. doi: 10.1007/s12021-018-9377-x absence of any commercial or financial relationships that could be construed as a
Xun, S., Li, D., Zhu, H., Chen, M., Wang, J., Li, J., et al. (2021). Generative potential conflict of interest.
adversarial networks in medical image segmentation: a review. Comput. Biol.
Med. 140, 105063. doi: 10.1016/j.compbiomed.2021.105063 The handling editor WZ declared a past co-authorship with CL.
Yang, C., Rangarajan, A., and Ranka, S. (2018). Visual explanations from deep
3D convolutional neural networks for Alzheimer’s disease classification. AMIA Publisher’s Note: All claims expressed in this article are solely those of the authors
Ann. Symp. Proc. 2018, 1571–1580. and do not necessarily represent those of their affiliated organizations, or those of
Yushkevich, P. A., Piven, J., Hazlett, H. C., Smith, R. G., Ho, S., Gee, J. C., et al. the publisher, the editors and the reviewers. Any product that may be evaluated in
(2006). User-guided 3D active contour segmentation of anatomical structures: this article, or claim that may be made by its manufacturer, is not guaranteed or
significantly improved efficiency and reliability. Neuroimage 31, 1116–1128. endorsed by the publisher.
doi: 10.1016/j.neuroimage.2006.01.015
Zhang, Y., Zhong, P., Jie, D., Wu, J., Zeng, S., Chu, J., et al. (2021). Copyright © 2022 De Asis-Cruz, Krishnamurthy, Jose, Cook and Limperopoulos.
Brain tumor segmentation from multi-modal MR images via This is an open-access article distributed under the terms of the Creative Commons
ensembling UNets. Front. Radio 1:704888. doi: 10.3389/fradi.2021.70 Attribution License (CC BY). The use, distribution or reproduction in other forums
4888 doi: 10.3389/fradi.2021.704888 is permitted, provided the original author(s) and the copyright owner(s) are credited
Zhao, L., De Asis-Cruz, J., Feng, X., Wu, Y., Kapse, K., Largent, A., et al. (2022). and that the original publication in this journal is cited, in accordance with accepted
Automated 3D fetal brain segmentation using an optimized deep learning academic practice. No use, distribution or reproduction is permitted which does not
approach. AJNR Am. J. Neuroradiol. 2022:7419. doi: 10.3174/ajnr.A7419 comply with these terms.

Frontiers in Neuroscience | www.frontiersin.org 9 June 2022 | Volume 16 | Article 887634

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy