
In clinical diagnostics, magnetic resonance imaging (MRI) technology plays a crucial role in the recognition of cardiac regions, serving as a pivotal tool to assist physicians in diagnosing cardiac diseases. Despite the notable success of convolutional neural networks (CNNs) in cardiac MRI segmentation, it remains a challenge to use existing CNNs-based methods to deal with fuzzy information in cardiac MRI. Therefore, we proposed a novel network architecture named DAFNet to comprehensively address these challenges.
The proposed method was used to design a fuzzy convolutional module, which could improve the feature extraction performance of the network by utilizing fuzzy information that was easily ignored in medical images while retaining the advantage of attention mechanism. Then, a multi-scale feature refinement structure was designed in the decoder portion to solve the problem that the decoder structure of the existing network had poor results in obtaining the final segmentation mask. This structure further improved the performance of the network by aggregating segmentation results from multi-scale feature maps. Additionally, we introduced the dynamic convolution theory, which could further increase the pixel segmentation accuracy of the network.
The effectiveness of DAFNet was extensively validated for three datasets. The results demonstrated that the proposed method achieved DSC metrics of 0.942 and 0.885, and HD metricd of 2.50mm and 3.79mm on the first and second dataset, respectively. The recognition accuracy of left ventricular end-diastolic diameter recognition on the third dataset was 98.42%.
Compared with the existing CNNs-based methods, the DAFNet achieved state-of-the-art segmentation performance and verified its effectiveness in clinical diagnosis.
Citation: Yuxin Luo, Yu Fang, Guofei Zeng, Yibin Lu, Li Du, Lisha Nie, Pu-Yeh Wu, Dechuan Zhang, Longling Fan. DAFNet: A dual attention-guided fuzzy network for cardiac MRI segmentation[J]. AIMS Mathematics, 2024, 9(4): 8814-8833. doi: 10.3934/math.2024429
[1] | S. Neelakandan, Sathishkumar Veerappampalayam Easwaramoorthy, A. Chinnasamy, Jaehyuk Cho . Fuzzy adaptive learning control network (FALCN) for image clustering and content-based image retrieval on noisy dataset. AIMS Mathematics, 2023, 8(8): 18314-18338. doi: 10.3934/math.2023931 |
[2] | Eman A. Al-Shahari, Marwa Obayya, Faiz Abdullah Alotaibi, Safa Alsafari, Ahmed S. Salama, Mohammed Assiri . Accelerating biomedical image segmentation using equilibrium optimization with a deep learning approach. AIMS Mathematics, 2024, 9(3): 5905-5924. doi: 10.3934/math.2024288 |
[3] | Yan Ma, Defeng Kong . Super-resolution reconstruction algorithm for dim and blurred traffic sign images in complex environments. AIMS Mathematics, 2024, 9(6): 14525-14548. doi: 10.3934/math.2024706 |
[4] | Yongyan Zhao, Jian Li . Integrating artificial intelligence with network evolution theory for community behavior prediction in dynamic complex systems. AIMS Mathematics, 2025, 10(2): 2042-2063. doi: 10.3934/math.2025096 |
[5] | Xia Chang, Haixia Zhao, Zhenxia Xue . MRI image enhancement based on feature clustering in the NSCT domain. AIMS Mathematics, 2022, 7(8): 15633-15658. doi: 10.3934/math.2022856 |
[6] | Ming Shi, Ibrar Hussain . Improved region-based active contour segmentation through divergence and convolution techniques. AIMS Mathematics, 2025, 10(1): 654-671. doi: 10.3934/math.2025029 |
[7] | Wen Sun . Fuzzy knowledge spaces based on $ \beta $ evaluation criteria. AIMS Mathematics, 2023, 8(11): 26840-26862. doi: 10.3934/math.20231374 |
[8] | Yuzi Jin, Soobin Kwak, Seokjun Ham, Junseok Kim . A fast and efficient numerical algorithm for image segmentation and denoising. AIMS Mathematics, 2024, 9(2): 5015-5027. doi: 10.3934/math.2024243 |
[9] | Alaa O. Khadidos . Advancements in remote sensing: Harnessing the power of artificial intelligence for scene image classification. AIMS Mathematics, 2024, 9(4): 10235-10254. doi: 10.3934/math.2024500 |
[10] | Pablo Fernández-López, Patricio García Báez, Ylermi Cabrera-León, Aleš Procházka, Carmen Paz Suárez-Araujo . Modeling the implications of nitric oxide dynamics on information transmission: An automata networks approach. AIMS Mathematics, 2023, 8(12): 30142-30181. doi: 10.3934/math.20231541 |
In clinical diagnostics, magnetic resonance imaging (MRI) technology plays a crucial role in the recognition of cardiac regions, serving as a pivotal tool to assist physicians in diagnosing cardiac diseases. Despite the notable success of convolutional neural networks (CNNs) in cardiac MRI segmentation, it remains a challenge to use existing CNNs-based methods to deal with fuzzy information in cardiac MRI. Therefore, we proposed a novel network architecture named DAFNet to comprehensively address these challenges.
The proposed method was used to design a fuzzy convolutional module, which could improve the feature extraction performance of the network by utilizing fuzzy information that was easily ignored in medical images while retaining the advantage of attention mechanism. Then, a multi-scale feature refinement structure was designed in the decoder portion to solve the problem that the decoder structure of the existing network had poor results in obtaining the final segmentation mask. This structure further improved the performance of the network by aggregating segmentation results from multi-scale feature maps. Additionally, we introduced the dynamic convolution theory, which could further increase the pixel segmentation accuracy of the network.
The effectiveness of DAFNet was extensively validated for three datasets. The results demonstrated that the proposed method achieved DSC metrics of 0.942 and 0.885, and HD metricd of 2.50mm and 3.79mm on the first and second dataset, respectively. The recognition accuracy of left ventricular end-diastolic diameter recognition on the third dataset was 98.42%.
Compared with the existing CNNs-based methods, the DAFNet achieved state-of-the-art segmentation performance and verified its effectiveness in clinical diagnosis.
Cardiac and vascular disease (CVD) poses a significant threat to human life with high rates of morbidity and mortality. According to the World Health Organization (WHO) statistics, there are over 500 million global cases of CVD, resulting in approximately 17.9 million deaths annually, making it the leading cause of mortality worldwide [1]. Current clinical diagnosis methods for CVD include the ultrasound (US), computed tomography (CT), and magnetic resonance imaging (MRI). The MRI has become a crucial tool for physicians due to its ability to capture characteristics of original 3D cross-sectional medical images without reconstruction [2]. However, in clinical practice, the manual identification of cardiac regions on MRI images introduces a risk of high subjective variation and poor reproducibility. Therefore, there is a pressing need for accurate, rapid, and batch cardiac segmentation methods, which hold widespread application and significant research importance for physicians in diagnosing cardiovascular diseases [3].
While traditional image processing methods yield satisfactory results for cardiac segmentation in clear images [4,5], they exhibit limitations when confronted with the multi-temporal, multi-modal and low-quality images. These shortcomings hinder their performance and fail to assist doctors effectively in achieving accurate diagnostic outcomes. Additionally, these methods require physician involvement, leading to a potential waste of medical resources and inefficiencies in disease diagnosis.
Recent advances in artificial intelligence technology have enabled scholars to apply deep learning to segmentation and detection tasks in medical images. Tran et al. [6] were among the pioneers to utilize the fully convolutional network (FCN [7]) for medical images segmentation, demonstrating that deep learning methods outperform traditional segmentation methods in terms of accuracy and speed. Khened et al. [8] introduced a long-skip and short-cut connection structure, which is efficient. However, medical images are characterized by sparse samples, low information density, and simple semantic content compared to the natural scene images, leading to overfitting problems when solely relying on FCN for cardiac segmentation. This limitation prevents network from obtaining more accurate performance. Ronneberger et al. [9] addressed this challenge by introducing a U-shaped structure known as U-Net. However, the simplicity of U-Net structure compromises its performance in extracting object features, as illustrated in Figure 1. Subsequent research has witnessed scholars enhancing and integrating U-Net, striving for performance improvement. Painchaud et al. [10] designed a variational self-encoder structure in the post-processing stage to rectify invalid cardiac shapes, achieving higher anatomical validity. Cheng et al. [11] developed a feature rectification and fusion module based on orientation fields to address issue of inter-class indistinguishability and intra-class inconsistency. Tong et al. [12] proposed a loop feedback architecture to enhance features and improve the features extraction performance of the network. Wang et al. [13] proposed a framework that combines semi-supervised learning and self-training to be able to train the network better using fewer samples. Gao et al. [14] propose a powerful hybrid Transformer architecture that integrates self-attention into a CNNs. Rahman et al. [15] introduce a diffusion model that produces multiple plausible outputs by learning a distribution of group insights.
Despite these improvements, the limited deterministic information in the medical images can negatively affect the features extraction performance of neural networks. The integration of attention mechanisms is the key to solving this problem. It can reduce the interference of irrelevant information and enable network to focus on the limited deterministic information. Schlemper et al. [16] designed a spatial attention gate on skip connection structure of U-Net, which weighted up-sampling results of the decoder by calculating the spatial attention of output feature maps from encoder, achieving better performance than U-Net. Hu et al. [17] proposed a squeeze-and-excitation network (SE-Net), which assigned the attention weight to feature maps in channel dimension, allowing network to ignore the irrelevant information. Woo et al. [18] designed a convolutional block attention module (CBAM) by combining channel and spatial attention to improve the segmentation performance. Although, these attention mechanisms can reduce the negative impact of irrelevant information, they also lead to the network disregarding the significance of fuzzy information in medical images. Fuzziness is a condition in which the boundary pixels cannot be segmented accurately due to the weak grayscale variation in medical images. Therefore, the pixels of each tissue boundary may contain fuzzy information. The fuzzy information in cardiac MRI can cause neural networks to extract incorrect features of the left ventricle, right ventricle, and myocardium, leading to poor segmentation results. This shortcoming obstacles the network's ability to sufficiently utilize the large amount of valuable fuzzy information, thus adversely affecting the overall performance of the network.
Therefore, in this study, we proposed a dual attention-guided fuzzy network called the DAFNet, and aimed to improve overall segmentation performance. A fuzzy convolution module (FCM) that combines spatial attention and channel attention is designed. FCM allowes the network to significantly utilize fuzzy information while retaining the advantages of attention mechanisms to obtain more accurate and informative object features. Then, we designe a multi-scale feature refinement structure (MSFR) in the decoder to aggregate segmentation maps on different decoding features. MSFR enables the network to achieve more accurate segmentation results. Moreover, a pixel segmentation prediction structure is designed to further enhance pixel segmentation performance of the network according to the dynamic convolutional theory [19]. We also provide a patient cardiac MRI dataset provided by a hospital.
We evaluated the segmentation performance of DAFNet using three cardiac MRI datasets. The dataset 1 (ACDC [20]) is an open-source and fully labeled dataset, comprising 150 patient scans each featuring a short-axis MRI acquired on a 1.5T Siemens Area and a 3.0T Siemens MR instruments. In data acquisition of dataset 1, only two-chamber scans of cardiac on conventional axial were included. Dataset 2 (M&Ms-1 [21]) contains MR data of 375 patients from six clinical centers in Spain, Canada, and Germany, acquired under four scanner vendors (1.5T Siemens, 3.0T Siemens, 1.5T Philips, 1.5T General Electric, and 1.5T Canon). The dataset 3 was collected by the Chongqing Traditional Chinese Medicine Hospital, consisted of 12 patient scans, each including a short-axis MRI acquired on the GE Healthcare SIGNA Architect 3.0T superconducting MR instrument. In data acquisition of dataset 3, two-chamber, three-chamber, and four-chamber scans of cardiac on sagittal, coronary, conventional axial were included. The basic information of cardiac MRI image on the three datasets is shown in Table 1. Since the resolution of these MRI images is not fixed, we uniformly resize them to 212×212 pixels during experiments.
Dataset | Frames | Train | Test | Resolution | Ground Truth |
dataset 1 | 2,979 | 1,902 | 1,076 | ⩽256×256 pixels | Background, LV, RV, Myo |
dataset 2 | 3,264 | 1,643 | 1,621 | ⩽256×256 pixels | Background, LV, RV, Myo |
dataset 3 | 4,649 | 0 | 4,649 | ⩽512×512 pixels | LV end-diastolic inner diameter |
Representative slices of these three cardiac MRI datasets are shown in Figure 2, from which it can be seen that dataset 3 has a higher level of noise and artifacts, and the overall image quality is not as good as that of dataset 1 and dataset 2. In addition, there is a phenomenon of insignificant difference in the grayscale of some of the image pixels between dataset 2 and dataset 3, which is more pronounced in dataset 3. These differences allow us to make a more comprehensive assessment of the overall performance of DAFNet.
We aimed to achieve accurate segmentation of the LV, RV and Myo categories based on MRI. Figure 3 shows that the proposed DAFNet was developed based on the U-Net [9] architecture, which consisted of encoder and decoder. Noticeably, we are not limited to U-Net architecture, these proposed modules can be migrated to any network architecture, but as a general rule, we improve and introduce on the U-Net architecture.
Specifically, encoder consists of two convolutional layers (Conv) with a kernel size of 3×3 and four FCM, and outputs five different scale feature F1,F2,F3,F4,F5. FCM comprised of a max-pooling layer (MaxPool), a spatial fuzzy convolutional layer (SFConv), and a channel fuzzy convolutional layer (CFConv). The decoder adopts the designed MSFR structure, takes the output feature maps of the encoder as input, and outputs four multi-scale feature maps O1,O2,O3,O4. The segmentation map is acquired through our designed dynamic pixels segmentation convolution (dyConv) structure. During training phase, the training loss of DAFNet was calculated as the sum of the losses of four segmentation maps and the label. In testing phase, the final cardiac segmentation results are obtained by aggregating the four segmentation maps in the MSFR structure.
The attention mechanisms (Attention) computes the attention weight of a feature map in spatial or channel dimension, aiming to enhance the positive influence of deterministic information and to mitigate the negative effects of the irrelevant information on object features extraction [18]. This directs CNNs-based methods toward the deterministic information, thereby improving the overall performance. However, medical images not only contain the limited directly usable deterministic information but also include the substantial amount of overlooked yet valuable fuzzy information. The introduction of the attentional mechanism makes networks ignore the fuzzy information because it is provided with a low attentional weight, which hinders the networks' performance. Addressing the challenge posed by the abundance of fuzzy information in cardiac MRI images, we proposed FCM that not only retained the benefits of attention mechanisms but also effectively leveraged fuzzy information to improve the network's performance. The FCM contained two convolutional layer structures: Spatial and channel fuzzy convolutional layer.
The first convolutional layer structure, a spatial fuzzy convolution, is illustrated in Figure 4. This structure enabling the network to learn and infer the fuzzy information from the spatial dimension. The forward inference process of SFConv is outlined as follows
Fi+1=Conv1×1((1−u)×Conv3×3(Fi)+u×Fi), | (1) |
where i∈[1,4], u is the fuzzification of each pixel in spatial dimension, and 1-u is the determinacy of each pixel. Similarly, the feature map Fi+1 and Fi in the encoder can also be replaced by the feature map Oi and Oi+1 in the decoder.
We utilized information entropy theory to compute the fuzzification of each feature. When the probability of a feature belonging to a particular category is higher, the information entropy is smaller, approaching 0, indicating higher deterministic. Conversely, when the probability of a feature belonging to some categories is similar, the information entropy increases, approaching 1, indicating higher fuzzification. Thus, the two branches in Eq (1) can simultaneously retain the deterministic and fuzzy information of the convolutional feature map.
The fuzzification in the spatial dimension is calculated as follows
μ=softmax(Conv1×1(Fi)), | (2) |
u=−1NN∑j=1μj⋅log2(μj), | (3) |
where μi represents the probability that each feature of Fi belongs to category j in the spatial dimension, and N is the categories number to be detected.
The second convolutional layer structure, a channel fuzzy convolution, is show in Figure 5. This structure enabling network to learn and infer the fuzzy information from the channel dimension. The forward inference process of CFConv is outlined as follows
Fi+1=Conv1×1((1−v)×Conv3×3(Fi)+v×Fi), | (4) |
where v is the fuzzification of each channel in the channel dimension, and 1-v is the determinacy of each channel. The fuzzification in the channel dimension is calculated as follows
η=softmax(Conv1×1(Fi)), | (5) |
v=−1NN∑j=1ηj⋅log2(ηj), | (6) |
where ηi represents the probability that each feature of Fi belongs to category j in the channel dimension.
In the convolutional neural networks, deep features have high-level semantics and generate strong responses to the detected object. While shallow features contain a large amount of local information and can provide a rich context to accurately identify object contour boundaries. allowing the network to access deep features, high-level semantics can be exploited to accurately extract object features. Allowing the network to access shallow features and rich context can be exploited to further refine the detected object contour boundaries [22]. However, the existing networks only focus on deep features and ignore the improvements of shallow features. To address this issue, we designed a MSFR structure in the decoder portion of the DAFNet, as shown in Figure 6a. This structure allowed the network to simultaneously access deep and shallow features, thereby significantly enhancing the performance of the network. The forward inference process of MSFR is outlined as follows
yseg=fseg1(O1)+fseg2(fup(O2))+fseg3(fup(O3))+fseg4(fup(O4)), | (7) |
where yseg is the final segmentation result, fsegi(⋅) is the dyConv structure we design for predicting each segmentation map, fup(⋅) is the bilinear interpolation method. The final pixel segmentation map yseg is aggregated from four predicted pixel segmentation maps fsegi(i∈[1,4]) in channel dimension by adding operation. The final mask is obtained by the argmax function on the final pixel segmentation map yseg.
Since using the Conv as a pixel segmentation convolutional structure will reduce the network performance, we designed a higher-performance dyConv structure based on the dynamic convolution theory. It calculates the attention of the feature map in channel or spatial dimension first, and then dynamically adjusts the convolution kernel and kernel parameters according to the attention. Then, a more accurate decoded feature map is extracted. The dyConv structure is shown in Figure 6b, and its forward inference process is outlined as follows
segi=fsegi(Oi)=h(g1×g2×Wi,Oi), | (8) |
g1=sigmoid(mlp1(GAvgPool(X)+GMaxPool(X))), | (9) |
g2=softmax(mlp2(GAvgPool(X)+GMaxPool(X))), | (10) |
where mlp(·) is a multi-layer perceptron, and Wi is the kernel coefficients matrix of convolutional layer. g1 is the calculated attention weight of the kernel coefficient matrix in spatial dimension, and g2 is the calculated attention weight of the kernel in channel dimension. h(·) is the forward inference process of the group convolutional layer.
We utilized a composite loss function by merging the Dice loss and the cross-entropy loss for the training of DAFNet. The Dice loss is employed to evaluate the level of overlap between the segmented maps and the labeled image. Moreover, the cross-entropy loss is utilized to evaluate the accuracy of pixel-wise classification. The computation of the composite loss is outlined as follows
Ltotal=Ldice+α×Lce, | (11) |
Ldice=1−12b4∑i=1b∑j=1N∑k=1(pijk×gjk)/(pijk+gjk), | (12) |
Lce=−14b4∑i=1b∑j=1N∑k=1(pijk×log2(gjk)+(1−pijk)×log2(1−gjk)), | (13) |
where Ldice and Lce represents the computed Dice loss and cross-entropy loss. Symbol α denotes the weight assigned for balancing each loss, which was set at 0.5. pijk and gjk is the predicted segmentation maps and corresponding label.
The experiments were carried out on Ubuntu 21.04 system using PyTorch 1.7.1. We incorporated random rotation, diagonal mirroring, and vertical mirroring for image data augmentation during training. Additionally, the input image was resized to 212×212 pixels. The batch size coefficient b was set to 16. Since the LV, RV, Myo and background pixels were segmented from the cardiac MRI in this study, the number of category N that the network needed to predict was set to 4. During the training of DAFNet on dataset 1, the Adam optimizer with a weight decay coefficient of 0.0001 and a cosine annealing decay strategy with a learning rate of 0.001 were employed. The training epoch was set to 30. Since neither dataset contains a validation set, the proposed method is trained without employing validation loss and early stopping. The detailed overview of the proposed DAFNet structure parameter settings is shown in Figure 7.
The Dice similarity coefficient (DSC) and Hausdorff distance (HD) metric are sensitive to the segmentation regions and boundarys. These metrics gauge similarity and distance by evaluating the correspondence between segmentation maps and their corresponding labels. The DSC and HD metrics are outlined as follows
DSC(Pi,Ti)=2×|Pi∩Ti|/(|Pi|+|Ti|)×100%, | (14) |
HD(Pi,Ti)=max(maxn∈Pi(minm∈Ti‖n−m‖),maxm∈Ti(minn∈Pi‖m−n‖)), | (15) |
where P represents the set of pixels in segmented map, while Ti is the set of pixels in corresponding label image. i signifies the category of LV, RV or Myo. |⋅| is utilized to obtain the number of elements in the set, ||⋅|| is utilized for calculating the distance between points in two sets.
The LV diastolic inner diameter typically falls within range of 45−55mm for males and 35−50mm for females. Abnormal increases in LV end-diastolic inner diameter (LVd) can potentially lead to cardiac failure. In clinical practice, the deviation from indicators such as LVd and normal values is commonly used to evaluate the cardiac function. For a more precise evaluation, we used accuracy to quantify the concordance between predicted LVd and the label. The accuracy metric is outlined as follows
accuracy=1−|TLVd−PLVd|/TLVd, | (16) |
where PLVd is the LVd computed from predicted segmentation map, and TLVd represents the LVd measured by experts.
For dataset 1, we recorded the qualitative results of the Attention, FCM, MSFR, and dyConv to verify the effectiveness of these proposed modules, as shown in Table 2. Since dataset 1 provided only the LV, RV, Myo, and background labels, we use the DSC and HD as metrics to evaluate the performance of methods. The results of row 1 show the cardiac segmentation performance of the baseline. Row 2 shows the performance after adding the FCM. Row 3 shows the performance after adding the FCM and MSFR. Row 4 shows the performance of the proposed DAFNet. The p-values in rows 2−5 indicate whether the baseline uses attention, or FCM, the methods of the third row and fourth row, and the method of the fourth row and fifth row, respectively. We found that the FCM, which enables the network to learn and utilize fuzzy information, MSFR, which aggregates multi-scale segmentation results, and dyConv, which improves pixel segmentation accuracy, can improve the segmentation performance of the network, thereby verifying the effectiveness of these proposed modules.
Baseline | Attention | FCM | MSFR | dyConv | mean DSC(%) | mean HD(mm) | p-values | FLOPs(G) |
√ | 89.23 | 4.94 | − | 7.85 | ||||
√ | √ | 90.86(+1.63) | 3.82(−1.12) | 0.2698 | 7.86 | |||
√ | √ | 92.28(+3.05) | 3.22(−1.72) | 0.0010 | 11.64 | |||
√ | √ | √ | 93.19(+3.96) | 2.73(−2.21) | 0.1045 | 11.64 | ||
√ | √ | √ | √ | 94.42(+4.94) | 2.50(−2.44) | 0.0340 | 11.73 |
In medical images, the difference in gray-scale values between the different categories contour boundaries pixel are exceedingly subtle, which brings a challenge to cardiac segmentation. Figure 8 visualizes segmentation results after progressively adding the FCM, MSFR, and dyConv to the baseline. The first and last columns are the segmentation results of the baseline and DAFNet respectively. The results demonstrated that the RV segmentation accuracy of the baseline is very low, while DAFNet achieve high segmentation accuracy. Therefore, it is verified that the segmentation performance is greatly improved after integrating these modules.
To further validate the effectiveness of FCM, we investigated the impact of different FCM structures on segmentation performance. Table 3 shows the segmentation performance of DAFNet based on different FCM structures. It can be found that both SFConv and CFConv effectively enhance the performance compared to the convolutional layer, and the combination of the last row has the best results. Therefore, this combination is used as the proposed FCM structure in this article.
Structure | DSC(%) | HD(mm) | p-values | ||||||
LV | RV | Myo | mean | LV | RV | Myo | mean | ||
Conv+Conv | 89.31 | 89.45 | 94.07 | 90.94 | 5.61 | 3.00 | 2.53 | 3.71 | − |
Conv+CFConv | 89.76 | 89.51 | 94.21 | 91.16 | 5.66 | 3.03 | 2.47 | 3.72 | 0.7831 |
CFConv+Conv | 90.16 | 89.35 | 93.77 | 91.09 | 5.15 | 3.04 | 2.53 | 3.57 | 0.8480 |
CFConv+CFConv | 89.38 | 89.97 | 93.79 | 91.05 | 6.42 | 3.34 | 2.84 | 4.20 | 0.5628 |
Conv+SFConv | 90.33 | 89.55 | 94.00 | 91.29 | 5.16 | 2.79 | 2.38 | 3.44 | 0.7843 |
SFConv+Conv | 90.15 | 89.29 | 94.08 | 91.17 | 5.27 | 2.98 | 2.19 | 3.48 | 0.5673 |
SFConv+SFConv | 88.48 | 89.06 | 94.06 | 90.53 | 5.96 | 2.93 | 2.39 | 3.76 | 0.3872 |
CFConv+SFConv | 93.95 | 90.49 | 94.29 | 92.91 | 3.90 | 2.79 | 2.36 | 3.02 | 0.0173 |
SFConv+CFConv | 95.23 | 91.53 | 95.76 | 94.17 | 3.20 | 2.34 | 1.96 | 2.50 | < 0.0001 |
In addition, to further illustrate that the FCM cannot only enable the network to obtain better encoding features, but also enable to obtain better object features in the decoding stage, we conducted relevant experiments. As shown in Table 4, the impact of the FCM on the segmentation performance during the encoding and decoding stages demonstrates that using FCM in both encoding and decoding stages can achieve the best performance improvement.
Encoder | Decoder | DSC(%) | HD(mm) | p-values | ||||||
LV | RV | Myo | mean | LV | RV | Myo | mean | |||
89.31 | 89.45 | 94.07 | 90.94 | 5.61 | 3.00 | 2.53 | 3.71 | − | ||
√ | 93.87 | 90.28 | 94.88 | 93.01 | 3.75 | 2.69 | 2.20 | 2.88 | 0.0021 | |
√ | 92.17 | 88.89 | 94.13 | 91.73 | 6.07 | 3.73 | 3.18 | 4.33 | 0.2526 | |
√ | √ | 95.23 | 91.53 | 95.76 | 94.17 | 3.20 | 2.34 | 1.96 | 2.50 | 0.0128 |
Figure 9 depicts the Dice loss and DSC changes of the baseline and DAFNet on dataset 1. The results showed that under the same experiment environment, DAFNet not only accelerates the convergence rate of the Dice loss, but also reduces the final convergence loss. Moreover, DAFNet achieves higher DSC faster than the baseline. The Dice loss and DSC in Figure 9 are the average of five additional experiments, as shown in Table 5.
Method | Numbers | 1 | 2 | 3 | 4 | 5 | mean | σ2 | |
Baseline | Dice loss | Train | 0.0574 | 0.0537 | 0.0609 | 0.0549 | 0.0526 | 0.0559 | 8.7960×10-6 |
Test | 0.0848 | 0.0834 | 0.0945 | 0.0799 | 0.0834 | 0.0852 | 2.4244×10-7 | ||
DSC(%) | Train | 93.02 | 93.61 | 92.58 | 93.42 | 93.73 | 93.27 | 0.1777 | |
Test | 89.39 | 89.66 | 88.12 | 90.09 | 89.68 | 89.39 | 0.4520 | ||
DAFNet | Dice loss | Train | 0.0468 | 0.0459 | 0.0457 | 0.0454 | 0.0449 | 0.0457 | 3.9600×10-7 |
Test | 0.0505 | 0.0500 | 0.0502 | 0.0481 | 0.0489 | 0.0495 | 8.1200×10-7 | ||
DSC(%) | Train | 94.53 | 94.68 | 94.68 | 94.72 | 94.77 | 94.68 | 0.0064 | |
Test | 94.05 | 94.13 | 94.09 | 94.36 | 94.24 | 94.17 | 0.0127 |
Table 6 records the performance of different methods on dataset 1. The proposed DAFNet obtained the best mean DSC and HD results. However, the results in Table 6 show that the right ventricle segmentation performance is not satisfactory. This is due to the complex shape, ill-defined thin edges, large variations among patients, and pathology of the right ventricle [24]. The results in lines 16-17 of Table 6 show that DAFNet can further improve the performance of other methods. Figure 10 shows representative segmentation results of different methods on dataset 1, indicating that DAFNet outperformed other methods in terms of the segmentation and obtained contour boundaries in various parts. Under the right ventricle, the results of existing methods are not accurate enough. Under the left ventricle and myocardium, the proposed method obtains better segmentation results, verifying that utilizing fuzzy information can solve the challenge of boundary pixels that cannot be accurately segmented due to fuzziness. The reduction in HD indirectly supported the effectiveness of the proposed network. Compared with other methods, the proposed DAFNet basically obtains the highest DSC metric and the lowest HD metric in the three categories. The DAFNet-Res18 architecture means adding these proposed modules in the ResNet-18 [25] architecture instead of the U-Net architecture described in Section 2.2. Table 7 and Figure 11 show the results of different methods on dataset 2.
Methods | LV | RV | Myo | mean | ||||||||||
DSC(%) | HD(mm) | DSC(%) | HD(mm) | DSC(%) | HD(mm) | DSC(%) | HD(mm) | |||||||
ED | ES | ED | ES | ED | ES | ED | ES | ED | ES | ED | ES | |||
Reference [9] | 90.38 | 85.61 | 6.74 | 6.93 | 87.27 | 86.78 | 3.88 | 5.06 | 95.57 | 89.80 | 3.23 | 3.81 | 90.94 | 3.71 |
Reference [12] | 96.60 | 91.80 | 7.95 | 8.18 | 94.80 | 89.80 | 11.67 | 12.84 | 90.50 | 91.50 | 9.50 | 9.57 | 92.50 | 9.95 |
Reference [16] | 96.13 | 92.37 | 3.62 | 4.37 | 90.14 | 90.65 | 2.48 | 3.03 | 97.27 | 93.36 | 1.98 | 2.67 | 93.32 | 3.03 |
Reference [14] | 96.31 | 92.84 | 3.13 | 3.97 | 90.65 | 90.93 | 2.76 | 3.96 | 96.79 | 93.23 | 1.88 | 2.55 | 93.46 | 3.04 |
Reference [15] | 95.59 | 91.94 | 3.82 | 4.33 | 90.03 | 90.43 | 3.04 | 4.03 | 96.37 | 92.67 | 2.38 | 3.14 | 93.36 | 3.46 |
Reference [26] | 96.70 | 92.80 | 6.40 | 7.60 | 93.60 | 88.90 | 13.30 | 14.40 | 89.10 | 90.40 | 8.30 | 9.60 | 91.92 | 9.93 |
Reference [27] | 96.10 | 91.80 | 7.50 | 9.60 | 92.80 | 87.20 | 11.90 | 13.40 | 87.50 | 89.40 | 11.10 | 10.70 | 90.80 | 10.70 |
Reference [28] | 96.30 | 91.10 | 6.50 | 9.20 | 93.20 | 88.30 | 12.70 | 14.70 | 89.20 | 90.10 | 8.70 | 10.60 | 91.37 | 10.40 |
Reference [29] | 96.70 | 92.80 | 5.50 | 6.90 | 94.60 | 90.40 | 8.80 | 11.40 | 89.60 | 91.90 | 7.60 | 7.10 | 92.67 | 7.88 |
DAFNet-Ref. [18] | 96.01 | 92.57 | 3.31 | 4.38 | 90.38 | 90.43 | 2.33 | 3.10 | 97.17 | 93.20 | 1.92 | 2.58 | 93.29 | 2.94 |
DAFNet-Ref. [19] | 96.42 | 92.66 | 3.43 | 4.03 | 92.04 | 92.34 | 2.18 | 2.89 | 97.52 | 94.15 | 1.87 | 2.43 | 94.18 | 2.81 |
DAFNet-Res18(Ours) | 96.53 | 92.78 | 3.01 | 3.89 | 92.12 | 90.89 | 2.90 | 2.84 | 96.89 | 91.99 | 2.10 | 3.62 | 93.53 | 3.06 |
DAFNet-Res34(Ours) | 96.41 | 92.89 | 2.96 | 3.77 | 91.97 | 91.03 | 2.99 | 2.65 | 97.13 | 92.78 | 2.03 | 3.03 | 93.70 | 2.91 |
DAFNet(Ours) | 96.72 | 93.74 | 2.78 | 3.61 | 91.68 | 91.38 | 2.09 | 2.59 | 97.53 | 93.99 | 1.70 | 2.22 | 94.17 | 2.50 |
Method | LV | RV | Myo | mean | ||||
DSC (%) | HD (mm) | DSC (%) | HD (mm) | DSC (%) | HD (mm) | DSC (%) | HD (mm) | |
Reference [9] | 87.69 | 6.93 | 81.13 | 4.57 | 85.63 | 3.64 | 84.82 | 5.05 |
Reference [12] | 89.21 | 5.68 | 83.68 | 4.28 | 87.25 | 3.29 | 86.71 | 4.42 |
Reference [16] | 90.94 | 5.34 | 85.59 | 3.79 | 87.74 | 3.12 | 88.09 | 4.08 |
Reference [30] | 91.25 | 9.10 | 85.30 | 11.70 | 88.50 | 12.25 | 88.35 | 11.03 |
Reference [31] | 90.90 | 9.40 | 84.55 | 11.85 | 87.95 | 12.65 | 87.80 | 11.30 |
Reference [32] | 90.50 | 10.00 | 84.10 | 12.45 | 87.50 | 12.65 | 87.37 | 11.70 |
DAFNet-Res18 | 92.34 | 5.23 | 84.28 | 3.24 | 87.23 | 3.23 | 87.95 | 3.90 |
DAFNet-Res34 | 91.23 | 5.63 | 84.47 | 3.68 | 87.84 | 2.67 | 87.85 | 3.99 |
DAFNet (Ours) | 92.58 | 4.87 | 84.17 | 3.97 | 88.68 | 2.54 | 88.47 | 3.79 |
Additionally, in clinical diagnosis, considering the different cardiac image acquisition methods of different instruments, different noise and artifacts will be introduced. Achieving robust segmentation across datasets is therefore critical to assist physicians in diagnosing cardiac disease. For dataset 3, since it provided only the Lvd label, we use the accuracy as a metric. We evaluate their performance using different methods trained on dataset 1 to predict LVd, as shown in Table 8. We found that the LVd obtained by DAFNet are in good agreement with the results manually measured by experts, which further confirms the effectiveness of DAFNet. Figure 12 shows the segmentation results of different methods on dataset 3. Under the left ventricle and myocardium, the other methods appear to under-segment or over-segment, and the proposed method is able to obtain better results by utilizing fuzzy information. Under the right ventricle, only eference [12] can barely segment, which is due to the fact that right ventricle segmentation is usually a difficult problem in cardiac MRI segmentation [24].
Methods | Patients | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | mean | σ2 |
TLVd | 44.00 | 47.00 | 58.00 | 42.00 | 51.00 | 54.00 | 68.00 | 44.00 | 45.00 | 47.00 | 61.00 | 38.00 | — | — | |
Reference [9] | PLVd | 41.72 | 43.78 | 53.33 | 40.77 | 54.34 | 51.81 | 65.93 | 46.32 | 43.65 | 44.14 | 59.74 | 39.84 | — | — |
accuracy | 94.82 | 93.15 | 91.95 | 97.07 | 93.45 | 95.94 | 96.96 | 94.73 | 97.00 | 93.91 | 97.93 | 95.16 | 95.17 | 1.77 | |
Reference [12] | PLVd | 42.32 | 45.37 | 57.88 | 41.30 | 50.80 | 52.72 | 66.94 | 45.22 | 43.79 | 45.44 | 59.26 | 39.06 | — | — |
accuracy | 96.18 | 96.53 | 99.79 | 98.33 | 99.61 | 97.63 | 98.44 | 97.23 | 97.31 | 96.68 | 97.15 | 97.21 | 97.67 | 1.22 | |
Reference [16] | PLVd | 42.30 | 45.18 | 56.97 | 40.85 | 50.55 | 53.30 | 66.66 | 45.64 | 43.34 | 45.13 | 59.79 | 39.01 | — | — |
accuracy | 96.14 | 96.13 | 98.22 | 97.26 | 99.12 | 98.70 | 98.03 | 96.27 | 96.31 | 96.02 | 98.02 | 97.34 | 97.30 | 1.13 | |
DAFNet-Res18 | PLVd | 44.38 | 48.25 | 56.88 | 41.78 | 51.21 | 54.79 | 67.32 | 45.39 | 44.13 | 45.39 | 61.79 | 38.67 | — | — |
accuracy | 99.14 | 97.34 | 98.07 | 99.48 | 99.59 | 98.54 | 99.00 | 96.84 | 98.07 | 96.57 | 98.70 | 98.24 | 98.30 | 0.94 | |
DAFNet-Res34 | PLVd | 45.79 | 48.17 | 57.03 | 41.42 | 51.43 | 55.67 | 68.77 | 43.75 | 45.76 | 46.21 | 62.68 | 38.94 | — | — |
accuracy | 95.93 | 97.51 | 98.33 | 98.62 | 99.16 | 96.91 | 98.87 | 99.43 | 98.31 | 98.32 | 97.25 | 97.53 | 98.01 | 0.97 | |
DAFNet(Ours) | PLVd | 45.33 | 48.07 | 57.94 | 42.51 | 50.71 | 55.26 | 68.59 | 42.14 | 45.88 | 46.83 | 61.69 | 38.26 | — | — |
accuracy | 96.98 | 97.72 | 99.90 | 98.79 | 99.43 | 97.67 | 99.13 | 95.78 | 98.04 | 99.64 | 98.87 | 99.32 | 98.44 | 1.18 |
In deep learning, the segmentation performance improvement can be achieved through four avenues. The excellent pre-processing pipeline [33] highlights feature through data augmentation methods. The effective post-processing pipeline [34] optimizes the prediction results through clustering algorithms. Excellent neural networks [12,16] extract precise features by improving network structure. Superior training skill [35] enable the network to fitting object better by fine-tuning. These avenues can complement each other to further enhance the performance.
In this study, we introduced the DAFNet method for cardiac segmentation by improving the networks structure to improve the segmentation performance, which consist of two pivotal components: The FCM and MSFR structure. The FCM excels in extracting more nuanced and informative features during both the encoding and decoding processes, while the MSFR provides different level of feature information during the decoding, thereby optimizing the final segmentation result. Both components synergistically enhance the overall performance and robustness of the network.
In ACDC and M&Ms-1 competition, due to the different instruments collected, there are image quality inconsistencies, artifacts, noise, and other problems between the data, which make it difficult to segment the organization in the image. To address this problem, many researchers have aimed to enhance U-Net to improve segmentation accuracy. Lossau et al. [36] propose a fully automatic dynamic pacemaker artifact reduction pipeline, which is built from three CNNs ensembles. Tong et al. [12] utilized the staggered attention module to effectively fuse multi-level contextual information, regulating the information transmitted to the decoding stage. Schlemper et al. [16] introduced an attention gate structure in the skip-connection of the network, enabling the network to learn crucial information from the original image and suppress unimportant regions. However, their methods involved only the simple insertion of attention mechanisms, limiting the segmentation performance of U-Net. As shown in Figures 10 and 11, despite the slight improvement in accuracy, the segmentation results of the boundaries region are not optimized. Tables 6 and 7 demonstrate that DAFNet outperformed other methods, achieving superior results in HD and DSC metric. In clinical datasets, segmentation of various cardiac tissues is a difficult task due to the older instruments used for acquisition and slightly poorer image quality. The excellent feature extraction structure of the DAFNet greatly contributes to the performance. In Table 8 and Figure 12, although the clinical data did not provide labeled images, the DAFNet outperforms the other methods by comparing it with the LVd in the hospital-collected data with high stability. Even in some of the basal slices, it performs well.
Although the proposed DAFNet can solve the fuzzy information challenge to obtain better results, the method has some limitations. First, in the FCM, the deterministic and fuzzy information is obtained by matrix multiplication, which means that the network computation will be more intensive. Second, DAFNet is a supervised learning method, which means that complete cardiac label need to be provided in order to train network. However, there will be differences in the cardiac images captured by different devices and the labels are costly to produce, which leads to the possibility that the network may be overfitted to a particular dataset during the training process. These factors limit the generalization ability and practical application of the proposed method in real scenarios.
In this study, we proposed a high-performance automatic cardiac segmentation method named DAFNet, which was designed to effectively address the challenges of accurate extracting cardiac feature on different datasets. The proposed fuzzy convolutional module greatly utilizes fuzzy information, which is easily ignored in medical images and deterministic information, is widely valued in attention mechanism, and improves the network's feature extraction performance. Compared with other networks that obtain segmentation results only on the shallowest decoded feature maps, the proposed multi-scale feature refinement structure further improves performance by aggregating segmentation results on different scale decoded feature maps. Extensive experiments for the three cardiac MRI datasets show that DAFNet demonstrate state-of-the-art performance compared to existing CNN-based methods. This achievement establishes robust foundation for applications in cardiac disease diagnosis, treatment planning, and postoperative measurements.
The authors declare that they have not used Artificial Intelligence (AI) tools in the creation of this article.
Authors are grateful to the reviewers and editors for their suggestions and comments to improve the manuscript.
Authors also thank the financial support from National Natural Science Foundation of China (No. 11461037), Yunnan Fundamental Research Projects (No. 202101BE070001-050; 202301AU070184), Chongqing Science and Health Joint Medical Research key Project (No. 2024ZDXM001), and Chongqing Natural Science Foundation project (No. cstc2021jcyj-msxm0727), and Xinglin Scholar Research Premotion Project of Chengdu University of Traditional Chinese Medicine (No. YYZX2022136) and Kunming University of Science and Technology Self discipline Talent Introduction Scientific Research Foundation Project (No. KKZ3202307033).
The authors declare no conflicts of interest in this paper.
Supplementary: Data availability statement
The ACDC dataset (dataset 1) that supports this research is available at: https://acdc.creatis.insa-lyon.fr/description/databases.html. The M&Ms-1 dataset (dataset 2) that supports this research is available at: https://www.ub.edu/mnms. The hospital clinical datasets (dataset 3) generated and analyzed during this research are not publicly available due to the nature of this research; participants did not agree for their data to be shared publicly.
[1] | World health statistics 2022: Monitoring health for the SDGs, Sustainable development goals, Geneva: World Health Organization, 2022. Available from: https://www.who.int/data/gho/publications/world-health-statistics. |
[2] |
A. F. Frangi, W. J. Niessen, M. A. Viergever, Three-dimensional modeling for functional analysis of cardiac images, a review, IEEE T. Med. Imaging, 20 (2001), 2−5. https://doi.org/10.1109/42.906421 doi: 10.1109/42.906421
![]() |
[3] |
S. J. Al'Aref, K. Anchouche, G. Singh, P. J. Slomka, K. K. Kolli, A. Kumar, et al., Clinical applications of machine learning in cardiovascular disease and its relevance to cardiac imaging, Eur. Heart J., 40 (2019), 1975−1986. https://doi.org/10.1093/eurheartj/ehy404 doi: 10.1093/eurheartj/ehy404
![]() |
[4] | G. I. Sanchez-Ortiz, A. Noble, Fuzzy clustering driven anisotropic diffusion: Enhancement and segmentation of cardiac MR images, In: 1998 IEEE Nuclear Science Symposium Conference Record, IEEE Nuclear Science Symposium and Medical Imaging Conference, 1998, 1873−1874. https://doi.org/10.1109/NSSMIC.1998.773901 |
[5] |
N. Paragios. A level set approach for shape-driven segmentation and tracking of the left ventricle, IEEE T. Med. Imaging, 22 (2003), 773−776. https://doi.org/10.1109/TMI.2003.814785 doi: 10.1109/TMI.2003.814785
![]() |
[6] | P. Tran, A fully convolutional neural network for cardiac segmentation in short-axis MRI, arXiv preprint, 2016. Available from: https://arXiv.org/abs/1604.00494. |
[7] |
E. Shelhamer, J. Long, T. Darrell, Fully convolutional networks for semantic segmentation, IEEE T. Pattern Anal., 39 (2017), 3431−3440. https://doi.org/10.1109/TPAMI.2016.2572683 doi: 10.1109/TPAMI.2016.2572683
![]() |
[8] |
M. Khened, V. Kollerathu, G. Krishnamurthi, Fully convolutional multi-scale residual DenseNets for cardiac segmentation and automated cardiac diagnosis using ensemble of classifiers, Med. Image Anal., 51 (2019), 21−45. https://doi.org/10.1016/j.media.2018.10.004 doi: 10.1016/j.media.2018.10.004
![]() |
[9] | O. Ronneberger, P. Fischer, T. Brox, U-net: Convolutional networks for biomedical image segmentation, In: Medical Image Computing and Computer-Assisted Intervention-MICCAI 2015: 18th International Conference, 2015,234−241. https://doi.org/10.1007/978-3-319-24574-4_28 |
[10] |
N. Painchaud, Y. Skandarani, T. Judge, O. Bernard, A. Lalande, P. M. Jodoin, Cardiac segmentation with strong anatomical guarantees, IEEE T. Med. Imaging, 39 (2020), 3703−3713. https://doi.org/10.1109/TMI.2020.3003240 doi: 10.1109/TMI.2020.3003240
![]() |
[11] | F. Cheng, C. Chen, Y. Wang, H. Shi, Y. Cao, D. Tu, et al., Learning directional feature maps for cardiac MRI segmentation, In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2020: 23rd International Conference, 2020,108−117. https://doi.org/10.1007/978-3-030-59719-1_11 |
[12] |
Q. Tong, C. Li, W. Si, X. Liao, Y. Tong, Z. Yuan, et al., RIANet: Recurrent interleaved attention network for cardiac MRI segmentation, Comput. Biol. Med., 109 (2019), 290−302. https://doi.org/10.1016/j.compbiomed.2019.04.042 doi: 10.1016/j.compbiomed.2019.04.042
![]() |
[13] |
W. Wang, Q. Xia, Z. Hu, Z. Yan, Z. Li, Y. Wu, et al., Few-shot learning by a cascaded framework with shape-constrained pseudo label assessment for whole heart segmentation, IEEE T. Med. Imaging, 40 (2021), 2629−2641. https://doi.org/10.1109/TMI.2021.3053008 doi: 10.1109/TMI.2021.3053008
![]() |
[14] | Y. Gao, M. Zhou, D. N. Metaxas, UTNet: A hybrid transformer architecture for medical image segmentation, In: Medical Image Computing and Computer Assisted Intervention, 2021, 61−71. https://doi.org/10.1007/978-3-030-87199-4_6 |
[15] | A. Rahman, J. Valanarasu, I. Hacihaliloglu, V. M. Patel, Ambiguous medical image segmentation using diffusion models, In: IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, 11536−11546. https://doi.org/10.1109/CVPR52729.2023.01110 |
[16] |
J. Schlemper, O. Oktay, M. Schaap, M. Heinrich, B. Kainz, B. Glocker, et al., Attention gated networks: Learning to leverage salient regions in medical images, Med. Image Anal., 53 (2019), 197−207. https://doi.org/10.1016/j.media.2019.01.012 doi: 10.1016/j.media.2019.01.012
![]() |
[17] |
J. Hu, L. Shen, G. Sun, Squeeze-and-excitation networks, IEEE T. Pattern Anal., 42 (2020), 2011−2023. https://doi.org/10.1109/TPAMI.2019.2913372 doi: 10.1109/TPAMI.2019.2913372
![]() |
[18] | S. Woo, J. Park, J. Y. Lee, I. S. Kweon, Cbam: Convolutional block attention module, In: European Conference on Computer Vision-ECCV 2018, Lecture Notes in Computer Science, 2018, 11211. https://doi.org/10.1007/978-3-030-01234-2_1 |
[19] | Y. Chen, X. Dai, M. Liu, D. Chen, L. Yuan, Z. Liu, Dynamic convolution: Attention over convolution kernels, In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, 11027−11036. https://doi.org/10.1109/CVPR42600.2020.01104 |
[20] |
O. Bernard, A. Lalande, C. Zotti, F. Cervenansky, X. Yang, P. A. Heng, et al., Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: Is the problem solved? IEEE T. Med. Imaging, 37 (2018), 2514−2525. https://doi.org/10.1109/TMI.2018.2837502 doi: 10.1109/TMI.2018.2837502
![]() |
[21] |
V. Campello, P. Gkontra; C. Izquierdo, C. Martin-Isla, A. Sojoudi, P. M. Full, et al., Multi-centre, multi-vendor and multi-disease cardiac segmentation: The M&Ms challenge, IEEE T. Med. Imaging, 12 (2021), 3543−3554. https://doi.org/10.1109/TMI.2021.3090082 doi: 10.1109/TMI.2021.3090082
![]() |
[22] | T. Lin, P. Dollar, R. Girshick, K. He, B. Hariharan, S. Belongie, Feature pyramid networks for object detection, In: IEEE Conference on Computer Vision and Pattern Recognition, 2017, 2117−2125. https://doi.org/10.1109/CVPR.2017.106 |
[23] | Y. Ioannou, D. Robertson, R. Cipolla, A. Criminisi, Deep roots: Improving CNN efficiency with hierarchical filter groups, In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, 2017, 1231−1240. https://doi.org/10.1109/CVPR.2017.633 |
[24] |
A. Ammari, R. Mahmoudi, B. Hmida, R. Saouli, M. H. Bedoui, A review of approaches investigated for right ventricular segmentation using short-axis cardiac MRI, IET Image Process., 15 (2021), 1845−1868. https://doi.org/10.1049/ipr2.12165 doi: 10.1049/ipr2.12165
![]() |
[25] | K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, In: 2016 IEEE Conference on Computer Vision and Pattern Recognition, 2016,770−778. https://doi.org/10.1109/CVPR.2016.90 |
[26] |
G. Simantiris, G. Tziritas, Cardiac MRI segmentation with a dilated CNN incorporating domainspecific constraints, IEEE J.-STSP, 14 (2020), 1235−1243. https://doi.org/10.1109/JSTSP.2020.3013351 doi: 10.1109/JSTSP.2020.3013351
![]() |
[27] | J. M. Wolterink, T. Leiner, M. A. Viergever, I. Išgum, Automatic segmentation and disease classification using cardiac cine MR images, In: Statistical Atlases and Computational Models of the Heart & ACDC and MMWHS Challenges, 2017, 1235−1243. https://doi.org/10.1007/978-3-319-75541-0_11 |
[28] | C. Baumgartner, L. Koch, M. Pollefeys, E. Konukoglu, An exploration of 2D and 3D deep learning techniques for cardiac MR image segmentation, In: Statistical Atlases and Computational Models the Heart & ACDC and MMWHS Challenges, 2018,111−119. https://doi.org/10.1007/978-3-319-75541-0_12 |
[29] | F. Isensee, P. F. Jaeger, P. M. Full, I. Wolf, S. Engelhardt, K. H. Maier-Hein, Automatic cardiac disease assessment on cine-MRI via time-series segmentation and domain specific features, In: Statistical Atlases and Computational Models of the Heart, ACDC and MMWHS Challenges, 2018,120−129. https://doi.org/10.1007/978-3-319-75541-0_13 |
[30] | P. Full, F. Isensee, P. Jager, K. Maier-Hein, Studying robustness of semantic segmentation under domain shift in cardiac MRI, In: Statistical Atlases and Computational Models of the Heart, M&Ms and EMIDEC Challenges, 2021,238−249. https://doi.org/10.1007/978-3-030-68107-4_24 |
[31] | Y. Zhang, J. Yang, F. Hou, Y. Liu, Y. Wang, J. Tian, et al., Semi-supervised cardiac image segmentation via label propagation and style transfer, In: Statistical Atlases and Computational Models of the Heart, M&Ms and EMIDEC Challenges, 2021,219−227. https://doi.org/10.1007/978-3-030-68107-4_22 |
[32] | J. Ma, Histogram matching augmentation for domain adaptation with application to multi-centre, multi-vendor and multi-disease cardiac image segmentation, In: Statistical Atlases and Computational Models of the Heart, M&Ms and EMIDEC Challenges, 2021,177−186. https://doi.org/10.1007/978-3-030-68107-4_18 |
[33] |
F. Isensee, P. Jaeger, S. Kohl, J. Petersen, K. H. Maier-Hein, nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation, Nat. Methods, 18 (2021), 203−211. https://doi.org/10.1038/s41592-020-01008-z doi: 10.1038/s41592-020-01008-z
![]() |
[34] |
M. Forouzanfar, N. Forghani, M. Teshnehlab, Parameter optimization of improved fuzzy c-means clustering algorithm for brain MR image segmentation, Eng. Appl. Artif. Intel., 23 (2010), 160−168. https://doi.org/10.1016/j.engappai.2009.10.002 doi: 10.1016/j.engappai.2009.10.002
![]() |
[35] |
N. Tajbakhsh, J. Shin; R. Suryakanth, R. T. Hurst, C. B. Kendall, M. B. Gotway, et al., Convolutional neural networks for medical image analysis: Full training or fine tuning? IEEE T. Med. Imaging, 35 (2016), 1299−1312. https://doi.org/10.1109/TMI.2016.2535302 doi: 10.1109/TMI.2016.2535302
![]() |
[36] |
T. Lossau, H. Nickisch, T. Wissel, M. Morlock, M. Grass, Learning metal artifact reduction in cardiac CT images with moving pacemakers, Med. Image Anal., 61 (2020), 101655. https://doi.org/10.1016/j.media.2020.101655 doi: 10.1016/j.media.2020.101655
![]() |
1. | Majdi Khalid, Sugitha Deivasigamani, Sathiya V, Surendran Rajendran, An efficient colorectal cancer detection network using atrous convolution with coordinate attention transformer and histopathological images, 2024, 14, 2045-2322, 10.1038/s41598-024-70117-y |
Dataset | Frames | Train | Test | Resolution | Ground Truth |
dataset 1 | 2,979 | 1,902 | 1,076 | ⩽256×256 pixels | Background, LV, RV, Myo |
dataset 2 | 3,264 | 1,643 | 1,621 | ⩽256×256 pixels | Background, LV, RV, Myo |
dataset 3 | 4,649 | 0 | 4,649 | ⩽512×512 pixels | LV end-diastolic inner diameter |
Baseline | Attention | FCM | MSFR | dyConv | mean DSC(%) | mean HD(mm) | p-values | FLOPs(G) |
√ | 89.23 | 4.94 | − | 7.85 | ||||
√ | √ | 90.86(+1.63) | 3.82(−1.12) | 0.2698 | 7.86 | |||
√ | √ | 92.28(+3.05) | 3.22(−1.72) | 0.0010 | 11.64 | |||
√ | √ | √ | 93.19(+3.96) | 2.73(−2.21) | 0.1045 | 11.64 | ||
√ | √ | √ | √ | 94.42(+4.94) | 2.50(−2.44) | 0.0340 | 11.73 |
Structure | DSC(%) | HD(mm) | p-values | ||||||
LV | RV | Myo | mean | LV | RV | Myo | mean | ||
Conv+Conv | 89.31 | 89.45 | 94.07 | 90.94 | 5.61 | 3.00 | 2.53 | 3.71 | − |
Conv+CFConv | 89.76 | 89.51 | 94.21 | 91.16 | 5.66 | 3.03 | 2.47 | 3.72 | 0.7831 |
CFConv+Conv | 90.16 | 89.35 | 93.77 | 91.09 | 5.15 | 3.04 | 2.53 | 3.57 | 0.8480 |
CFConv+CFConv | 89.38 | 89.97 | 93.79 | 91.05 | 6.42 | 3.34 | 2.84 | 4.20 | 0.5628 |
Conv+SFConv | 90.33 | 89.55 | 94.00 | 91.29 | 5.16 | 2.79 | 2.38 | 3.44 | 0.7843 |
SFConv+Conv | 90.15 | 89.29 | 94.08 | 91.17 | 5.27 | 2.98 | 2.19 | 3.48 | 0.5673 |
SFConv+SFConv | 88.48 | 89.06 | 94.06 | 90.53 | 5.96 | 2.93 | 2.39 | 3.76 | 0.3872 |
CFConv+SFConv | 93.95 | 90.49 | 94.29 | 92.91 | 3.90 | 2.79 | 2.36 | 3.02 | 0.0173 |
SFConv+CFConv | 95.23 | 91.53 | 95.76 | 94.17 | 3.20 | 2.34 | 1.96 | 2.50 | < 0.0001 |
Encoder | Decoder | DSC(%) | HD(mm) | p-values | ||||||
LV | RV | Myo | mean | LV | RV | Myo | mean | |||
89.31 | 89.45 | 94.07 | 90.94 | 5.61 | 3.00 | 2.53 | 3.71 | − | ||
√ | 93.87 | 90.28 | 94.88 | 93.01 | 3.75 | 2.69 | 2.20 | 2.88 | 0.0021 | |
√ | 92.17 | 88.89 | 94.13 | 91.73 | 6.07 | 3.73 | 3.18 | 4.33 | 0.2526 | |
√ | √ | 95.23 | 91.53 | 95.76 | 94.17 | 3.20 | 2.34 | 1.96 | 2.50 | 0.0128 |
Method | Numbers | 1 | 2 | 3 | 4 | 5 | mean | σ2 | |
Baseline | Dice loss | Train | 0.0574 | 0.0537 | 0.0609 | 0.0549 | 0.0526 | 0.0559 | 8.7960×10-6 |
Test | 0.0848 | 0.0834 | 0.0945 | 0.0799 | 0.0834 | 0.0852 | 2.4244×10-7 | ||
DSC(%) | Train | 93.02 | 93.61 | 92.58 | 93.42 | 93.73 | 93.27 | 0.1777 | |
Test | 89.39 | 89.66 | 88.12 | 90.09 | 89.68 | 89.39 | 0.4520 | ||
DAFNet | Dice loss | Train | 0.0468 | 0.0459 | 0.0457 | 0.0454 | 0.0449 | 0.0457 | 3.9600×10-7 |
Test | 0.0505 | 0.0500 | 0.0502 | 0.0481 | 0.0489 | 0.0495 | 8.1200×10-7 | ||
DSC(%) | Train | 94.53 | 94.68 | 94.68 | 94.72 | 94.77 | 94.68 | 0.0064 | |
Test | 94.05 | 94.13 | 94.09 | 94.36 | 94.24 | 94.17 | 0.0127 |
Methods | LV | RV | Myo | mean | ||||||||||
DSC(%) | HD(mm) | DSC(%) | HD(mm) | DSC(%) | HD(mm) | DSC(%) | HD(mm) | |||||||
ED | ES | ED | ES | ED | ES | ED | ES | ED | ES | ED | ES | |||
Reference [9] | 90.38 | 85.61 | 6.74 | 6.93 | 87.27 | 86.78 | 3.88 | 5.06 | 95.57 | 89.80 | 3.23 | 3.81 | 90.94 | 3.71 |
Reference [12] | 96.60 | 91.80 | 7.95 | 8.18 | 94.80 | 89.80 | 11.67 | 12.84 | 90.50 | 91.50 | 9.50 | 9.57 | 92.50 | 9.95 |
Reference [16] | 96.13 | 92.37 | 3.62 | 4.37 | 90.14 | 90.65 | 2.48 | 3.03 | 97.27 | 93.36 | 1.98 | 2.67 | 93.32 | 3.03 |
Reference [14] | 96.31 | 92.84 | 3.13 | 3.97 | 90.65 | 90.93 | 2.76 | 3.96 | 96.79 | 93.23 | 1.88 | 2.55 | 93.46 | 3.04 |
Reference [15] | 95.59 | 91.94 | 3.82 | 4.33 | 90.03 | 90.43 | 3.04 | 4.03 | 96.37 | 92.67 | 2.38 | 3.14 | 93.36 | 3.46 |
Reference [26] | 96.70 | 92.80 | 6.40 | 7.60 | 93.60 | 88.90 | 13.30 | 14.40 | 89.10 | 90.40 | 8.30 | 9.60 | 91.92 | 9.93 |
Reference [27] | 96.10 | 91.80 | 7.50 | 9.60 | 92.80 | 87.20 | 11.90 | 13.40 | 87.50 | 89.40 | 11.10 | 10.70 | 90.80 | 10.70 |
Reference [28] | 96.30 | 91.10 | 6.50 | 9.20 | 93.20 | 88.30 | 12.70 | 14.70 | 89.20 | 90.10 | 8.70 | 10.60 | 91.37 | 10.40 |
Reference [29] | 96.70 | 92.80 | 5.50 | 6.90 | 94.60 | 90.40 | 8.80 | 11.40 | 89.60 | 91.90 | 7.60 | 7.10 | 92.67 | 7.88 |
DAFNet-Ref. [18] | 96.01 | 92.57 | 3.31 | 4.38 | 90.38 | 90.43 | 2.33 | 3.10 | 97.17 | 93.20 | 1.92 | 2.58 | 93.29 | 2.94 |
DAFNet-Ref. [19] | 96.42 | 92.66 | 3.43 | 4.03 | 92.04 | 92.34 | 2.18 | 2.89 | 97.52 | 94.15 | 1.87 | 2.43 | 94.18 | 2.81 |
DAFNet-Res18(Ours) | 96.53 | 92.78 | 3.01 | 3.89 | 92.12 | 90.89 | 2.90 | 2.84 | 96.89 | 91.99 | 2.10 | 3.62 | 93.53 | 3.06 |
DAFNet-Res34(Ours) | 96.41 | 92.89 | 2.96 | 3.77 | 91.97 | 91.03 | 2.99 | 2.65 | 97.13 | 92.78 | 2.03 | 3.03 | 93.70 | 2.91 |
DAFNet(Ours) | 96.72 | 93.74 | 2.78 | 3.61 | 91.68 | 91.38 | 2.09 | 2.59 | 97.53 | 93.99 | 1.70 | 2.22 | 94.17 | 2.50 |
Method | LV | RV | Myo | mean | ||||
DSC (%) | HD (mm) | DSC (%) | HD (mm) | DSC (%) | HD (mm) | DSC (%) | HD (mm) | |
Reference [9] | 87.69 | 6.93 | 81.13 | 4.57 | 85.63 | 3.64 | 84.82 | 5.05 |
Reference [12] | 89.21 | 5.68 | 83.68 | 4.28 | 87.25 | 3.29 | 86.71 | 4.42 |
Reference [16] | 90.94 | 5.34 | 85.59 | 3.79 | 87.74 | 3.12 | 88.09 | 4.08 |
Reference [30] | 91.25 | 9.10 | 85.30 | 11.70 | 88.50 | 12.25 | 88.35 | 11.03 |
Reference [31] | 90.90 | 9.40 | 84.55 | 11.85 | 87.95 | 12.65 | 87.80 | 11.30 |
Reference [32] | 90.50 | 10.00 | 84.10 | 12.45 | 87.50 | 12.65 | 87.37 | 11.70 |
DAFNet-Res18 | 92.34 | 5.23 | 84.28 | 3.24 | 87.23 | 3.23 | 87.95 | 3.90 |
DAFNet-Res34 | 91.23 | 5.63 | 84.47 | 3.68 | 87.84 | 2.67 | 87.85 | 3.99 |
DAFNet (Ours) | 92.58 | 4.87 | 84.17 | 3.97 | 88.68 | 2.54 | 88.47 | 3.79 |
Methods | Patients | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | mean | σ2 |
TLVd | 44.00 | 47.00 | 58.00 | 42.00 | 51.00 | 54.00 | 68.00 | 44.00 | 45.00 | 47.00 | 61.00 | 38.00 | — | — | |
Reference [9] | PLVd | 41.72 | 43.78 | 53.33 | 40.77 | 54.34 | 51.81 | 65.93 | 46.32 | 43.65 | 44.14 | 59.74 | 39.84 | — | — |
accuracy | 94.82 | 93.15 | 91.95 | 97.07 | 93.45 | 95.94 | 96.96 | 94.73 | 97.00 | 93.91 | 97.93 | 95.16 | 95.17 | 1.77 | |
Reference [12] | PLVd | 42.32 | 45.37 | 57.88 | 41.30 | 50.80 | 52.72 | 66.94 | 45.22 | 43.79 | 45.44 | 59.26 | 39.06 | — | — |
accuracy | 96.18 | 96.53 | 99.79 | 98.33 | 99.61 | 97.63 | 98.44 | 97.23 | 97.31 | 96.68 | 97.15 | 97.21 | 97.67 | 1.22 | |
Reference [16] | PLVd | 42.30 | 45.18 | 56.97 | 40.85 | 50.55 | 53.30 | 66.66 | 45.64 | 43.34 | 45.13 | 59.79 | 39.01 | — | — |
accuracy | 96.14 | 96.13 | 98.22 | 97.26 | 99.12 | 98.70 | 98.03 | 96.27 | 96.31 | 96.02 | 98.02 | 97.34 | 97.30 | 1.13 | |
DAFNet-Res18 | PLVd | 44.38 | 48.25 | 56.88 | 41.78 | 51.21 | 54.79 | 67.32 | 45.39 | 44.13 | 45.39 | 61.79 | 38.67 | — | — |
accuracy | 99.14 | 97.34 | 98.07 | 99.48 | 99.59 | 98.54 | 99.00 | 96.84 | 98.07 | 96.57 | 98.70 | 98.24 | 98.30 | 0.94 | |
DAFNet-Res34 | PLVd | 45.79 | 48.17 | 57.03 | 41.42 | 51.43 | 55.67 | 68.77 | 43.75 | 45.76 | 46.21 | 62.68 | 38.94 | — | — |
accuracy | 95.93 | 97.51 | 98.33 | 98.62 | 99.16 | 96.91 | 98.87 | 99.43 | 98.31 | 98.32 | 97.25 | 97.53 | 98.01 | 0.97 | |
DAFNet(Ours) | PLVd | 45.33 | 48.07 | 57.94 | 42.51 | 50.71 | 55.26 | 68.59 | 42.14 | 45.88 | 46.83 | 61.69 | 38.26 | — | — |
accuracy | 96.98 | 97.72 | 99.90 | 98.79 | 99.43 | 97.67 | 99.13 | 95.78 | 98.04 | 99.64 | 98.87 | 99.32 | 98.44 | 1.18 |
Dataset | Frames | Train | Test | Resolution | Ground Truth |
dataset 1 | 2,979 | 1,902 | 1,076 | ⩽256×256 pixels | Background, LV, RV, Myo |
dataset 2 | 3,264 | 1,643 | 1,621 | ⩽256×256 pixels | Background, LV, RV, Myo |
dataset 3 | 4,649 | 0 | 4,649 | ⩽512×512 pixels | LV end-diastolic inner diameter |
Baseline | Attention | FCM | MSFR | dyConv | mean DSC(%) | mean HD(mm) | p-values | FLOPs(G) |
√ | 89.23 | 4.94 | − | 7.85 | ||||
√ | √ | 90.86(+1.63) | 3.82(−1.12) | 0.2698 | 7.86 | |||
√ | √ | 92.28(+3.05) | 3.22(−1.72) | 0.0010 | 11.64 | |||
√ | √ | √ | 93.19(+3.96) | 2.73(−2.21) | 0.1045 | 11.64 | ||
√ | √ | √ | √ | 94.42(+4.94) | 2.50(−2.44) | 0.0340 | 11.73 |
Structure | DSC(%) | HD(mm) | p-values | ||||||
LV | RV | Myo | mean | LV | RV | Myo | mean | ||
Conv+Conv | 89.31 | 89.45 | 94.07 | 90.94 | 5.61 | 3.00 | 2.53 | 3.71 | − |
Conv+CFConv | 89.76 | 89.51 | 94.21 | 91.16 | 5.66 | 3.03 | 2.47 | 3.72 | 0.7831 |
CFConv+Conv | 90.16 | 89.35 | 93.77 | 91.09 | 5.15 | 3.04 | 2.53 | 3.57 | 0.8480 |
CFConv+CFConv | 89.38 | 89.97 | 93.79 | 91.05 | 6.42 | 3.34 | 2.84 | 4.20 | 0.5628 |
Conv+SFConv | 90.33 | 89.55 | 94.00 | 91.29 | 5.16 | 2.79 | 2.38 | 3.44 | 0.7843 |
SFConv+Conv | 90.15 | 89.29 | 94.08 | 91.17 | 5.27 | 2.98 | 2.19 | 3.48 | 0.5673 |
SFConv+SFConv | 88.48 | 89.06 | 94.06 | 90.53 | 5.96 | 2.93 | 2.39 | 3.76 | 0.3872 |
CFConv+SFConv | 93.95 | 90.49 | 94.29 | 92.91 | 3.90 | 2.79 | 2.36 | 3.02 | 0.0173 |
SFConv+CFConv | 95.23 | 91.53 | 95.76 | 94.17 | 3.20 | 2.34 | 1.96 | 2.50 | < 0.0001 |
Encoder | Decoder | DSC(%) | HD(mm) | p-values | ||||||
LV | RV | Myo | mean | LV | RV | Myo | mean | |||
89.31 | 89.45 | 94.07 | 90.94 | 5.61 | 3.00 | 2.53 | 3.71 | − | ||
√ | 93.87 | 90.28 | 94.88 | 93.01 | 3.75 | 2.69 | 2.20 | 2.88 | 0.0021 | |
√ | 92.17 | 88.89 | 94.13 | 91.73 | 6.07 | 3.73 | 3.18 | 4.33 | 0.2526 | |
√ | √ | 95.23 | 91.53 | 95.76 | 94.17 | 3.20 | 2.34 | 1.96 | 2.50 | 0.0128 |
Method | Numbers | 1 | 2 | 3 | 4 | 5 | mean | σ2 | |
Baseline | Dice loss | Train | 0.0574 | 0.0537 | 0.0609 | 0.0549 | 0.0526 | 0.0559 | 8.7960×10-6 |
Test | 0.0848 | 0.0834 | 0.0945 | 0.0799 | 0.0834 | 0.0852 | 2.4244×10-7 | ||
DSC(%) | Train | 93.02 | 93.61 | 92.58 | 93.42 | 93.73 | 93.27 | 0.1777 | |
Test | 89.39 | 89.66 | 88.12 | 90.09 | 89.68 | 89.39 | 0.4520 | ||
DAFNet | Dice loss | Train | 0.0468 | 0.0459 | 0.0457 | 0.0454 | 0.0449 | 0.0457 | 3.9600×10-7 |
Test | 0.0505 | 0.0500 | 0.0502 | 0.0481 | 0.0489 | 0.0495 | 8.1200×10-7 | ||
DSC(%) | Train | 94.53 | 94.68 | 94.68 | 94.72 | 94.77 | 94.68 | 0.0064 | |
Test | 94.05 | 94.13 | 94.09 | 94.36 | 94.24 | 94.17 | 0.0127 |
Methods | LV | RV | Myo | mean | ||||||||||
DSC(%) | HD(mm) | DSC(%) | HD(mm) | DSC(%) | HD(mm) | DSC(%) | HD(mm) | |||||||
ED | ES | ED | ES | ED | ES | ED | ES | ED | ES | ED | ES | |||
Reference [9] | 90.38 | 85.61 | 6.74 | 6.93 | 87.27 | 86.78 | 3.88 | 5.06 | 95.57 | 89.80 | 3.23 | 3.81 | 90.94 | 3.71 |
Reference [12] | 96.60 | 91.80 | 7.95 | 8.18 | 94.80 | 89.80 | 11.67 | 12.84 | 90.50 | 91.50 | 9.50 | 9.57 | 92.50 | 9.95 |
Reference [16] | 96.13 | 92.37 | 3.62 | 4.37 | 90.14 | 90.65 | 2.48 | 3.03 | 97.27 | 93.36 | 1.98 | 2.67 | 93.32 | 3.03 |
Reference [14] | 96.31 | 92.84 | 3.13 | 3.97 | 90.65 | 90.93 | 2.76 | 3.96 | 96.79 | 93.23 | 1.88 | 2.55 | 93.46 | 3.04 |
Reference [15] | 95.59 | 91.94 | 3.82 | 4.33 | 90.03 | 90.43 | 3.04 | 4.03 | 96.37 | 92.67 | 2.38 | 3.14 | 93.36 | 3.46 |
Reference [26] | 96.70 | 92.80 | 6.40 | 7.60 | 93.60 | 88.90 | 13.30 | 14.40 | 89.10 | 90.40 | 8.30 | 9.60 | 91.92 | 9.93 |
Reference [27] | 96.10 | 91.80 | 7.50 | 9.60 | 92.80 | 87.20 | 11.90 | 13.40 | 87.50 | 89.40 | 11.10 | 10.70 | 90.80 | 10.70 |
Reference [28] | 96.30 | 91.10 | 6.50 | 9.20 | 93.20 | 88.30 | 12.70 | 14.70 | 89.20 | 90.10 | 8.70 | 10.60 | 91.37 | 10.40 |
Reference [29] | 96.70 | 92.80 | 5.50 | 6.90 | 94.60 | 90.40 | 8.80 | 11.40 | 89.60 | 91.90 | 7.60 | 7.10 | 92.67 | 7.88 |
DAFNet-Ref. [18] | 96.01 | 92.57 | 3.31 | 4.38 | 90.38 | 90.43 | 2.33 | 3.10 | 97.17 | 93.20 | 1.92 | 2.58 | 93.29 | 2.94 |
DAFNet-Ref. [19] | 96.42 | 92.66 | 3.43 | 4.03 | 92.04 | 92.34 | 2.18 | 2.89 | 97.52 | 94.15 | 1.87 | 2.43 | 94.18 | 2.81 |
DAFNet-Res18(Ours) | 96.53 | 92.78 | 3.01 | 3.89 | 92.12 | 90.89 | 2.90 | 2.84 | 96.89 | 91.99 | 2.10 | 3.62 | 93.53 | 3.06 |
DAFNet-Res34(Ours) | 96.41 | 92.89 | 2.96 | 3.77 | 91.97 | 91.03 | 2.99 | 2.65 | 97.13 | 92.78 | 2.03 | 3.03 | 93.70 | 2.91 |
DAFNet(Ours) | 96.72 | 93.74 | 2.78 | 3.61 | 91.68 | 91.38 | 2.09 | 2.59 | 97.53 | 93.99 | 1.70 | 2.22 | 94.17 | 2.50 |
Method | LV | RV | Myo | mean | ||||
DSC (%) | HD (mm) | DSC (%) | HD (mm) | DSC (%) | HD (mm) | DSC (%) | HD (mm) | |
Reference [9] | 87.69 | 6.93 | 81.13 | 4.57 | 85.63 | 3.64 | 84.82 | 5.05 |
Reference [12] | 89.21 | 5.68 | 83.68 | 4.28 | 87.25 | 3.29 | 86.71 | 4.42 |
Reference [16] | 90.94 | 5.34 | 85.59 | 3.79 | 87.74 | 3.12 | 88.09 | 4.08 |
Reference [30] | 91.25 | 9.10 | 85.30 | 11.70 | 88.50 | 12.25 | 88.35 | 11.03 |
Reference [31] | 90.90 | 9.40 | 84.55 | 11.85 | 87.95 | 12.65 | 87.80 | 11.30 |
Reference [32] | 90.50 | 10.00 | 84.10 | 12.45 | 87.50 | 12.65 | 87.37 | 11.70 |
DAFNet-Res18 | 92.34 | 5.23 | 84.28 | 3.24 | 87.23 | 3.23 | 87.95 | 3.90 |
DAFNet-Res34 | 91.23 | 5.63 | 84.47 | 3.68 | 87.84 | 2.67 | 87.85 | 3.99 |
DAFNet (Ours) | 92.58 | 4.87 | 84.17 | 3.97 | 88.68 | 2.54 | 88.47 | 3.79 |
Methods | Patients | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | mean | σ2 |
TLVd | 44.00 | 47.00 | 58.00 | 42.00 | 51.00 | 54.00 | 68.00 | 44.00 | 45.00 | 47.00 | 61.00 | 38.00 | — | — | |
Reference [9] | PLVd | 41.72 | 43.78 | 53.33 | 40.77 | 54.34 | 51.81 | 65.93 | 46.32 | 43.65 | 44.14 | 59.74 | 39.84 | — | — |
accuracy | 94.82 | 93.15 | 91.95 | 97.07 | 93.45 | 95.94 | 96.96 | 94.73 | 97.00 | 93.91 | 97.93 | 95.16 | 95.17 | 1.77 | |
Reference [12] | PLVd | 42.32 | 45.37 | 57.88 | 41.30 | 50.80 | 52.72 | 66.94 | 45.22 | 43.79 | 45.44 | 59.26 | 39.06 | — | — |
accuracy | 96.18 | 96.53 | 99.79 | 98.33 | 99.61 | 97.63 | 98.44 | 97.23 | 97.31 | 96.68 | 97.15 | 97.21 | 97.67 | 1.22 | |
Reference [16] | PLVd | 42.30 | 45.18 | 56.97 | 40.85 | 50.55 | 53.30 | 66.66 | 45.64 | 43.34 | 45.13 | 59.79 | 39.01 | — | — |
accuracy | 96.14 | 96.13 | 98.22 | 97.26 | 99.12 | 98.70 | 98.03 | 96.27 | 96.31 | 96.02 | 98.02 | 97.34 | 97.30 | 1.13 | |
DAFNet-Res18 | PLVd | 44.38 | 48.25 | 56.88 | 41.78 | 51.21 | 54.79 | 67.32 | 45.39 | 44.13 | 45.39 | 61.79 | 38.67 | — | — |
accuracy | 99.14 | 97.34 | 98.07 | 99.48 | 99.59 | 98.54 | 99.00 | 96.84 | 98.07 | 96.57 | 98.70 | 98.24 | 98.30 | 0.94 | |
DAFNet-Res34 | PLVd | 45.79 | 48.17 | 57.03 | 41.42 | 51.43 | 55.67 | 68.77 | 43.75 | 45.76 | 46.21 | 62.68 | 38.94 | — | — |
accuracy | 95.93 | 97.51 | 98.33 | 98.62 | 99.16 | 96.91 | 98.87 | 99.43 | 98.31 | 98.32 | 97.25 | 97.53 | 98.01 | 0.97 | |
DAFNet(Ours) | PLVd | 45.33 | 48.07 | 57.94 | 42.51 | 50.71 | 55.26 | 68.59 | 42.14 | 45.88 | 46.83 | 61.69 | 38.26 | — | — |
accuracy | 96.98 | 97.72 | 99.90 | 98.79 | 99.43 | 97.67 | 99.13 | 95.78 | 98.04 | 99.64 | 98.87 | 99.32 | 98.44 | 1.18 |