Skip to main content

ORIGINAL RESEARCH article

Front. Oncol., 25 March 2022
Sec. Cancer Imaging and Image-directed Interventions

BUSnet: A Deep Learning Model of Breast Tumor Lesion Detection for Ultrasound Images

  • 1Faculty of Electronic Information and Electrical Engineering, Dalian University of Technology, Dalian, China
  • 2Department of Surgery, The Second Hospital of Dalian Medical University, Dalian, China

Ultrasound (US) imaging is a main modality for breast disease screening. Automatically detecting the lesions in US images is essential for developing the artificial-intelligence-based diagnostic support technologies. However, the intrinsic characteristics of ultrasound imaging, like speckle noise and acoustic shadow, always degenerate the detection accuracy. In this study, we developed a deep learning model called BUSnet to detect the breast tumor lesions in US images with high accuracy. We first developed a two-stage method including the unsupervised region proposal and bounding-box regression algorithms. Then, we proposed a post-processing method to enhance the detecting accuracy further. The proposed method was used to a benchmark dataset, which includes 487 benign samples and 210 malignant samples. The results proved the effectiveness and accuracy of the proposed method.

Introduction

Breast cancer is the most frequently diagnosed among Chinese women (1). Early detection of breast cancer is an effective method to decrease the morality rate dramatically (2). Because the ultrasound imaging (US) technique is a low-cost way to offer favorable sensitivity and detection rates for early cancer, it is a widely applied modality for breast cancer detection in China (3). Unlike computed tomography or magnetic resonance imaging, quality control is a critical issue for US. Especially when handheld ultrasound is used to screen the whole breast, the imaging quality is positively related to the skill levels of the radiologists (4). In addition, the artifacts in US, including the speckle noise and the acoustic shadow, are another issue degenerating the imaging quality. Thus, it is imperative to develop US diagnostic support technologies to resolve these operation-dependent difficulties.

Over the past decade, machine learning (ML) is being increasingly applied in the research and application of medical imaging techniques (4, 5). Furthermore, deep learning (DL) is an advantage subset of ML, in which convolutional neural networks (CNN) were initially designed for image analysis (6). The state-of-the-art algorithms of ML and DL are extensively applied for image classification (6), object detection (7), and segmentation (8), which offers the potential to develop US diagnostic support technologies.

Classification estimates the label for the entire image (9). constructed CNN to classify the US images into benign and malignant breast tumors. CNN was further improved by introducing a matching layer for breast mass classification (10). The support vector machine, VGG, and ResNet were also applied to the classification issues for the breast US images (1113). Object detection can automatically identify the specified targets in the images (12). used VGG and ResNet for the classification and lesion detection for US images (14). evaluated the performance of detecting breast cancer lesions in US images for three deep learning models. Segmentation is used for the further precise measurements and structures of targets (15). proposed a U-net-based segmentation algorithm for suspicious breast masses in US images. Object detection can offer more detailed information than classification (16) and can also save more costs of labeling data and training networks than segmentation (17). For these reasons, object detection is often used for the breast disease lesion identification for US images, which can help the radiologist make a diagnosis efficiently with expected high sensitivity and specificity (14).

Although several ingenious methods for the breast tumor lesions detection for US images have been reported, as indicated by (4), the progress of medical artificial intelligence (AI) research and development in US is slower than that in other modalities. This is mainly because a single ML or DL method cannot tackle the problem caused by the artifacts in US images. Effective ML and DL methods for lesion detection for US images should integrate preprocessing, like denoising, and post-processing, like redundancy reduction. However, such integration has not been considered enough in the previous work. In this research, we proposed a breast tumor lesion detection method called BUSnet, which includes a preprocessing procedure, a deep learning model for the lesion detection, and a post-processing procedure. In the preprocessing procedure, we combined the unsupervised Canny edge detection (18) with the selective search (19) to determine the region of interest (RoI) candidates. We trained a G-CNN (20) for the bounding-box regression. We proposed a novel post-processing method by improving the non-maximum suppression method (21) to enhance the detection accuracy. The dataset obtained from (22) was used to demonstrate the outstanding performance of our method.

Methods

Data

The dataset, including 780 samples (133 normal samples, 487 benign samples, and 210 malignant samples), was achieved from (22). In this research, the benign and malignant samples were used to develop BUSnet. 80% of samples were used as the training data, and the rest were used as the test data. A sample example is shown in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1 An US image sample. (A) Original image. (B) Original segmentation ground truth in binary mask form provided by (22). (C) Conversion to bounding box as the ground truth for RoI detection and localization.

Algorithm Framework

Figure 2 shows the algorithm. RoI candidates are screened out by a preprocessing method and an unsupervised region proposal method. The classification and bounding-box regression networks were constructed for the RoI candidates. We select all the RoI candidates with the probability of being lesion higher than 0.9 according to (23) and adjust their position with bounding-box regression. Finally, we aggregate these bounding boxes and achieve the final output.

FIGURE 2
www.frontiersin.org

Figure 2 Framework of BUSnet.

Pre-Processing

We first conduct the histogram equalization, the normalization, and the wavelet-based denoising to the US images to ease the effect of the uncertainties like the speckle noise. Then, we clip 30% lower part of the images to reduce the redundant RoIs according to (24).

Region Proposal and Classification

We propose an unsupervised region proposal method by using the Canny edge detector and the selective search to obtain the RoI candidates. The Canny edge detector is carried out as follows:

1. The Gaussian-smoothing is applied to the images.

2. The Sobel operators in two directions

   Sobelx=(121000121), Sobely=(101202101).

   are used to the smoothed images to detect the edges.

3. Non-maxima suppression is conducted. The gradient intensity of the current pixel and that of the two pixels along the positive and negative gradient directions are compared. If the gradient intensity of the current pixel is larger than the other two pixels, the current pixel is retained as an edge point; otherwise, the current pixel will be discarded.

4. A two-threshold test is applied. A high threshold and a low threshold are set. If the gradient value of the edge pixel is higher than the high threshold, it will be marked as a strong edge pixel; if the gradient value of the edge pixel is between the two thresholds, it will be marked as a weak edge pixel; if the gradient value of the edge pixel is less than the low threshold, it will be suppressed.

The Canny edge detection removes redundant information and noise. Then, the selective search method is applied to identify RoI candidates. In the selective search method, for the regions i and j, let sgray(i,j), stexture(i,j), ssize(i,j), and sfill(i,j) denote the similarities of gray level, texture, size, and fullness, respectively. The sum s(i,j) = sgray(i,j) + stexture(i,j) + ssize(i,j) + sfill(i,j) is used to measure the total similarity between the regions i and j. In the following procedure, the total similarity is used to emerge regions:

1. Initialize the region set R = {1,2,…,n} and similarity set S = ∅.

2. Calculate pair-wised total similarities s(i,j) and conduct S = Ss(i, j).

3. Let s(p,q) = max S, and t = pq.

4. Remove all the similarities associated with regions p and q, i.e., S = S\{s(p,*), s(*,q)}.

5. Take t as a new region and R = Rt and S = S ∪ {s(t,*)}.

6. Repeat step 2-5, until S = ∅.

After finishing the region proposal, the RoIs are classified into lesion and normal types using ResNet50. For the RoIs with probability larger than 0.9, their position will be further adjusted by using bounding-box regression.

Bounding-Box Regression

In the following, the RoIs classified as lesions are denoted as bounding box and used as the input of the bounding-box regression. The procedure is illustrated in Figure 3. Because ResNet50 has been proven to be competent for the feature extraction of breast US images (25), we use ResNet50 as the backbone network for the bounding-box regression.

FIGURE 3
www.frontiersin.org

Figure 3 Process of iterative bounding-box regression. The red rectangle represents the bounding box, and the blue one represents the target box. At each step, the network modifies the bounding box, and the target box coordinates and uses the new bounding box as the input.

Let B = (x,y,w,h) denote the bounding box location, in which x and y are the x-axis and y-axis locations of the upper-left vertex of the box, respectively; w and h are the width and height of the box, respectively. With respect to these denotations, let G = (x*,y*,w*,h*) denote the location of the true region. The error between the bounding box and the truth box δ = (tx,ty,tw,th) is obtained by the following function:

δ=(tx,ty,tw,th)=(xxwa,yyha,log(ww),log(hh))(1)

For convenience, let δ = Δ(B,G) denote the transformation (1) from the locations B and G to the difference δ. Consequently, G can be calculated from

G=(x,y,w,h)=(tw+x,th+y,wexp (tw), h exp (th))(2)

Let G = Γ(B, δ) denote the transformation (2) from the bounding box location B and the error δ.

Motivated by (20), an iterative method is applied for the bounding-box regression. At first, the total iteration s is prefixed. For s = 1,2,…,,S, the target bounding box is obtained by

Φ(Bs,G,s)=Bs+GBsSs+1,

in which Bs is updated at each iteration as the following:

Bs=Γ(Bs1,δs1)

Then, we construct the loss function for the bounding-box regression as the following:

Lreg=s=1Si=1N[I(BisBbg)×L1s(δisΔ(Bis,Gi,s))],

where N is the total of bounding boxes; Bi1 is the initialization of the ith bounding box; and Bbg is the set of background bounding boxes with the Jaccard index being smaller than 0.2 with respect to the true box, with the definition of the Jaccard index for pixel sets A and B

J(A,B)=|AB||AB|,

(BisBbg)

is the indicator function

(BisBbg)={1,ifBisBbg0,   otherwise;

L1s is the smooth l1 loss as the following:

L1s(x)={0.5x2,        if |x|<1|x|0.5,   otherwise.

Aggregation

In the aggregation, the regressive bounding boxes are integrated into one box to eliminate the redundant information. Let ℬ denote the set of all the bounding boxes and initialize index k = 0. The aggregation procedure is as follows:

1. If ℬ ≠ ∅, let k = k + 1. Initialize Ak = { Bk0} with Bk0 being the box nearest to the center of the image and ℬ = ℬ\{ℬk0}.

2. For each ℬ ∈ ℬ, calculate pairwise Jaccard indices with the elements in 𝒜k. If any Jaccard index is larger than 0.5, 𝒜k = 𝒜k ∪ {B} and ℬ = ℬ\{ℬ}.

3. If ℬ becomes ∅, assume {𝒜k|k = 1,2,…, K} is obtained. Let 𝒜* denote the bounding box set with the largest size among {𝒜k}.

4. Calculate all the pairwise Jaccard indices of the elements in 𝒜*. Assume that B1, B2 ∈ 𝒜 are the two bounding boxes with the largest Jaccard index. If J (B1, B2) ≥ 0.7, B = B1B2. If 0.5 ≤ J(B1, B2) < 0.7, construct the smallest bounding box B that can cover B1B2.

5. 𝒜* = 𝒜*\{B1}, 𝒜* = 𝒜*\{B2}, and 𝒜* = 𝒜*B.

6. Repeat aggregations 4 and 5, until 𝒜* contains only one bounding box.

Note that multiple 𝒜*s can be achieved in step 3. In this case, all the 𝒜*s are aggregated by repeating steps 4 and 5. Then, the bounding box with the largest classification confidence is determined as the final aggregated bounding box.

Performance Metrics

The Jaccard index between the predicted set of lesion pixels and the truth ground, accuracy, precision, recall, and F1 score are used as the prediction performance metrics. Accuracy, precision, recall, and F1 score are defined as follows:

accuracy=TPTP+FP+FN,
precision=TPTP+FP,
recall=TPTP+FN,
F1 score=2recall×precisionrecall+precision,

where TP is true positive for a total of bounding boxes with a Jaccard index larger than 0.5; FP is false positive for a total bounding boxes with a Jaccard index smaller than 0.5; and FN is false negative for a total of the images without a correctly detected bounding box. Note that the true negative is set to be 0 according to (23).

Results

Ablation for Pre-Processing

An ablation study was conducted to prove the effect of the proposed preprocessing method. Figure 4 is the ablation study result. The vertical axis is the ratio of images, whose Jaccard index measuring RoI and ground truth exceeds the corresponding abscissa value.

FIGURE 4
www.frontiersin.org

Figure 4 Ablation study to prove effect of proposed preprocessing procedure.

Furthermore, Table 1 lists the numbers of RoIs achieved by the region proposal method with and without the preprocessing procedure. The results suggest that the proposed preprocessing procedure can reduce a total of RoIs and save the subsequent computation cost.

TABLE 1
www.frontiersin.org

Table 1 Statistics of RoIs achieved by the region proposal method with and without the preprocessing procedure.

RoI Extraction and Bounding-Box Regression Results

Figure 5 shows the results of RoI extraction, in which (A) is the original image, (B) is the edges obtained by the Canny edge detector, (C) is the RoIs obtained by selective search, and (D) is the RoIs on the original image. Figure 6 shows the results of the bounding-box regression for four samples. The green box is the ground truth in each sub-figure, and the red gradient boxes are the bounding-box regression results. The darkest red box is the initial bounding box, and the lightest red box is the final iteratively regressed bounding box. The results suggest that the iteration method can constantly extend the intersection area of the regressed box and the ground truth.

FIGURE 5
www.frontiersin.org

Figure 5 Step-by-step process of RoI extraction. (A) Original image. (B) Edges obtained by Canny edge detector. (C) RoIs obtained by selective search. (D) RoIs on the original image.

FIGURE 6
www.frontiersin.org

Figure 6 (A–D) Four examples of bounding-box regression iterations.

Prediction Performance

The proposed BUSnet was testified by the dataset in (22). The two-stage methods included Faster R-CNN (26) and CASCADE R-CNN (27), and the single-stage method included RetinaNet (28), YoLo V2 (29), and a single-shot detector (30). Note that these methods have been proven to be effective to objection detection in US images. Table 2 illustrates the performance metrics for the six methods, in which the best values are bolded. The results prove the excellent performance of the proposed BUSnet.

TABLE 2
www.frontiersin.org

Table 2 Performance metrics for six methods.

Figure 7 shows the prediction result when a normal US image is inputted into BUSnet. No lesion box is identified in the output.

FIGURE 7
www.frontiersin.org

Figure 7 Prediction for a normal US image. (A) Original image. (B) RoIs obtained after selective search. (C) Final output.

Discussion

US imaging is an effective technique for breast cancer screening. This paper proposes a deep learning model, BUSnet, to automatically detect the breast tumor lesions in US images. According to the characteristics of breast US images, the preprocess procedure and the iterative bounding-box regression were integrated in the proposed BUSnet, which were proved to be able to the improve the detection accuracy of lesions.

By far, ML and DL methods have been widely investigated to develop the AI-based US diagnosis techniques. Like the Bayesian network (31) and support vector machine (32), the classic ML methods were used as the classifier in two-stage lesion detection methods (14). proved that the DL methods can achieve better performance than the ML methods for breast US lesion detection. The well-developed DL models, like Faster R-CNN and YoLo, have been applied to breast US lesion detection (33). However, AI research and development for US images are still slow, compared with other modalities (4).

To improve the performance of DL-based lesion detection for breast US images, the preprocessing, the classification and bounding-box regression, and the post-processing should be integratedly considered. The imaging quality problems in US, like speckle noise, can disturb the classification of RoIs. Thus, the preprocessing procedure included the histogram equalization, normalization, and wavelet domain denoising. Because most breast tumors are located in the upper part of the US images (24), we cut out 30% of the area under the US images in the preprocessing procedure to save the subsequent computation burden. After preprocessing, we adopted a two-stage approach, including the region proposal and bounding-box regression, for the lesion detection. To effectively extract RoIs in the US images, we combined the Canny edge detector with the selective search in the region proposal stage. Because the ultrasound image is monochrome, the edge information is very important for detecting the lesion area. Thus, we first used a Canny edge detector to capture edges in images. Note that a Canny edge detector captures a huge number of edges in images. To tackle this problem, the selective search is used after the Canny edge detector to decrease the redundant edges. The RoIs obtained by selective search can reinforce the features of lesion for the further detection. According to (25), ResNet50 showed outstanding performance for the breast US images and acted as the backbone network for the classification of RoIs and bounding-box regression in the proposed BUSnet. The aggregation was applied as the post-processing procedure, which was proven to be able to exclude the background of US images from the lesions effectively. Our method discarded the bounding boxes identified as lesions with the probability less than 0.9. Thus, when a normal image is inputted into BUSnet, BUSnet will output no lesion box in the image.

The experimental results indicate that the proposed BUSnet worked well as expected. The ablation study shown in Figure 4 confirms the proposed preprocessing procedure. Figure 6 proves that the iteration regression strategy helped to continuously improve the bounding boxes. The comparison between our proposed BUSnet and other advanced methods is shown in Table 2. Overall, all the two-stage approaches, including our BUSnet, Faster R-CNN, and CASCADE R-CNN, performed better than the one-stage approaches, including Retina Net, YoLo V2, and SSD300. Furthermore, BUSnet achieved better performance than the other two-stage methods.

In this research, the dataset size is still small. Extending the capacity of the labeled dataset will be a crucial issue in the future. Furthermore, the semi-supervised learning and the transfer learning strategies will be considered for BUSnet for the incompletely labelled datasets and the various fields collected datasets, respectively.

Conclusions

US is an effective modality for the breast cancer screening. This paper proposed a deep learning model called BUSnet for the lesion detection in US images. BUSnet integrates preprocessing, region proposal, bounding-box regression, and post-processing. BUSnet can achieve satisfying lesion detection accuracy, which can be further applied to develop AI-based diagnostic support technologies for breast disease screening.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found as follows: https://scholar.cu.edu.eg/?q=afahmy/pages/dataset.

Author Contributions

HG and HW conceived the idea for this study. YL worked on the end-to-end implementation of the study. JW provided relevant insights on the clinical impact of the research work and handled the redaction of the paper. PQ managed the project and provided the funding for the research. All authors contributed to the article and approved the submitted version.

Funding

This work was supported by the National Natural Science Foundation of China (grant number 61633006 and 81872247) and the Fundamental Research Funds for the Central Universities, China (grant number DUT21YG118).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Lei S, Zheng R, Zhang S, Chen R, Wang S, Sun K, et al. Breast Cancer Incidence and Mortality in Women in China: Temporal Trends and Projections to 2030. Cancer Biol Med (2021) 18:900. doi: 10.20892/j.issn.2095-3941.2020.0523

CrossRef Full Text | Google Scholar

2. Coleman C. Early Detection and Screening for Breast Cancer. Semin Oncol Nurs (Elsevier) (2017) 33:141–55. doi: 10.1016/j.soncn.2017.02.009

CrossRef Full Text | Google Scholar

3. Xin Y, Zhang X, Yang Y, Chen Y, Wang Y, Zhou X, et al. A Multicenter, Hospital-Based and Non-Inferiority Study for Diagnostic Efficacy of Automated Whole Breast Ultrasound for Breast Cancer in China. Sci Rep (2021) 11:1–9. doi: 10.1038/s41598-021-93350-1

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Komatsu M, Sakai A, Dozen A, Shozu K, Yasutomi S, Machino H, et al. Towards Clinical Application of Artificial Intelligence in Ultrasound Imaging. Biomedicines (2021) 9:720. doi: 10.3390/biomedicines9070720

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Takahashi S, Takahashi M, Kinoshita M, Miyake M, Kawaguchi R, Shinojima N, et al. Fine-Tuning Approach for Segmentation of Gliomas in Brain Magnetic Resonance Images With a Machine Learning Method to Normalize Image Differences Among Facilities. Cancers (2021) 13:1415. doi: 10.3390/cancers13061415

PubMed Abstract | CrossRef Full Text | Google Scholar

6. LeCun Y, Bengio Y, Hinton G. Deep Learning. Nature (2015) 521:436–44. doi: 10.1038/nature14539

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Zhao ZQ, Zheng P, St XU, Wu X. Object Detection With Deep Learning: A Review. IEEE Trans Neural Networks Learn Syst (2019) 30:3212–32. doi: 10.1109/TNNLS.2018.2876865

CrossRef Full Text | Google Scholar

8. Lateef F, Ruichek Y. Survey on Semantic Segmentation Using Deep Learning Techniques. Neurocomputing (2019) 338:321–48. doi: 10.1016/j.neucom.2019.02.003

CrossRef Full Text | Google Scholar

9. Han S, Kang HK, Jeong JY, Park MH, Kim W, Bang WC, et al. A Deep Learning Framework for Supporting the Classification of Breast Lesions in Ultrasound Images. Phys Med Biol (2017) 62:7714. doi: 10.1088/1361-6560/aa82ec

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Byra M, Galperin M, Ojeda-Fournier H, Olson L, O’Boyle M, Comstock C, et al. Breast Mass Classification in Sonography With Transfer Learning Using a Deep Convolutional Neural Network and Color Conversion. Med Phys (2019) 46:746–55. doi: 10.1002/mp.13361

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Antropova N, Huynh BQ, Giger ML. A Deep Feature Fusion Methodology for Breast Cancer Diagnosis Demonstrated on Three Imaging Modality Datasets. Med Phys (2017) 44:5162–71. doi: 10.1002/mp.12453

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Tanaka H, Chiu SW, Watanabe T, Kaoku S, Yamaguchi T. Computer-Aided Diagnosis System for Breast Ultrasound Images Using Deep Learning. Phys Med Biol (2019) 64:235013. doi: 10.1088/1361-6560/ab5093

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Zhu YC, AlZoubi A, Jassim S, Jiang Q, Zhang Y, Wang YB, et al. A Generic Deep Learning Framework to Classify Thyroid and Breast Lesions in Ultrasound Images. Ultrasonics (2021) 110:106300. doi: 10.1016/j.ultras.2020.106300

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Yap MH, Pons G, Martí J, Ganau S, Sentís M, Zwiggelaar R, et al. Automated Breast Ultrasound Lesions Detection Using Convolutional Neural Networks. IEEE J Biomed Health Inf (2017) 22:1218–26. doi: 10.1109/JBHI.2017.2731873

CrossRef Full Text | Google Scholar

15. Kumar V, Webb JM, Gregory A, Denis M, Meixner DD, Bayat M, et al. Automated and Real-Time Segmentation of Suspicious Breast Masses Using Convolutional Neural Network. PloS One (2018) 13:e0195816. doi: 10.1371/journal.pone.0195816

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Alom MZ, Taha TM, Yakopcic C, Westberg S, Sidike P, Nasrin MS, et al. A State-of-the-Art Survey on Deep Learning Theory and Architectures. Electronics (2019) 8:292. doi: 10.3390/electronics8030292

CrossRef Full Text | Google Scholar

17. Hariharan B, Arbeláez P, Girshick R, Malik J. Springer, Cham: Simultaneous Detection and Segmentation, In: European Conference on Computer Vision. Zürich, Switzerland: Springer (2014). p. 297–312.

Google Scholar

18. Canny J. A Computational Approach to Edge Detection, In: IEEE Transactions on Pattern Analysis and Machine Intelligence. (1986) PAMI-8(6):679–98. doi: 10.1109/TPAMI.1986.4767851

CrossRef Full Text | Google Scholar

19. Uijlings JR, Van De Sande KE, Gevers T, Smeulders AW. Selective Search for Object Recognition. Int J Comput Vision (2013) 104:154–71. doi: 10.1007/s11263-013-0620-5

CrossRef Full Text | Google Scholar

20. Najibi M, Rastegari M, Davis LS. G-cnn: An Iterative Grid Based Object Detector, in: Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas: IEEE (2016). p. 2369–77.

Google Scholar

21. Bodla N, Singh B, Chellappa R, Davis LS. Soft-nms–Improving Object Detection With One Line of Code, In: 2017 Proceedings of the IEEE International Conference on Computer Vision. IEEE: Hawaii (2017). 5561–9 p.

Google Scholar

22. Al-Dhabyani W, Gomaa M, Khaled H, Fahmy A. Dataset of Breast Ultrasound Images. Data Brief (2020) 28:104863. doi: 10.1016/j.dib.2019.104863

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Yap MH, Goyal M, Osman F, Marti R, Denton E, Juette A, et al. Breast Ultrasound Region of Interest Detection and Lesion Localisation. Artif Intell Med (2020) 107:101880. doi: 10.1016/j.artmed.2020.101880

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Yap MH, Edirisinghe EA, Bez HE. Fully Automatic Lesion Boundary Detection in Ultrasound Breast Images. In: Medical Imaging 2007: Image Processing. San Diego, CA: International Society for Optics and Photonics (2007). 6512:p. 65123I.

Google Scholar

25. Gómez-Flores W, de Albuquerque Pereira WC. A Comparative Study of Pre-Trained Convolutional Neural Networks for Semantic Segmentation of Breast Tumors in Ultrasound. Comput Biol Med (2020) 126:104036. doi: 10.1016/j.compbiomed.2020.104036

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Ren S, He K, Girshick R, Sun J. Faster R-CNN: Towards Real-Time Object Detection With Region Proposal Networks. IEEE Trans Pattern Anal Mach Intell (2017) 39(6):1137–49. doi: 10.1109/TPAMI.2016.2577031.

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Cai Z, Vasconcelos N. Cascade R-CNN: Delving Into High Quality Object Detection, In: Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition: Salt Lake City, Utah: IEEE (2018). 6154–62 p.

Google Scholar

28. Lin TY, Goyal P, Girshick R, He K, Dollár P. Focal Loss for Dense Object Detection, In: Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition. Hawaii: IEEE (2017). 2980–8 p.

Google Scholar

29. Redmon J, Farhadi A. Yolo9000: Better, Faster, Stronger, In: Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition. Hawaii: IEEE (2017). 7263–71 p.

Google Scholar

30. Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY, et al. Ssd: Single Shot Multibox Detector, In: European Conference on Computer Vision. Amsterdam: Springer (2016). 21–37 p.

Google Scholar

31. Drukker K, Giger ML, Horsch K, Kupinski MA, Vyborny CJ, Mendelson EB. Computerized Lesion Detection on Breast Ultrasound. Med Phys (2002) 29:1438–46. doi: 10.1118/1.1485995

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Pons G, Martí R, Ganau S, Sentís M, Martí J. (2013). Feasibility Study of Lesion Detection Using Deformable Part Models in Breast Ultrasound Images, In: Iberian Conference on Pattern Recognition and Image Analysis. Berlin Heidelberg: Springer Berlin Heidelberg (2013).269–76 p.

Google Scholar

33. Cao Z, Duan L, Yang G, Yue T, Chen Q. An Experimental Study on Breast Lesion Detection and Classification From Ultrasound Images Using Deep Learning Architectures. BMC Med Imaging (2019) 19:1–9. doi: 10.1186/s12880-019-0349-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: breast ultrasound, lesion detection, deep learning, unsupervised pre-processing, bounding-box regression

Citation: Li Y, Gu H, Wang H, Qin P and Wang J (2022) BUSnet: A Deep Learning Model of Breast Tumor Lesion Detection for Ultrasound Images. Front. Oncol. 12:848271. doi: 10.3389/fonc.2022.848271

Received: 09 January 2022; Accepted: 23 February 2022;
Published: 25 March 2022.

Edited by:

Po-Hsiang Tsui, Chang Gung University, Taiwan

Reviewed by:

Mohammad Rahman, Mawlana Bhashani Science and Technology University, Bangladesh
Prabhakar T, GMR Institute of Technology, India

Copyright © 2022 Li, Gu, Wang, Qin and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Pan Qin, qp112cn@dlut.edu.cn; Jia Wang, wangjia77@hotmail.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.