Skip to main content

A hybrid features fusion-based framework for classification of breast micronodules using ultrasonography

Abstract

Background

Breast cancer is one of the leading diseases worldwide. According to estimates by the National Breast Cancer Foundation, over 42,000 women are expected to die from this disease in 2024.

Objective

The prognosis of breast cancer depends on the early detection of breast micronodules and the ability to distinguish benign from malignant lesions. Ultrasonography is a crucial radiological imaging technique for diagnosing the illness because it allows for biopsy and lesion characterization. The user’s level of experience and knowledge is vital since ultrasonographic diagnosis relies on the practitioner’s expertise. Furthermore, computer-aided technologies significantly contribute by potentially reducing the workload of radiologists and enhancing their expertise, especially when combined with a large patient volume in a hospital setting.

Method

This work describes the development of a hybrid CNN system for diagnosing benign and malignant breast cancer lesions. The models InceptionV3 and MobileNetV2 serve as the foundation for the hybrid framework. Features from these models are extracted and concatenated individually, resulting in a larger feature set. Finally, various classifiers are applied for the classification task.

Results

The model achieved the best results using the softmax classifier, with an accuracy of over 95%.

Conclusion

Computer-aided diagnosis greatly assists radiologists and reduces their workload. Therefore, this research can serve as a foundation for other researchers to build clinical solutions.

Peer Review reports

Introduction

Currently, the most malignant tumor in the world is breast cancer, which is a highly heterogeneous tumor [1, 2]. The biomarker expression assessed by immunohistochemistry (IHC) was used by the 2013 St. Gallen to divide breast cancer into five subtypes [3]. Breast cancer biomarker expression and various subtypes are significant prognostic factors [4]. After lung, stomach, liver, and colon cancers, breast cancer is the fifth most prevalent cause of cancer-related fatalities in the modern era. It is the leading cause of cancer death in women [5]. In 2005 alone, 519,000 deaths from breast cancer were reported [6]. This indicates that breast cancer caused one in every 100 fatalities globally and nearly one in every fifteen deaths from cancer. Among females, one in four cancer diagnoses is breast cancer, and one in six cancer fatalities is related to breast cancer [7]. Breast cancer also carries a hefty financial cost. The yearly cost of healthcare for women with breast cancer is $13,000 more than that of those without the disease [8]. In 2020, there were 10.9 million cancer-related deaths (9.9 million excluding NMSC, excluding basal cell carcinoma) and 19.3 million new cases of cancer (18.1 million excluding NMSC, excluding basal cell carcinoma) worldwide [7].

A tumor can be classified as malignant (cancerous) or benign (uncancerous). Benign tumors are not harmful because they do not spread cancer; their cells develop slowly, resemble normal tissue, and do not invade nearby tissues or injure other body components. On the other hand, cancerous tumors pose a threat. They eventually grow larger than the initial tumor and target other body areas if they are not treated [9].

One crucial non-radiation imaging method for identifying and categorizing breast tumors is breast ultrasonography. Patients tolerate it well and integrating it into interventional procedures for patient treatments is a simple process [10]. Breast US accuracy depends on the operator’s technical proficiency and experience; however, it is still limited. A standardized vocabulary and reporting system for evaluating breast mass and characterizing its characteristics are provided by the Breast Imaging Reporting and Data System (BI-RADS). Although several BI-RADS US descriptors are linked to both benign and malignant lesions, especially category 4 breast masses, it has been demonstrated to be an excellent approach for discriminating between benign and malignant masses [11].

The identification of breast tumors using combined B-mode and ultrasound elastography is currently limited. The experience of the doctors has a major impact on how ultrasounds operate [11]. It is indisputable that measurement errors result from differences in probe placement/orientation and annotation across and among observers [12, 13]. Furthermore, it can be challenging to determine the boundary between benign and malignant lesions as well as between normal and tumor tissue. Necrosis and liquefaction in malignant lesions or mechanization or calcification in benign lesions might compromise the accuracy of the malignancy rating method [14, 15].

Many computer-aided diagnosis (CAD) systems have been created in the literature for different medical diagnoses [16,17,18] and to help diagnose breast cancer patients by differentiating between benign and malignant tumors on images [19,20,21]. These methods have previously been shown to improve diagnostic precision and reduce observer variability [22,23,24]. Feature extraction, selection, and classification are involved in the classification process when utilizing conventional CAD systems [25, 26]. Effective feature extraction is the main challenge with these systems, as noted by [27], and it has an impact on total performance. Radiologists manually segmented the region of interest (ROI) without using any pre-processing techniques, as reported by Moon et al. [28]. Following various pre-processing methods, manual segmentation was performed in the work presented in [29]. The GAD and response diffusion (RD) based level set segmentation were combined by Zhang et al. [30]. Yu et al.‘s [31] important addition was the pre-processing procedures they took, which included dyadic wavelet transform, active contour, and k-means clustering. A suggested method [32] used a deep learning architecture and included feature extraction, segmentation, and classification, with a focus on Convolutional Neural Networks (CNNs). A variety of CNN models, such as Xception [19], InceptionV3 [21], InceptionNesNetV2 [33], DenseNet1 [20], DenseNet161 [34], and NASNetMobile [20], were tested and compared by Fujioka et al. [35]. Misra et al. [36], in contrast, used two CNN models (AlexNet [37] and ResNet [38]) and used ensemble learning to combine the models with ultrasound modalities (B-mode and SWE). However, even though CNN was used, Zhang et al. [39] and Zhou et al. [40] configured the feature extraction and classifier independently. Many of the recent studies [41,42,43] have also performed the diagnosis of diseases using various deep-learning techniques and achieved good results in this area. Identifying significant elements in an image requires a lot of work and effort. Furthermore, it might be difficult to optimize conventional CAD’s overall performance.

In this research, we have proposed a Hybrid Convolutional Neural Network (HCNN) system for the diagnosis of benign and malignant breast cancer lesions. The models InceptionV3 and MobilenetV2 serve as the foundation for the hybrid framework. These models’ characteristics are retrieved and concatenated individually. As a result, more features can be utilized. Finally, various classifiers are applied to the classification task.

The remaining structure of the manuscript is as follows: the second section contains details about the dataset as well as the detailed research techniques. The third section highlights the important results of the manuscript along with a discussion of the given results. Finally, the fourth section concludes the research.

Materials and methods

Dataset

One of the most crucial techniques for determining the sensitivity and efficacy of a screening modality is ultrasonography. The dataset used in this work is publicly available and was published in 2020 [44]. Breast ultrasound images were collected from women between the ages of 25 and 75 as part of the baseline data collection which was collected in 2018. There are 600 female patients in total. The collection contains 780 images, with an average pixel size of 500 by 500. The images are in PNG format. Original images are displayed with the ground truth images. The images are divided into three categories: normal, benign, and malignant. The benign class contains 437 images, malignant contains 210 images, and the normal contains 133 images. Fig. 1 shows the sample images from the ultrasonography images dataset.

Fig. 1
figure 1

Sample images from dataset

Methodology

The initial step involved preprocessing the images which is followed by applying the transfer learning techniques to extract relevant features using the Inception V3 and MobileNet V2 models. Next, a serial-based fusion approach was employed to combine the extracted features which enhanced the model capability to capture the enriched patterns. The fused features are then used for classification in the final step. The overall methodology is illustrated in Fig. 2.

Fig. 2
figure 2

Methodology for Classification of Breast Ultrasonography

In the pre-processing step, the number of images is increased through the data augmentation task and the image quality is also enhanced.

Dataset preprocessing

Images from breast ultrasonography are included in the dataset, which is utilized for testing and training. A few preparation processes must be completed before the dataset is passed to the models. The steps taken as part of the dataset preparation are listed below:

  1. 1.

    Initially, the dataset is divided into various ratios for training and testing, i.e. 90/10, 80/20, 70/30, and 60/40. This step helps the model to generalize the data and provide the actual performance of the model using variation in the train-to-test ratio.

  2. 2.

    Images are resized to match the models’ input dimensions.

  3. 3.

    A special filter is applied to remove additional information from the images, such as lines etc. In this step, we used a long horizontal probing element to remove the vertical lines from the images and a long vertical probing element to remove the horizontal lines from the images.

  4. 4.

    Data augmentation is applied using geometric transformations. The geometric augmentations applied include rotation with multiple values, scaling, shifting width and height, and zooming in as well as out. After applying data augmentation, images increased to 500 images in each class.

  5. 5.

    To ensure that the pixel values fall within a range that works with the selected model, normalization is applied. While pre-trained models are designed to handle images in the range of (0, 255), scratch CNN models are built to use input in the range of (0, 1) floating point.

  6. 6.

    The contrast of the images is enhanced and stretched using the gray-level contrast stretching technique.

  7. 7.

    It was confirmed that preprocessing methods did not alter the dataset images and that they remain relevant.

Pre-trained deep models

Because of their excellent performance over the last several years—state-of-the-art networks have gone from having ten layers to over a thousand layers—CNNs have become increasingly complex. By using some of these cutting-edge architectures that have already been trained on ImageNet, we applied transfer learning from natural images to images from breast cancer ultrasonography in this study. The main advantage of MobileNet V2 is that it handles the gradient vanishing problem much better; moreover, it is also a lightweight model. Furthermore, the motivation for choosing the Inception V3 model is that it also reduces the parameters set and is also helpful in the case of smaller datasets.

Inception V3

Research suggests that the Inception V3 model may be effectively utilized for the identification and categorization of novel images by modifying the fully connected layer design and preserving the convolution layer parameters. On the ImageNet dataset, the image recognition model Inception V3 has demonstrated accuracy levels above 78.1%. The model is the result of several concepts that have been developed over time by various scholars.

In terms of object recognition, the Inception V3 model performs better than GoogleNet’s Inception V1. The enhanced Inception module, the classifier, and the fundamental convolutional block make up the Inception V3 model. For feature extraction, the fundamental convolutional block that alternates between convolutional and max-pooling layers is utilized. The improved Startup module is based on Intra-Mesh, which further cascades the convolution results of each branch and performs multi-scale convolution in parallel. The auxiliary group is used to improve gradient convergence and produce consistent training results while also minimizing overfitting and underfitting problems. Inception V3 mainly uses a 1*1 convolution kernel to reduce the number of unique channels and speed up training. Additionally, dividing the large convolution into smaller convolutions reduces the computational cost and number of parameters. In conclusion, Inception V3’s state-of-the-art object identification performance is a result of its distinctive Inception architecture. Therefore, this approach is frequently employed for transfer learning.

Basic convolutional blocks enhanced Inception modules, and task-specific classifiers are cascaded based on the Inception-v3 model. Particularly low-performance maps were learned by the ensemble method using 1 × 1 and 3 × 3 kernels. In the initial module, the multi-scale feature representation is combined and fed to an auxiliary classifier using a set of convolution kernels (i.e. 1 × 1, 1 × 3, 3 × 1, 3 × 3, 5 × 5, 1 × 7, and 7 × 7 filters to improve convergence). Eleven inception modules act as a linked process and transform multiple performance vectors into a single vector. Finally, the Softmax classifier produces a heat vector. Figure 3 shows the Inception v3 architecture. The model extracts a total of 2048 features.

Fig. 3
figure 3

Inception V3 Architecture. The model takes an input in the format 299 × 299 × 3 and outputs the 8 × 8 × 2048. Inception V3 introduced the factorized 7 × 7 convolutions as well as batch normalization in the auxiliary classifier and performed label smoothing

MobileNet V2

MobileNet is a CNN-based model that is widely used for image classification, in contrast to MobileNet V2. The primary benefit of utilizing the MobileNet architecture is that, in comparison to the traditional CNN model, it requires significantly less computing power, which makes it appropriate for use with mobile devices and PCs with less processing power. The convolution layer of the MobileNet model, which is a streamlined structure, is useful for differentiating details based on two controllable parameters that efficiently flip between latency and accuracy. Reducing the size of the network is a benefit of the MobileNet architecture.

The architecture of MobileNet is as effective while using as few resources as possible, including Palmprint Recognition. MobileNet is designed in a depth-wise manner. The basic structure is built on several abstraction layers, which are a part of various convolutions that appear to be the quantized configuration that thoroughly evaluates the difficulty of a typical problem. Point-wise complexity is the complexity of a 1 × 1 convolution. In-depth platforms are made with abstraction layers that pass through a conventional rectified linear unit (ReLU) and feature deep structural elements. To reduce the dimensionality of the input image and the internal representation of each layer, the resolution multiplier variable 𝜔 is applied.

Depending on the context, two hyperparameters, width multipliers and a resolution multiplier help translate the ideal viewing area into a more accurate estimate. The recommended input image size in the model is 224 × 224 × 3. This value must be greater than 32. It can be seen that there are three dimensions as the third value. The size of each of the 32 filters in the design is 3 × 3 × 3 × 32.

The idea behind MobileNet architectures is to replace complex convolutional layers with simpler ones. Each layer consists of a size 3 × 3 convolutional layer that buffers the input data and a size 1 × 1 pointwise convolutional layer that uses these filtered parameters to create a new component, as seen in Fig. 4. Simplifying the model and making it quicker than the standard convolutional model is the idea outlined above. There are 1,210 features that have been extracted from this model in total.

Fig. 4
figure 4

MobileNet V2 Architecture. The model takes an input of 224 × 224 × 3, and it is based on inverted residual connections between the bottleneck layers. The model in intermediately uses the depth-wise convolutions as a source of non-linearity

After the features of both models are extracted, these features are fused using a serial-based approach. The primary benefit of serial-based fusion is that it produces significant information reduction and enables real-time processing. The total features after fusion are 2,458. For reliable prediction, these fused features are then fed through a few classifiers, including Softmax, linear support vector machine (SVM), and Bayesian. The Softmax classifier utilizes cross-entropy loss for its computations. The Softmax classifier is named after the Softmax function, which transforms raw class scores into normalized positive values that sum to one, making it possible to apply cross-entropy loss effectively. SVMs are effective across a wide range of tasks because they can handle high-dimensional data and capture non-linear relationships efficiently. The success of SVMs lies in their ability to find the optimal hyperplane that maximizes the separation between different classes in the dataset. Bayesian classifiers rely on Bayes’ decision theory and use fundamental statistical principles. The core idea is that if the class is known, the values of the other characteristics may be predicted. Conversely, if the class is unknown, Bayes’ rule can be applied to predict the class label based on the provided feature values. These classifiers employ probabilistic models to determine the class label for new samples.

Results and discussion

The dataset is split into various train-to-test ratios to facilitate experimentation. Keras, PyTorch, and Matplotlib are the primary tools and libraries used in this process. This study evaluates the proposed method using several performance criteria commonly used in breast ultrasonography categorization techniques. Accuracy and F-measure are the performance measures. Accuracy is a reliable assessment metric for evenly distributed, non-skewed classification tasks without any class imbalance. Generally speaking, accuracy can dangerously show exaggerated and overly optimistic results, particularly in imbalanced datasets. Since the F-measure is the harmonic mean of recall and accuracy, it maintains the classifier’s balance between the two [45].

The results are collected on different train-to-test ratios. Initially, results are collected on a 90 to 10 train/test ratio as shown in Table 1. The best results are achieved using the Softmax classifier with 81.26% F-measure and 82.31% accuracy. However, the Bayesian classifier produced the second-best results.

Table 1 Classification performance on 90/10 Train/Test ratio

The results collected on an 80 to 20 train/test ratio are shown in Table 2. The best results are achieved using the softmax classifier with 83.71% F-measure and 85.62% accuracy. The performance of the Bayesian classifier can’t be overlooked as it produced the second-best results.

Table 2 Classification performance on 80/20 Train/Test ratio

The results obtained on a 70 to 30 train/test ratio are shown in Table 3. The best results are achieved using the Softmax classifier with 93.99% F-measure and 95.82% accuracy. In a 70/30 ratio distribution, we can see that linear SVM produced the second-best results, and the results are notable; therefore, the performance of this classifier cannot be ignored.

Table 3 Classification performance on 70/30 Train/Test ratio

The results obtained on a 60 to 40 train/test ratio are shown in Table 4. The best results are achieved using the Softmax classifier with 83.99% F-measure and 95.82% accuracy. In a 70/30 ratio distribution, we can see that linear SVM produced the second-best results.

Table 4 Classification performance on 60/40 Train/Test ratio

As we have collected the results on different train/test ratios, we can see that the results achieved with a 70/30 train/test ratio are the best, while those with a 90/10 train/test ratio are the worst. This is because the model might be overfitted with a 90/10 train/test ratio. The best results are achieved with a 70/30 train/test ratio because the model gets the best fit and provides a perfect balance between bias and variance. This happens because the 70/30 split shows its effectiveness by training the model on a large portion of the dataset while still maintaining sufficient data for validation/testing. Moreover, we can see that the softmax classifier performed best overall. The softmax classifier provides a well-defined probability distribution as it reacts to low simulations with a normal distribution and high simulations with near to zero or one.

To compare the train and test accuracies among the 70/30 train/test ratio, which also provided the best results, a comparative graph is shown in Fig. 5. Upon analyzing the graph, it is evident that initially, the model accuracy increased along with epochs. Moreover, training and testing accuracies did not differ significantly throughout the training process, which shows the model is generalizing well. Another thing to note is that the model accuracies remained constant after 250 epochs so it is justified to stop the training at 300 epochs otherwise model may start overfitting on the data.

Fig. 5
figure 5

Accuracy Comparison on 70/30 Train/Test Ratio

Ablation study

To compare the results on individual models, we have performed ablation studies. Initially, we gathered the results separately for both Inception V3 and MobileNet V2 models. Table 5 shows the classification results obtained on the Inception V3 model. It can be seen that accuracy is not up to par.

Table 5 Classification performance on Inception V3 Model

The classification performance on the MobileNet V2 model is shown in Table 6.

Table 6 Classification performance on MobileNet V2 Model

The results without the preprocessing steps are shown in Table 7 which provides a detailed analysis and highlights the need for the steps.

Table 7 Classification performance without Preprocessing Step

Fig. 6 illustrates the accuracy comparison obtained in this work. It is evident that after feature fusion, the performance of the model improved significantly. The model can be used to explore further research and for initial clinical trials. In future work, exploring further research as well as improving the explainability of the model may be considered.

Fig. 6
figure 6

Classification Performance Comparison

Conclusion

Breast cancer has a high occurrence worldwide and is life-threatening. Computer-aided diagnosis greatly helps radiologists and reduces their workload. Therefore, in this research, we have proposed a hybrid CNN system for the diagnosis of benign and malignant breast cancer lesions. The models InceptionV3 and MobilenetV2 serve as the foundation for the hybrid framework. These models’ characteristics are extracted and concatenated individually. As a result, more features are utilized. Finally, various classifiers are applied to the classification task. The model has achieved the best results using a softmax classifier and a 70/30 train/test ratio. In future work, we will explore explainable AI techniques on the data present the findings. We will also investigate multi-modal data.

Data availability

The dataset analyzed during the current study is available in the [https://www.kaggle.com/datasets/aryashah2k/breast-ultrasound-images-dataset] repository.

References

  1. Siegel RL, Miller KD, Jemal A. Cancer statistics, 2018. Cancer J Clin. 2018;68:7–30.

    Article  Google Scholar 

  2. Zhang L, Li J, Xiao Y, Cui H, Du G, Wang Y, et al. Identifying ultrasound and clinical features of breast cancer molecular subtypes by ensemble decision. Sci Rep. 2015;5:11085.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  3. Goldhirsch A, Winer EP, Coates A, Gelber R, Piccart-Gebhart M, Thürlimann B, et al. Personalizing the treatment of women with early breast cancer: highlights of the St Gallen International Expert Consensus on the primary therapy of early breast Cancer 2013. Ann Oncol. 2013;24:2206–23.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  4. Zhou J, Jin A-q, Zhou S-c, Li J-w, Zhi W-x, Huang Y-x, et al. Application of preoperative ultrasound features combined with clinical factors in predicting HER2-positive subtype (non-luminal) breast cancer. BMC Med Imaging. 2021;21:1–13.

    Article  Google Scholar 

  5. Majnaric L, Vcev A. Prevention and early detection of Cancer–A Public Health View. Cancer Management. ed: IntechOpen; 2012.

  6. Gokhale S. Ultrasound characterization of breast masses. Indian J Radiol Imaging. 2009;19:242–7.

    Article  PubMed  PubMed Central  Google Scholar 

  7. Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram I, Jemal A, et al. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. Cancer J Clin. 2021;71:209–49.

    Article  Google Scholar 

  8. Mao Y-J, Lim H-J, Ni M, Yan W-H, Wong DW-C, Cheung JC-W. Breast Tumour Classification Using Ultrasound Elastography with Machine Learning: A Systematic Scoping Review, Cancers, vol. 14, p. 367, 2022.

  9. Mobark N, Hamad S, Rida SZ. CoroNet: deep neural network-based end-to-end training for breast Cancer diagnosis, Applied Sciences, 12, p. 7080, 2022.

  10. Merritt C. Combined screening with Ultrasound and Mammography vs Mammography alone in women at elevated risk of breast Cancer. Breast Diseases: Year Book Q. 2009;4:401–2.

    Google Scholar 

  11. Liberman L, Menell JH. Breast imaging reporting and data system (BI-RADS). Radiologic Clin. 2002;40:409–30.

    Google Scholar 

  12. Jiang W-w, Li A-h, Zheng Y-P. A semi-automated 3-D annotation method for breast ultrasound imaging: system development and feasibility study on phantoms. Ultrasound Med Biol. 2014;40:434–46.

    Article  PubMed  Google Scholar 

  13. Samir AE, Dhyani M, Vij A, Bhan A. K., Halpern E. F., Méndez-Navarro J, et al. Shear-wave elastography for the estimation of liver fibrosis in chronic liver disease: determining accuracy and ideal site for measurement. Radiology. 2015;274:888–96.

    Article  PubMed  Google Scholar 

  14. Kerridge WD, Kryvenko ON, Thompson A, Shah BA. Fat necrosis of the breast: a pictorial review of the mammographic, ultrasound, CT, and MRI findings with histopathologic correlation, Radiology research and practice, vol. 2015, 2015.

  15. Bae MS, Han W, Koo HR, Cho N, Chang JM, Yi A, et al. Characteristics of breast cancers detected by ultrasound screening in women with negative mammograms. Cancer Sci. 2011;102:1862–7.

    Article  CAS  PubMed  Google Scholar 

  16. Noor MN, Ashraf I, Nazir M. Analysis of GAN-Based Data Augmentation for GI-Tract Disease classification, in. Advances in Deep Generative models for Medical Artificial Intelligence. ed: Springer; 2023. pp. 43–64.

  17. Noor MN, Nazir M, Ashraf I, Almujally NA, Aslam M, Fizzah Jilani S. GastroNet: a robust attention-based deep learning and cosine similarity feature selection framework for gastrointestinal disease classification from endoscopic images. CAAI Trans Intell Technol, 2023.

  18. Nouman Noor M, Nazir M, Ashraf I, Song O-Y. Localization and classification of gastrointestinal tract disorders using explainable AI from endoscopic images. Appl Sci. 2023;13:9031.

    Article  CAS  Google Scholar 

  19. Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D et al. ,., Going deeper with convolutions, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 1–9.

  20. Zoph B, Vasudevan V, Shlens J, Le QV. Learning transferable architectures for scalable image recognition, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 8697–8710.

  21. Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z. Rethinking the inception architecture for computer vision, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 2818–2826.

  22. Huang Y-L, Chen D-R. Support vector machines in sonography: application to decision making in the diagnosis of breast cancer. Clin Imaging. 2005;29:179–84.

    Article  PubMed  Google Scholar 

  23. Singh S, Maxwell J, Baker JA, Nicholas JL, Lo JY. Computer-aided classification of breast masses: performance and interobserver variability of expert radiologists versus residents, Radiology, vol. 258, pp. 73–80, 2011.

  24. Fallenberg E, Dromain C, Diekmann F, Engelken F, Krohn M, Singh J et al. ,Contrast-enhanced spectral mammography versus MRI: initial results in the detection of breast cancer and assessment of tumour size, European radiology, vol. 24, pp. 256–264, 2014.

  25. Yang M-C, Moon WK, Wang Y-CF, Bae MS, Huang C-S, Chen J-H, et al. Robust texture analysis using multi-resolution gray-scale invariant features for breast sonographic tumor diagnosis. IEEE Trans Med Imaging. 2013;32:2262–73.

    Article  Google Scholar 

  26. Han S, Kang H-K, Jeong J-Y, Park M-H, Kim W, Bang W-C, et al. A deep learning framework for supporting the classification of breast lesions in ultrasound images. Phys Med Biol. 2017;62:7714.

    Article  PubMed  Google Scholar 

  27. Newell D, Nie K, Chen J-H, Hsu C-C, Yu HJ, Nalcioglu O, et al. Selection of diagnostic features on breast MRI to differentiate between malignant and benign lesions using computer-aided diagnosis: differences in lesions presenting as mass and non-mass-like enhancement. Eur Radiol. 2010;20:771–81.

    Article  PubMed  Google Scholar 

  28. Moon WK, Choi JW, Cho N, Park SH, Chang JM, Jang M, et al. Computer-aided analysis of ultrasound elasticity images for classification of benign and malignant breast masses. Am J Roentgenol. 2010;195:1460–5.

    Article  Google Scholar 

  29. Wu J-X, Chen P-Y, Lin C-H, Chen S, Shung KK. Breast benign and malignant tumors rapidly screening by ARFI-VTI elastography and random decision forests based classifier. IEEE Access. 2020;8:54019–34.

    Article  Google Scholar 

  30. Zhang Q, Song S, Xiao Y, Chen S, Shi J, Zheng H. Dual-mode artificially-intelligent diagnosis of breast tumours in shear-wave elastography and B-mode ultrasound using deep polynomial networks. Med Eng Phys. 2019;64:1–6.

    Article  PubMed  Google Scholar 

  31. Yu Y, Xiao Y, Cheng J, Chiu B. Breast lesion classification based on supersonic shear-wave elastography and automated lesion segmentation from B-mode ultrasound images. Comput Biol Med. 2018;93:31–46.

    Article  PubMed  Google Scholar 

  32. O’Mahony N, Campbell S, Carvalho A, Harapanahalli S, Hernandez GV, Krpalkova L et al. ,Deep learning vs. traditional computer vision, in Advances in Computer Vision: Proceedings of the 2019 Computer Vision Conference (CVC), Volume 1 1, 2020, pp. 128–144.

  33. Szegedy C, Ioffe S, Vanhoucke V, Alemi A. Inception-v4, inception-resnet and the impact of residual connections on learning, in Proceedings of the AAAI conference on artificial intelligence, 2017.

  34. Huang G, Liu Z, Van Der Maaten L, Weinberger KQ. Densely connected convolutional networks, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 4700–4708.

  35. Fujioka T, Katsuta L, Kubota K, Mori M, Kikuchi Y, Kato A, et al. Classification of breast masses on ultrasound shear wave elastography using convolutional neural networks. Ultrason Imaging. 2020;42:213–20.

    Article  PubMed  Google Scholar 

  36. Misra S, Jeon S, Managuli R, Lee S, Kim G, Yoon C, et al. Bi-modal transfer learning for classifying breast cancers via combined b-mode and ultrasound strain imaging. IEEE Trans Ultrason Ferroelectr Freq Control. 2021;69:222–32.

    Article  PubMed  Google Scholar 

  37. Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst, 25, 2012.

  38. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.

  39. Zhang X, Liang M, Yang Z, Zheng C, Wu J, Ou B, et al. Deep learning-based radiomics of b-mode ultrasonography and shear-wave elastography: improved performance in breast mass classification. Front Oncol. 2020;10:1621.

    Article  PubMed  PubMed Central  Google Scholar 

  40. Zhou Y, Xu J, Liu Q, Li C, Liu Z, Wang M, et al. A radiomics approach with CNN for shear-wave elastography breast tumor classification. IEEE Trans Biomed Eng. 2018;65:1935–42.

    Article  PubMed  Google Scholar 

  41. Noor MN, Nazir M, Dilshad V, Haneef F. A Framework for Multi-Grade Classification of Ulcerative-Colitis Using Deep Neural Networks, in 2023 25th International Multitopic Conference (INMIC), 2023, pp. 1–4.

  42. Carriero A, Groenhoff L, Vologina E, Basile P, Albera M. Deep Learning in Breast Cancer Imaging: State of the Art and Recent Advancements in Early 2024, Diagnostics, vol. 14, p. 848, 2024.

  43. Uysal F, Köse MM. Classification of breast cancer ultrasound images with deep learning-based models. Eng Proc. 2022;31:8.

    Google Scholar 

  44. Al-Dhabyani W, Gomaa M, Khaled H, Fahmy A. Dataset of breast ultrasound images. Data Brief. 2020;28:104863.

    Article  PubMed  Google Scholar 

  45. Noor MN, Nazir M, Ashraf I. Emerging Trends and advances in the diagnosis of gastrointestinal diseases. BioScientific Rev. 2023;5:118–43.

    Article  Google Scholar 

Download references

Acknowledgements

The author would like to extend his appreciation to King Saud University for funding this work through the Researchers Supporting Project number (RSP2024R458), King Saud University, Riyadh, Saudi Arabia.

Funding

The author would like to extend his appreciation to King Saud University for funding this work through the Researchers Supporting Project number (RSP2024R458), King Saud University, Riyadh, Saudi Arabia.

Author information

Authors and Affiliations

Authors

Contributions

M.A. was responsible for the conceptualization, methodology, validation, original draft writing, review and editing, and provided the funding acquisition.

Corresponding author

Correspondence to Mousa Alhajlah.

Ethics declarations

Ethical approval

This research is done on publicly available datasets that are already published and are therefore exempt from approval.

Consent for publication

The research is conducted on a publicly available anonymous dataset therefore this section is not applicable for this research.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Alhajlah, M. A hybrid features fusion-based framework for classification of breast micronodules using ultrasonography. BMC Med Imaging 24, 253 (2024). https://doi.org/10.1186/s12880-024-01425-y

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12880-024-01425-y

Keywords