Abstract

Breast cancer is the most frequent disease in women, with one in every 19 women at risk. Breast cancer is the fifth leading cause of cancer death in women around the world. The most effective and efficient technique of controlling cancer development is early identification. Mammography helps in the early detection of cancer, which saves lives. Many studies conducted various tests to categorize the tumor and obtained positive findings. However, there are certain limits. Mass categorization in mammography is still a problem, although it is critical in aiding radiologists in establishing correct diagnoses. The purpose of this study is to develop a unique hybrid technique to identify breast cancer mass pictures as benign or malignant. The combination of two networks helps accelerate the categorization process. This study proposes a novel-based hybrid approach, CNN-Inception-V4, based on the fusing of these two networks. Mass images are used in this research from the CBIS-DDSM dataset. 450 images are taken for benign, and 450 images are used for malignant. The images are first cleaned by removing pectoral muscles, labels, and white borders. Then, CLAHE is used to these images to improve their quality in order to produce promising classification results. Following preprocessing, our model classifies cancer in mammography pictures as benign or malignant abnormalities. Our proposed model’s accuracy is 99.2%, with sensitivity of 99.8%, specificity of 96.3%, and F1-score of 97%. We also compared our proposed model to CNN, Inception-V4, and ResNet-50. Our proposed model outperforms existing classification models, according to the results.

1. Introduction

Breast cancer becomes one of the most serious diseases and leading causes of mortality in women worldwide [1]. Breast cancer develops when aberrant cells in the breast proliferate excessively [2]. Breast cancer is classified into two types: benign and malignant. Benign is treatable, but it can become hazardous over time. Malignant is perilous [3]. It has the potential to spread to other organs of the body and cause death. In developed countries, the proportion of cancer cases is growing. This is a major issue in healthcare. As a result, early identification of breast cancer can lower the risk of patient death [4].

Mammography is indeed the good diagnostic imaging technique for detecting breast problems early on [5]. Cancer mortality rates could rise by 50% to 15 million by 2030, according to World Health Organization (WHO) predictions. Developing countries account for over 70% of new cancer cases. X-ray scans of each breast are acquired from two perspectives during a mammography screening. Human readers, such as an expert radiologist, assess and examine screening mammograms [6]. Dual reading was discovered to enhance evaluation results and has been implemented in many countries. Repeated readings with up to ten readers can boost diagnostic performance even further. Beyond the dual reading, there is much of room for improvement in mammography grading.

Based on the invasive phase of the breast anomaly, the irregularities may be benign or malignant. Detecting lumps in breast tissue is more difficult than detecting microcalcification. Many industrialised countries are adopting the smart cities concept in the modern era [7]. All of that in a smart city is controlled by a computer. Every industry in the world is moving toward the computer. Humans want computers to rule everything. In a smart city environment, decision-making is a significant difficulty [8]. This present era is sometimes referred to as the data world. Because consumers contribute data on a daily basis. This information is quite beneficial in terms of making the environment smarter. Data-driven decision AI (artificial intelligence) [9, 10] enables computers to make decisions. Data must be available to the computer in order for it to make conclusions.

Healthcare is a major global challenge [1115]. The use of computers, smart systems, and intelligent devices is critical in healthcare. The diagnosis of sickness at an early stage can save lives [16, 17]. New breakthroughs have the potential to lessen the risk to human life [18].

Computer-Aided Diagnosis (CAD) [19] systems are used to assist radiologists in the detection and diagnosis of cancer. CAD systems have been shown in studies to be useful; yet, accurate breast cancer detection remains a difficult issue [20]. Because of the intricacy of mammography and the large number of tests performed by radiologist, incorrect diagnoses occur. The CAD system attempts to reduce incorrect diagnoses, which in turn reduces unneeded biopsies, patient anxiety, increased health-care costs, and superfluous review by radiologists. Moving beyond traditional machine learning (ML) [21, 22], current efforts have focused on the application of deep learning (DL) approaches [2224]. Through statistical analysis, they are able to more correctly diagnose suspected lesions. DL models are intended for huge and diverse datasets (for example, the ImageNet dataset [25]), whereas mammography contains tiny datasets that are publically available. As a result, when learning from the start with minimal training samples, the capability and complexity of such networks might have a considerable negative impact on simulation results.

Many researchers played an important role in the classification of mass images of breast cancer as benign and malignant, but still, mass classification in mammography remains a significant issue, yet it is essential in assisting radiologists in making accurate diagnoses [26]. There is also a need for such a system that can speed up the process of classification of tumors with good results, and this is possible with a unique hybrid technique. So, using the CBIS-DDSM (Curated Breast Imaging Subset of DDSM) dataset [27], this research provides a novel hybrid approach CNN-Inception-V4 for classifying breast cancer as benign or malignant. The paper’s main contributions are based on the following points. (i)To propose such a unique strategy for properly classifying breast cancer(ii)To propose a hybrid technique that can speed up the process of breast cancer classification along with good results(iii)To perform techniques such as image preprocessing, data augmentation, and hyperparameters (e.g., batch size, learning rate), which are appealing choices for reducing overfitting and enhancing model generalization(iv)To compare our model to existing state-of-the-art networks

The other sections of the research are organized in the following manner: related work is done in Section 2, the proposed method and dataset used are described in Section 3, the classification results are evaluated in Section 4, and the conclusions are presented in the last section.

Many researchers used various strategies to reach promising findings. The authors in [28] employed computer vision techniques to help a radiologist evaluate mammography images and diagnose a malignancy in an acceptable amount of time. They extracted various characteristics from the mammogram’s selected area, which the surgeon subjectively labeled. These characteristics are supplied into a classification engine, which is then used to train and construct the suggested architecture classification models. The classification procedure by either SVM or NB with its optimum parameters resulted in 100% accuracy for the classification process. The authors in [29] proposed Fisher linear discrimination analysis (FLDA) and quadratic discrimination assessment (QDA) inside random projection (RP) filtration to evaluate our performance and efficiency with SVM classifier (SVM), -nearest neighbour (KNN), and multilayer perceptron (MLP) in IoT-based platforms. The modelling results demonstrate not only increased accuracy but also decreased duration. In our chosen dataset simulated for a relation to health and safety, runtime is reduced by 20 times when compared to SVM, KNN, and MLP in QDA and FLDA. At the IoT business, RP filtering in the preprocessing step of feature extraction, which fulfils the model’s runtime, is a point of view.

The authors in [30] proposed Blockchain-Integrated Remote Database (BIRD) paradigm that connects blockchain to a Relational Database Management System (RDBMS) that employs the Remote Database Access (RDA) protocol and a Cloud Server. This report gives a comprehensive history of blockchain and industry generations. Finally, this research proposes a next-generation blockchain architecture in which a smart government linked to Industry 4.0 monitors price increases and fraud.

The authors in [31] proposed synchronization technique based on AI to address smart grid timing difficulties. Backpropagation neural networks are used as an AI technology that utilizes time approximations and error corrections for exact results. In order to link the external time server, the unique AIFS system considers radio communication features. The suggested AIFS scheme’s performance is tested using a MATLAB-based simulation technique. The simulation results suggest that the proposed approach outperforms the present system.

The authors in [27] proposed a modified network of YOLOv5 to detect and categorize breast cancers. The process was carried out with an 8-batch size, an activation function of 0.01, a velocity of 0.843, and an interval value of 300. To evaluate the suggested effectiveness of the algorithm, it is contrasted to YOLOv3 and a faster RCNN, with 93.50% MCC value and 96.50% accuracy. The suggested model beats YOLOv3 and quicker RCNN. The results show that the proposed prediction model detects and diagnoses breast tumors while also addressing previous researchers’ limitations by decreasing the FPR and FNR and raising the MCC value.

The authors in [32] proposed an efficient, lightweight encryption method to provide a safe picture encryption approach for the healthcare business. To protect medical pictures, the proposed lightweight encryption solution uses two permutation algorithms. In terms of security and execution time, the suggested approach is investigated, assessed, and then compared to commonly encrypted ones. Several test photos were utilized to evaluate the proposed algorithm’s performance. Several studies demonstrate that the suggested algorithm for picture cryptosystems is more efficient than existing techniques.

The authors in [33] present a study of deep learning-based load forecasting strategies from 2015 to 2020. This study addresses research in the load forecasting process that uses deep learning techniques, Distributed Deep Learning (DDL) techniques, backpropagation- (BP-) based works, and non-BP-based works. Following the survey, it was concluded that data aggregation dependence would be advantageous for lowering computing time in load forecasting. As a result, a conceptual model of DDL for smart grids has been provided, with the HSIC (Hilbert-Schmidt Independence Criterion) bottleneck approach integrated to give more accuracy.

The authors in [32] offer an experimental analysis of cryptographic algorithms in order to categorize encryption methods as symmetric or asymmetric. It provides a thorough examination of AES, DES, 3DES, RSA, and Blowfish in terms of time complexity, size, encryption, and decryption performance. The simulation technique was used for the evaluation, and the speed of encryption and decryption of the chosen encryption algorithms was examined. The experiments ran the same encryption using the same plaintext five times for each encryption and decryption, and the average time is compared. Each encryption algorithm’s key size is the maximum number of bytes that the cypher may accept. The average time taken by the three devices to compute the method is utilized for comparison. In the simulation for the experimental test, a set of plaintexts—password-sized text and paragraph-sized text—achieves goal fair outcomes when compared to the existing methods in real-time deep learning networks for IoT applications.

The authors in [34] presented a model for breast cancer categorization, using the CAD technique. MIAS and DDSM datasets are used in this study. Preprocessing, categorization, gathering, and operation grouping are all used in this system. The CAD system includes a CNN model with eight coevolutionary, 4 max pooling, and two totally connected layers. The obtained findings are compared to AlexNet and VGG16, demonstrating that the suggested CNN [35] beat both of these models in terms of accuracy and AUC. The suggested model achieved accuracies of 92.54%, 96.47%, and 95% for MIAS, DDSM, and the self-collected dataset, respectively.

The authors in [36] developed an intelligent system that included picture normalization, microcalcification detection, and mammogram categorization. The best classification accuracy of the system in distinguishing between benign and malignant breasts can reach 0.8124 on the training set and 0.7237 on the test set, with an area under curve (AUC) of receiver operating characteristic of 0.8042. The sensitivity of malignant breast prediction is 0.8891 on the training dataset and 0.7778 on the testing dataset. The highest accuracy was 0.7511 on the microcalcification lesion degree, with an AUC of 0.7627 on the testing dataset.

The authors in [37] proposed a strategy to categorize breast invasive cancer using AGAN for data augmentation and CNN for tumor categorization. In many visual recognition applications, deep CNN models and other predictive analysis and ML algorithms have outperformed the state of the art. This method produced 89.17% accurate results but a 19.41% true alarm rate of classification.

The authors in [38] investigated the achievement of three types of ML methods: (1) NB, (2) NN, and (3) SVM. SVM delivers the most satisfactory results in correctly characterising the condition of the breast as “Good” or “Malignant,” with a reliability of 99.4% in training and 98.76% in testing, according to the outcomes of ten experiments. In the study, the SVM classifier outperformed the NN and NB models, suggesting that SVM is an appropriate choice for assessing the state of the breast at a preliminary phase.

The authors in [39] proposed a method to identify breast tumor utilizing two deep learning model networks, Res-Net-50 and VGG-16 using the IRMA dataset and then comparing their outcomes. They attained an accuracy of 94% for VGG-16 and 91.7% for Res-Net50, as well as a sensitivity of 94% for VGG-16 and 94% for Res-Net50. Their restriction is that for VGG16 and Res-Net50, their accuracy value lowers to 89% and 88%, respectively.

The authors in [40] proposed a hierarchical twitter sentiment model (HTSM) to show people’s opinions in short texts. The suggested model’s quality is confirmed by applying it to a popular product and a widely discussed issue. The findings reveal that the suggested model beats state-of-the-art approaches for successfully evaluating people’s opinions in brief text.

The authors in [41] provide a novel CNN structure for classifying histopathological cancer pictures that integrates spectral characteristics acquired via a multiresolution wavelet technique with CNN spatial features. Furthermore, batch normalization is employed after each layer in the convolution network to solve CNN’s poor convergence problem, and CNN’s deep layers are trained with spectral–spatial data. Using the BreaKHis dataset and the Breast Cancer Classification Challenge 2015 dataset, the suggested structure is evaluated on malignant histology pictures of the breast for both binary and multiclass tissue classifications. Experiment findings suggest that combining spectral–spatial data enhances CNN network classification accuracy and needs fewer training parameters than well-known models. On both datasets, the suggested structure achieves an average accuracy of 97.58% and 97.45% with 7.6 million training parameters.

The authors in [42] describe how the VGG16 framework of DL methodology was used to categorize breast cancer in mammography. The VGG16 classifier is constructed and tested using 322 photographs from the MIAS dataset. It beats the models AlexNet, EfficientNet, and GoogleNet. The categorization of mammograms will improve mammography screening effectiveness and act as a valuable mechanism for surgeons. The authors in [43] promptly made available breast cancer diagnoses at a preliminary phase of malignancy. The authors in [44] employed quicker R-CNN to identify abnormalities in mammograms to categorize masses into benign and malignant.

The authors in [45] utilize image fragmentation and limited preprocessing methods to classify cancer. In semantic segmentation, binarization is used to compare the edge detection strategy to the RCNN methodology. The authors in [46] employed a standard diversity CNN and then incorporated two improvement techniques: (i) batch normalization (BN) and (ii) drop-outs (DO). They employed rank-based stochastic pooling. As a consequence, BDR-CNN was developed, which is a hybrid of CNN, BN, DO, and RSP. This BDR-CNN was hybridised with a GCN to create our BDR-CNN-GCN framework, which was then deployed as a data augmentation strategy for breast mammography assessment. The authors in [47] explained HRD, a combination model of a radial basis function network with DT. This approach was evaluated versus three prominent algorithms: kNN, NB, and SVM. The observations reveal that their recommended technique achieved a high level of precision. The authors in [48] created a collection with chaotic tags or corrupted data and contrasted AlexNet against GoogleNet (GN) for cancer identification. The proportion of chaotic training pictures was changed to control the equilibrium of their system’ memory and cognitive.

The authors in [49] developed a unique hybrid optimal feature selection (HOFS) approach to figure out the relevant characteristics to achieve maximum efficiency for this classification. To train the neural network, a number of carefully chosen characteristics are used. In this proposed technique, the searchable relevant information from the minimammographic data capture organization (MIAS) collection was included. The authors in [50] develop a framework for assessing cancer in mammograms. Chroma noise, fragmentation, and clustering are among the phases in the suggested model. The image enhancement is used to improve the image quality from digital mammograms. The term “fuzzy” refers to the employment of a clustering algorithm to segment a breast tumor. The Bi-LSTM method generates EHO to diagnose cancer with performance characteristics. In this study, the MIAS dataset is employed. The results are then contrasted to CNN and DCNN, as well as Bi-LSTM, with EHOBi-LSTM beating everyone else, although FPR and FNR levels need to be decreased. The authors in [51] suggested that system is a combination stages: pretreatment and development of a Convolution Neural Network (CNN). The preparation process sets the mammograms data for the CNN building phase, which encompasses format unity, denoising, image enhancement, ROI retrieval, augmentation, and image scaling. In the CNN development phase, the proposed CNN model is built from the ground up in addition to learning features and identify tumor progression in mammograms data.

According to the discussion in related work, several researchers were critical in classifying mass pictures of breast cancer as benign or malignant. Many studies suggested various CNN models with varying parameters and obtained results ranging from 90 to 98%. However, there are other constraints that must be addressed. The major gap in mammography is mass classification using such a model that can categorize it effectively and quickly. It is possible to do this using a hybrid strategy. This study uses the CBIS-DDSM dataset to develop a unique hybrid approach CNN-Inception-V4 for identifying breast cancer masses as benign or malignant.

3. Proposed Methodology

In this research, we developed a novel hybrid strategy for classifying breast cancer as benign or malignant by combining CNN and Inception-V4 networks as showing in Figure 1. The CBIS-DDSM dataset, which contains mammography images, is used in this study. 450 images are utilized for benign, and 450 images are used for malignant. The images are first cleaned by removing pectoral muscles, labels, and white borders. Then, CLAHE is used to these images to improve their quality in order to produce promising classification results. where is the lowest limit of the data point and is a distribution adjustment factor that is determined for each input picture as showing in Equation (1) [52]. CNN is a form of neural network model that helps us to extract more accurate representations of picture material. Unlike traditional image recognition, which requires you to define the image characteristics, CNN processes the input pixel data from the picture, trains the model, and then extracts the feature representation for better categorization.

Images are cleaned and enhanced before being fed into the hybrid model using filters . where is the sequence such that as mentioned in Equation (2). Equations (3)–(5) [53] show the impulse response.

The information is sent to the following stage, which extracts features from Convolution (feature map ) using batch normalization [54]+ReLU and stride. A convolution runs the window over pictures and computes the input and filtration system dot product pixel values. This enables convolution to highlight the essential characteristics. The data is then transferred to max pool (feature map ) with a 25% dropout rate to extract features more effectively.

is an 8-dimensional vector, is a 3-dimensional vector, and is a 6-dimensional vector. Then, we make an matrix that encodes the values of so that as shown in Equation (6). To lower data size and processing time, CNN employs max pooling to replace output with a max summary. This helps you to identify the elements that have the greatest influence while reducing the danger of overfitting. Max pooling takes two hyperparameters into account: stride and size.

Even in this discrete example, the cumulative distribution functions of, , and (also known as expected values) are shown in Equations (7)–(9) [52]. The stride determines the skipping of value pools, whereas the length determines the amount of the value pools in each skip. The collected features are then fed into a convolution of with feature map with batch normalization+ReLU and stride. The results of the convolution layer was then routed to a max pool layer of with a feature map. Inception-V4 is the result of the merger of various networks.

Inception-v4 is a convolutional neural network structure that improves on prior generations of the Inception family by simplifying the design and employing additional inception modules than Inception-v3. Inception-v4, which evolved from GoogLeNet/Inception-v1, has a more consistent streamlined design and more inception modules than Inception-v3. This network has numerous blocks, including the Stem block, Inception-A block, Inception-B block, and Inception-C block, as well as reduction blocks.

The first block is the stem of the Inception V4 network, in which input is transmitted to a convolution of 32 2 V filters with size, then to convolution with size, and finally to with convolution of 64 kernels.

Another image matrix is the convolution of the image with the kernel as shown in Equation (10). It then splits into two branches. The first is the max pooling branch, which has 96 2 V, and the second is the convolution branch, which has 2 V. The output of these two branches is in size. Then, it splits into two branches once more. The first is convolution (64), while the second isconvolution (64). The first branch advances to various convolution layers, including convolution (64), convolution (64), and convolution (96 V). The second branch goes to convolution (96 V). The output is concatenated once again with dimensions. Following that, it separates into two branches once more. The first is max pool with a voltage of 2 V, and the second is convolution with a voltage of 192 V, and the output from these two branches concatenates to a size. The following block is Inception-A, in which the output from the stem is sent into the Inception-A block through the concatenation function. Then, it proceeds to four branches. The first is convolution (64), the second is convolution (96), and the fourth is AvgPooling. The input from the first branch is routed to convolution with 96 and then to convolution (96). The input from the second branch is routed to convolution (96), while the input from the fourth branch is sent to convolution (96). Then, using the filter concatenation function, all inputs from subbranches are concatenated. The input from this block is sent to the Inception-B block, which also contains branches and subbranches as shown in Equations (11)–(14). Input passes to convolution (192) from the first branch convolution (192), then to convolution (224), then to convolution (224), and and finally to convolution (256). From the second branch, the input is sent to convolution (192), then to convolution (224), and finally to convolution (256). convolution is the third branch (384).

The input from this block is sent to the Inception-B block, which also contains branches and subbranches as showing in Equations (11)–(14). Input passes to convolution (192) from the first branch convolution (192), then to convolution (224), then to convolution (224), and finally to convolution (256). From the second branch, the input is sent to convolution (192), then to convolution (224), and finally to convolution (256). convolution is the third branch (384). Input is sent to convolution from the fourth branch of AvgPooling (128). Then, using the filter concatenation function, all inputs from subbranches are concatenated. The following block is Inception-C. The input is routed through multiple branches and subbranches in this block. From the first branch, convolution (384), input is passed to convolution (448), then to convolution (512), then to convolution (256), and finally to convolution (256). convolution (384), convolution (256), convolution (256), and finally to convolution (256). Input is sent to convolution from the third branch of AvgPooling (256). Then, using the filter concatenation function, all inputs from subbranches are concatenated. After going through CNN and the Inception model, the input enters the classification phase, which has a 25% dropout rate and a ReLU activation function.

After passing through classification neurons, it employs Softmax for binary classification and then predicts benign and aggressive breast cancer. This suggested network was trained and tested using 50 epochs and 32 batch sizes, yielding encouraging findings that are reported in the next section.

4. Results and Discussion

In this study, we used the CBIS-DDSM dataset to classify breast cancer as benign or malignant. And, as described in Proposed Methodology, we employed a CNN-Inception-V4 hybrid methodology for this purpose. 450 images are employed for benign, and 450 images are used for malignant.

Testing and training images of benign and malignant are shown in Table 1.

50 epochs and 32 batch size for training and testing are used. We utilized accuracy, sensitivity [55], specificity, and -score [55] as evaluation metrics. Accuracy is the overall prediction of model as shown in the following equation:

Sensitivity is defined as the proportion of accurately identified real positives as shown in the following equation:

Specificity is defined as the proportion of accurately identified real negatives as shown in the following equation:

The F1-score is a mix of the model’s accuracy and recall as shown in the following equation:

Our suggested model’s accuracy is 99.2%, with a sensitivity of 99.8%, specificity of 96.3%, and F1-score of 97%. We compared our proposed model with CNN, Inception-V4, and ResNet-50. Our proposed model outperforms existing classification models, according to the findings as shown in Figure 2. The training and testing accuracy and loss of CNN-Inception-V4 are shown in Figures 3 and 4, respectively. The -axis shows the number of epochs and the -axis shows the accuracy for the model. Training results are shown in blue color, and testing results are shown in orange shade.

Our proposed model 99.2% accurately classifies breast cancer into benign and malignant. The results showing actual and predicted values of benign and malignant are shown in Figure 5.

5. Conclusion and Future Work

Many researchers played an important role in the classification of mass images of breast cancer as benign and malignant, but still, mass classification in mammography remains a significant issue, yet it is essential in assisting radiologists in making accurate diagnoses. There is also a need for such a system that can speed up the process of classification of tumors with good results, and this is possible with a unique hybrid technique. So, using the CBIS-DDSM (Curated Breast Imaging Subset of DDSM) dataset, this research provides a novel hybrid approach CNN-Inception-V4 for classifying breast cancer as benign or malignant. The CBIS-DDSM dataset, which comprises mammography images, is employed. 450 images are used for benign, and 450 images are used for malignant. The images are first cleaned by removing pectoral muscles, labels, and white borders. Then, CLAHE is used to these images to improve their quality in order to produce promising classification results. Following preprocessing, our model diagnoses cancer in mammography pictures as benign or malignant abnormalities (MGI). To lower data size and processing time, CNN employs max pooling to replace output with a max summary. This helps you to identify the elements that have the greatest influence while reducing the danger of overfitting. 50 epochs and 32 batch size for training and testing are used. We utilized accuracy, sensitivity, specificity, and -score as evaluation metrics. Our suggested model’s accuracy is 99.2%, with a sensitivity of 99.8%, specificity of 96.3%, and F1-score of 97%. We also compared our proposed model to CNN, Inception-V4, and ResNet-50. Our proposed model outperforms existing classification models, according to the outcomes. In our research, we employed mass images. As a result, calcification images can be used for further study in the future.

Data Availability

The data used to support the findings of the study are included within the article.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This research was supported by the Information Systems Department, Faculty of Management Comenius University in Bratislava, Odbojárov 10, 82005 Bratislava 25, Slovakia.