Design of ensemble recurrent model with stacked fuzzy ARTMAP for breast cancer detection

Abhishek Das (Department of Electronics and Communication Engineering, Institute of Technical Education and Research, Siksha ‘O' Anusandhan (Deemed to be University), Bhubaneswar, India)
Mihir Narayan Mohanty (Department of Electronics and Communication Engineering, Institute of Technical Education and Research, Siksha ‘O' Anusandhan (Deemed to be University), Bhubaneswar, India)

Applied Computing and Informatics

ISSN: 2634-1964

Article publication date: 21 June 2022

540

Abstract

Purpose

In time and accurate detection of cancer can save the life of the person affected. According to the World Health Organization (WHO), breast cancer occupies the most frequent incidence among all the cancers whereas breast cancer takes fifth place in the case of mortality numbers. Out of many image processing techniques, certain works have focused on convolutional neural networks (CNNs) for processing these images. However, deep learning models are to be explored well.

Design/methodology/approach

In this work, multivariate statistics-based kernel principal component analysis (KPCA) is used for essential features. KPCA is simultaneously helpful for denoising the data. These features are processed through a heterogeneous ensemble model that consists of three base models. The base models comprise recurrent neural network (RNN), long short-term memory (LSTM) and gated recurrent unit (GRU). The outcomes of these base learners are fed to fuzzy adaptive resonance theory mapping (ARTMAP) model for decision making as the nodes are added to the F_2ˆa layer if the winning criteria are fulfilled that makes the ARTMAP model more robust.

Findings

The proposed model is verified using breast histopathology image dataset publicly available at Kaggle. The model provides 99.36% training accuracy and 98.72% validation accuracy. The proposed model utilizes data processing in all aspects, i.e. image denoising to reduce the data redundancy, training by ensemble learning to provide higher results than that of single models. The final classification by a fuzzy ARTMAP model that controls the number of nodes depending upon the performance makes robust accurate classification.

Research limitations/implications

Research in the field of medical applications is an ongoing method. More advanced algorithms are being developed for better classification. Still, the scope is there to design the models in terms of better performance, practicability and cost efficiency in the future. Also, the ensemble models may be chosen with different combinations and characteristics. Only signal instead of images may be verified for this proposed model. Experimental analysis shows the improved performance of the proposed model. This method needs to be verified using practical models. Also, the practical implementation will be carried out for its real-time performance and cost efficiency.

Originality/value

The proposed model is utilized for denoising and to reduce the data redundancy so that the feature selection is done using KPCA. Training and classification are performed using heterogeneous ensemble model designed using RNN, LSTM and GRU as base classifiers to provide higher results than that of single models. Use of adaptive fuzzy mapping model makes the final classification accurate. The effectiveness of combining these methods to a single model is analyzed in this work.

Keywords

Citation

Das, A. and Mohanty, M.N. (2022), "Design of ensemble recurrent model with stacked fuzzy ARTMAP for breast cancer detection", Applied Computing and Informatics, Vol. ahead-of-print No. ahead-of-print. https://doi.org/10.1108/ACI-03-2022-0075

Publisher

:

Emerald Publishing Limited

Copyright © 2022, Abhishek Das and Mihir Narayan Mohanty

License

Published in Applied Computing and Informatics. Published by Emerald Publishing Limited. This article is published under the Creative Commons Attribution (CC BY 4.0) licence. Anyone may reproduce, distribute, translate and create derivative works of this article (for both commercial and non-commercial purposes), subject to full attribution to the original publication and authors. The full terms of this licence may be seen at http://creativecommons.org/licences/by/4.0/legalcode


1. Introduction

Cancer is triggered by various genetic and environmental factors. In most cases, cancers originated as malignant tumors. Rapid growth in these tumors leads to cancer. Breast cancer comes under this category and is deadly. Non-invasive kinds of cancer are observed within the milk ducts with no growth and no spread to nearby tissues, whereas invasive ductal carcinoma (IDC) is just opposite to the previous type and appears in 80% of all cases of breast cancer. Breast cancer is causing the highest mortality rate among all other cancers as per the report shown in Ref. [1] till 2020. In time detection of breast cancer may lead to an increase in the life span of the affected person. The cancer detection technique involves various imaging methods to picture the location, shape and size of the affected tissue producing cancer. The physicians manually observe the scanning reports which may lead to a time-consuming process. Automatic and accurate detection of breast cancer from the images obtained through various means can be preferred in the era of artificial intelligence.

Different deep learning models are preferred in various fields of medical image processing, signal processing, speech enhancement, speech recognition, image generation and also in many other areas. Researchers are attracted to these methods due to their human-like training algorithms as well as for high performance. Initially, single models like convolutional neural network (CNN), recurrent neural network (RNN), long short-term memory (LSTM) and gated recurrent unit (GRU) have been designed for specific purposes. Then, the era of the concatenation of these models started to study the performance. Few concatenated models were designed by concatenating the LSTM layers with convolution layers [2], and even, the traditional machine learning model support vector machine (SVM) [3] has also been attached as the classifier with convolutional layers.

Ensemble learning emerged as the highly improved performing method as compared to single deep learning models. Ensemble learning models are broadly classified as homogeneous and heterogeneous depending upon the types of base learners chosen for initial stage processing. An ensemble model designed with the same base models is termed homogeneous, whereas different base models lead to a heterogeneous ensemble model [4]. The heterogeneous ensemble learning method has been mostly preferred over homogeneous ones. A deep ensemble learning-based approach for accurate detection of breast cancer is used. The main contributions of this work are summarized in the points as follows:

  1. The proposed model is utilized for denoising and to reduce the data redundancy so that the feature selection is done using KPCA;

  2. Training and classification are performed using a heterogeneous ensemble model designed using RNN, LSTM and GRU as base classifiers to provide higher results than that of single models and

  3. The final classification by a fuzzy adaptive resonance theory mapping (ARTMAP) model that controls the number of nodes depending upon the performance makes it robust for accurate classification.

The rest part of the paper is arranged subsequently: The second section describes the various works proposed in the field of breast cancer detection. The third section provides a detailed description of the proposed technique designed for breast cancer detection. The fourth section delivers the outcomes found from the proposed model. In the fifth section, conclusions are drawn along with the future scope of this work followed by the references.

2. Literature survey

The deep learning algorithm has appeared as the most chosen method for the detection of breast cancer. Deep learning is performing with more accuracy on image and signal data in single [5, 6] and hybrid forms [7, 8]. Principal component analysis (PCA) was used to reduce the data size followed by feature extraction by using a multilayer perceptron (MLP) model. The final classification was done by using SVM [9]. Semantic segmentation before detection using SVM and MLP has been used for breast cancer detection [10] from histopathology images. Color normalization followed by enhancement was carried out as preprocessing step. Refinement (R), correlation (C) and adaptive (A) algorithms were used for fine-tuning the features and classification by AdaBoost-oriented tree model [11].

Deep learning techniques are gaining the interest of researchers these days. Mitosis count-based cancer detection from breast histopathology images has been proposed using the Atrous Fully Connected Neural Network (A-FCNN) for segmentation and multi-scale and region-based CNN (MS-RCNN) model for detection [12]. A multi-instant pooling layer-based CNN (MI-CNN) model [13] has been suggested for breast cancer detection from histopathology images. Mislabeled patch correction and classification of histopathology images have been proposed for breast cancer detection [14]. The anomaly detection using generative adversarial network (Ano-GAN) was used for anomaly detection, whereas classification was done using DenseNet-121. A deep learning framework has been proposed for mitotic cell count-based cancer detection from histopathology images using Recurrent Residual U-Net (R2U-Net) for segmentation and Inception Recurrent Residual CNN (IRRCNN) for classification [15]. An auto-encoder designed with the concept of the residual-CNN model [16] has been proposed for cancer detection from hematoxylin and Eosin-stained histopathology images. The use of CNN as a feature extractor and Extreme Machine Learning (ELM) as a classifier has been found in Ref. [17] for malignant versus benign classification. A combination of SVM as anomaly detector and FCNN as a classifier has been proposed in Ref. [18] for breast cancer detection from histopathology images. Labeling the data and then classification has been proposed for cancer detection from histopathology images of the BreastHis Dataset using a deep active learning model [19]. Mitotic cell detection for breast cancer classification has been proposed using two CNNs connected in parallel [20]. Application of empirical wavelet transform (EWT), as well as variational mode decomposition (VMD) as preprocessors followed by ensemble of three CNNs and MLP [21], has been proposed for breast cancer detection from histopathology images. That work also used gene data from breast cancer for image conversion using the DeepInsight framework [22]. For bioimage classification, an ensemble model containing CNNs as a base classifier and a sum rule for final decision-making has been proposed [23]. A modified residual neural network [24] has been proposed for breast cancer detection from histopathology images using modified ResNet34 and modified ResNet50 models. A Stochastic Dilated Residual Ghost (SDRG) method [25] has been proposed for cancer detection from breast histopathology images. In recent work, the combination of the attention technique and residual CNN model [26] has been utilized for breast cancer detection. A combination of the Xception model as feature extractor and radial basis function (RBF) kernel-based SVM as classifier [27] has been proposed for breast cancer detection from histopathology images. The authors have also studied the effect of magnification factors on performance. A combined deep learning model [28] designed using inception and residual blocks has been proposed for cancer detection from histopathology images. The magnification values are also considered for images as a preprocessing step.

From the literature analysis, it is found that ensemble models in this field have not yet been explored. In this work, we have considered the detection of breast cancer from breast histopathology images using data denoising by kernel principal component analysis (KPCA) followed by ensemble recurrent models stacked with fuzzy ARTMAP.

3. Proposed method

It has been studied in various literature that the heterogeneous form of ensemble learning provides better performance in comparison to homogeneous ensemble models and single models due to the use of different fine-tunes algorithms [29–31]. KPCA [32] is used to denoise the normalized raw signal and reduce the redundant information. The ensemble model is of three recurrent models, i.e. RNN, LSTM and GRU. The outcomes are fed to the ARTMAP model for final detection. The workflow diagram of the proposed model is shown in Figure 1.

The parameters which are set before training to obtain better performance and accuracy are termed hyper-parameters. In this work, each image of the dataset has 2500 numbers of pixels. Such enormous data will cause higher processing time. The presence of noise in the image data also increases the data complexity. To lower this effect of data complexity, KPCA is used to select the 500 highest eigenvalues. The data are mapped into higher space with the use of KPCA. These features are used to train each base model for further processing and feature extraction. To receive the 500 features provided by the KPCA, each neural network model has 500 nodes in the input layer. Each base model has 2 hidden layers with 32 and 64 numbers of nodes for deep feature extraction. Each model provides two predictions on the same data with two nodes in the output layer. The prediction values are the corresponding probabilities generated by the base models along with the labels of data. These parameters are concatenated to form six values (two values from three base classifiers) as metadata used to train the meta classifier fuzzy ARTMAP for final classification.

3.1 Dataset

The proposed model is experienced with the dataset [33]. The dataset contains 198,738 non-IDC images and 787860 images of the IDC category. 0 as class value represents non-IDC and IDC is considered as 1. Sample images from each category are shown in Figure 2. As 80% of the total dataset is used for training, the rest 20% is used for testing. However, the training set is again divided into train and validation subsets using 80:20 ratios. Validation data are used for training, whereas testing data are neither used for training nor validation. Test data are separate and meant for testing only. The accuracy along with the robustness is explained in the result section.

3.2 Preprocessing

The color images of the breast histopathology are converted to grayscale images to reduce the computational requirements. The height and width values of each image remain the same as it is. Only the depth of each image was reduced from 3 to 1, i.e. the 50 × 50 × 3 images are converted to 50 × 50 × 1 sized images. Figure 3 shows the sample image after conversion from a color image to a gray image.

The grayscale images with attributes (d) are then normalized using the mean normalization technique as given in Eq. (1).

(1)d=faverage(d)max(d)min(d)

3.3 Image denoising and feature selection

RNN model may pass through vanishing gradient problems due to the large size of data, whereas the LSTM and GRU models are free from this problem [34]. Each image of the dataset has 2,500 numbers of pixels. Such enormous data will cause higher processing time. The presence of noise in the image data also increases the data complexity. To lower this effect of data complexity, KPCA is used to select the 500 highest eigenvalues. Cosine kernel is considered as it is a metric that measures similar documents irrespective of size.

Let the data {di}D i=1n, of the high dimension D to higher dimension space ϕ(di). ϕ(di) is the kernelized version of the input data space and the ability to capture the reduced form of the data. It is common to consider the feature space to have zero mean, which is given by Eq. (2).

(2)1N i=1Nϕ(di)=0

The covariance matrix is then computed using Eq. (3).

(3)=1NI=1Nϕ(di)ϕ(di)T

The corresponding eigenvalues and eigenvectors are given by Eq. (4).

(4)ui=λkui   i=1,,k

Substituting Eq. (3) in Eq. (4), we get

(5)1NI=1Nϕ(di){ϕ(di)Tuk}=λkuk
(6)Hence,uk=i=1Naki ϕ(di)
(7)Whereaki=1λkN ϕ(di )Tuk

According to Mercer’s theory, the kernel κ is represented by

(8)κ(di,dj)=ϕ(di)Tϕ(di)

Substituting Eq. (6) into Eq. (5) and multiplying ϕ(di). On both the sides of equations, we get

(9)1Ni=1Nϕ(dl)ϕ(di)j=1Nakjϕ(di)ϕ(dj)=λkakiϕ(dl)ϕ(di)

Rewriting Eq. (9) in terms of the kernel, yields

(10)1Nκ(dl,di)i=1Nakjκ(di,dj)=λki=1Nκ(dl,di)
(11)K2ak=λkNak
(12)WhereK= κ(dl,di)
and ak represents the N-dimensional column eigenvectors of aki.

The denoised form obtained using KPCA is given by

(13)d^=ϕ(d)Tuk=i=1Nakiκ(d,di)

After computing the zero mean of the kernel, we obtained

(14)K^=K11NKK11N+11NK11N
where K= Kij.K^ is known as the Gram matrix.

In this work, we have used a cosine kernel in the KPCA. The cosine kernel is given by Eq. (15).

(15)κ(di,dj)=didjTdidj

3.4 Training

The denoised data are used to train the three base learners connected in parallel to each other. All the base models are designed with four layers. The output layer has two nodes activated with the SoftMax activation function as two classes are there for classification. These deep learning models are evaluated using the binary cross-entropy loss (BCE) that is mathematically represented by Eq. (16) as follows:

(16)BCE(yi,yi^)=I=12yi logyi^
where yi represents the actual prediction given as target and yi^ represents the experimental prediction done by the model.

From Figure 1, it can be visualized that the base classifiers are indirectly creating the features for ARTMAP. Majority voting, MLP [35] and Fuzzy Min-Max [36] models have been used as the meta classifier in various works. In those models, the number of nodes in each hidden layer is fixed to a certain value before training and cannot be updated according to the training requirements. But in fuzzy ARTMAP, the nodes are added to the F2a layer if the winning criteria are fulfilled. That makes the ARTMAP model more robust in comparison to earlier ones. ARTMAP is designed with the structure as shown in Figure 4.

Fuzzy ARTMAP is designed with two ART models, i.e. ARTa and ARTb, connected by a map field [37] that is capable of developing maps between the clusters generated in the input domain ARTa and output domain ARTb. In Figure 4, F0,F1,  and F2 represent the normalization layer, input layer and recognition layer, respectively, for both the domains.

The ARTa module receives the concatenated outputs a of M dimension from ensemble recurrent models and forms a 2M-dimensional complement-coded vector A as given in Eq. (17).

(17)A=(a,ac)=(a1,,aM,1a1,.1aM)

The category nodes in the network are selected using the category choice function (CCF) given by Eq. (18).

(18)|A Λ Wja||Wja|+αa=Zj(A)
where Λ represents the min operator, Wja are the weights learned by the jth node and αa represents the ARTa choice function.

Once the winning node is selected, a similarity check vigilance test is carried out against a vigilance parameter ρa, where

(19)|A Λ Wja||A|ρa

Resonance occurs when the above winning criterion is fulfilled and the weights are updated using the weight update equation given in Eq. (20).

(20)wjnew=β(A Λ wjold)+(1β)wjold
where β is the learning rate.Figures

A new node is generated when the winning criterion is fulfilled in the F2a layer. The following algorithm is utilized in training the proposed model. The symbols used in the algorithm are as follows:

  • di: image data;

  • yi: target;

  • n: dataset size and

  • B: base classifier

4. Results and discussion

4.1 Denoised features for training and testing

The size of each image of the dataset is denoised and the 2,500 numbers of pixel values are converted to high priority 500 values using KPCA. It is maintained because a low number of features degrade the performance of deep learning models.

4.2 Training, validation and classification results

The three recurrent models are trained with the denoised data and the well-fitted models are stored for stacking with each other. The performance of each model is represented in terms of accuracy in Figure 5. The overall performance of the stacked ensemble model is also shown in the same figure to show the improvement in the performance.

The details of accuracies are provided in Table 1. Figure 6 provides the confusion matrix for each base model and the whole proposed model for validation data.

The performance of the whole model is increasing in comparison to base models due to the second stage training by the ARTMAP which helps in decision making.

The evaluation parameters such as accuracy, F1-score, recall, precision, sensitivity and specificity are calculated using the mathematical expressions given in Eq. (21) to Eq. (26).

(21)Accuracy=TP+TNTP+TN+FP+FN × 100
(22)Precision=TPTP+FP×100
(23)Recall= TPTP+FN×100 
(24)F1 Score=2×Recall ×PrecisionRecall+Precision
(25)Sensitivity= TPTP+FN
(26)Specificity= TNTN+FP

The proposed model is also compared to a few state-of-the-art models and is provided in Table 2. The results in italics show the highest result.

The training and validation loss graphs of the proposed stacked ensemble model are shown in Figure 7.

Very low loss values, i.e. 0.0192 and 0.0331 in training and validation, respectively, represent the efficiency of the proposed model in classifying breast histopathology images classification.

4.3 Discussion

The number of base learners is chosen as three for certain reasons. Two or four base models may affect the performance of the meta classifier if half of them will provide opposite results to that of another half. Three base learners are chosen so that the training of the meta classifier will not be adversely affected and the computation time will be less in comparison to five base classifiers.

Figure 5 shows the accuracy of base models with the proposed stacked ensemble model. As it is observed, the accuracy of each individual starts from 40% and increased gradually. However, the stacked model provides 99.36% accuracy, which is given in Table 1. Also, this is proved in terms of the confusion matrix given in Figure 6. A total of 156 images were taken from the test data including both categories. In total,114 images were from non-IDC images, whereas 42 number images belong to IDC. From this figure, it is observed that the three base learners have the results of a maximum of 105 images as true positive (TP), whereas the stacked ensemble models detected 113 images as TP. Similarly, the proposed model predicted 41 images as true negative (TN) out of 42 IDC images, whereas this count is 33, 36 and 35, respectively, for RNN, LSTM and GRU models. Very low numbers of data are classified as false positive (FP) and false negative (FN) by the proposed model. Data overfitting is analyzed from the validation loss plot shown in Figure 7. The proposed model is free from data overfitting which is verified through convergence. Within 200 epochs, the convergence is found for both training and validation that proves the efficacy. Table 1 and Table 2 are showing the improvement in performance due to the proposed stacked ensemble method in comparison to both base models and state-of-the-art methods, respectively. The proper utilization of data preprocessing to avoid data redundancy and a suitable number of base classifiers of the proposed ensemble model combined together improved the performance.

5. Conclusion

In this work, we have utilized the data denoising by KPCA to reduce the data complexity burden from the classification model for faster processing. Features are extracted using the best form of deep learning, i.e. the deep ensemble learning developed using three recurrent models. The final classification is done with the stacked fuzzy ARTMAP. The proposed model is also free from data overfitting by considering a suitable number of iterations. It provides that the histopathology images are efficiently classified into IDC and non-IDC with 99.36% training and 98.72% validation accuracy.

Research in the field of medical applications is an ongoing method. More advanced algorithms are being developed for better classification. Still, the scope is there to design the models in terms of better performance, practicability and cost efficiency in the future. Also, the ensemble models may be chosen with different combinations and characteristics. Only signals instead of images may be verified for this proposed model. Experimental analysis shows the improved performance of the proposed model. This method needs to be verified using practical models and is kept for future work. Also, the practical implementation will be carried out for its real-time performance and cost efficiency.

Figures

Workflow diagram of the proposed model

Figure 1

Workflow diagram of the proposed model

Samples of breast histopathology images

Figure 2

Samples of breast histopathology images

(a) Dataset image of size 50 × 50 × 3 and (b) images of size 50 × 50 after conversion

Figure 3

(a) Dataset image of size 50 × 50 × 3 and (b) images of size 50 × 50 after conversion

Structure of the ARTMAP

Figure 4

Structure of the ARTMAP

Workflow diagram of the proposed model

Figure 1

Workflow diagram of the proposed model

Accuracy comparison plot

Figure 5

Accuracy comparison plot

Confusion matrix

Figure 6

Confusion matrix

Loss plots for training and validation

Figure 7

Loss plots for training and validation

Performance comparison of base models and proposed model

ModelTrainingValidation accuracy (%)
Accuracy (%)F1-scoreRecallPrecisionSensitivitySpecificity
RNN88.460.91150.90350.91960.90350.785787.18
LSTM93.590.93330.92100.94590.92110.857190.38
GRU90.030.92920.92100.93750.92110.833389.74
Proposed stacked ensemble99.360.99120.99120.99120.99120.976298.72

Performance comparison of the proposed model with related works

WorkMethod usedAccuracy (%)F1-score
PreprocessingFeature extractionClassification
Chiu et al. [9]Data reduction using PCAMLPSVM86.97
Rashmi et al. [10]Color normalizationTexture FeatureSVM + MLP0.83
Li et al. [11]RCAAdaBoost93.30
Kausar et al. [12]Segmentation using A-FCNNMS-RCNN0.902
Das et al. [13]MI-CNN96.63
Alom et al. [15]Segmentation using R2U-NetIRRCNN0.878
Brancati et al. [16]Residual-CNN89.57
Wang et al. [17]CNNELM86.50
Li et al. [18]SVMCNN76
Qi et al. [19]Pre-trained AlexNet92.84
Sebai et al. [20]Ensemble of 2 CNNs0.698
Das et al. [21]EWT and VMDEnsemble of 3 CNNsMLP98.080.98
Sharma et al. [27]MagnificationXceptionSVM R96.250.96
Proposed WorkData denoising using kPCAEnsemble of RNN, LSTM and GRUFuzzy ARTMAP98.720.985

References

1Available from: http://gco.iarc.fr/ (accessed 15 February, 2022).

2Wang, J, Yu, LC, Lai, KR, Zhang, X (2016). Dimensional sentiment analysis using a regional CNN-LSTM model. in Proceedings of the 54th annual meeting of the association for computational linguistics (volume 2: Short papers), 225-30.

3Xue, DX, Zhang, R, Feng, H, Wang, YL (2016). CNN-SVM for microvascular morphological type recognition with data augmentation. J Med Biol Eng, 36(6), 755-64.

4Nguyen, TT, Nguyen, MP, Pham, XC, Liew, AWC (2018). Heterogeneous classifier ensemble with fuzzy rule-based meta learner. Inf Sci, 422, 144-60.

5Mambou, SJ, Maresova, P, Krejcar, O, Selamat, A, Kuca, K (2018). Breast cancer detection using infrared thermal imaging and a deep learning model. Sensors, 18(9), 2799.

6Shen, L, Margolies, LR, Rothstein, JH, Fluder, E, McBride, R, Sieh, W (2019). Deep learning to improve breast cancer detection on screening mammography. Scientific Rep, 9(1), 1-12.

7Ragab, DA, Sharkas, M, Marshall, S, Ren, J (2019). Breast cancer detection using deep convolutional neural networks and support vector machines. PeerJ, 7, e6201.

8Wang, D, Khosla, A, Gargeya, R, Irshad, H, Beck, AH (2016). Deep learning for identifying metastatic breast cancer. arXiv preprint arXiv:1606.05718.

9Chiu, H-J, Li, T-HS, Kuo, P-H (2020). Breast cancer-detection system using PCA, multilayer perceptron, transfer learning, and support vector machine. IEEE Access, 8, 204309-24.

10Rashmi, R, Prasad, K, Udupa, CBK, Shwetha, V (2020). A comparative evaluation of texture features for semantic segmentation of breast histopathological images. IEEE Access, 8, 64331-46.

11Li G, Yuan T, Li C, Zhuo J, Jiang Z, Wu J, Ji D, Zhang H. Effective breast cancer recognition based on fine-grained feature selection. IEEE Access. 2020; 8: 227538-55.

12Kausar, T, Wang, M, Ashraf, MA, Kausar, A (2020). SmallMitosis: small size mitotic cells detection in breast histopathology images. IEEE Access, 9, 905-22.

13Das, K, Conjeti, S, Chatterjee, J, Sheet, D (2020). Detection of breast cancer from whole slide histopathological images using deep multiple instance CNN. IEEE Access, 8, 213502-11.

14Man, R, Yang, P, Xu, B (2020). Classification of breast cancer histopathological images using discriminative patches screened by generative adversarial networks. IEEE Access, 8, 155362-77.

15Alom, MZ, Aspiras, T, Taha, TM, Bowen, T, Asari, VK (2020). MitosisNet: end-to-end mitotic cell detection by multi-task learning. IEEE Access, 8, 68695-710.

16Brancati, N, De Pietro, G, Frucci, M, Riccio, D (2019). A deep learning approach for breast invasive ductal carcinoma detection and lymphoma multi-classification in histological images. IEEE Access, 7, 44709-20.

17Wang Z, Li M, Wang H, Jiang H, Yao Y, Zhang H, Xin J. Breast cancer detection using extreme learning machine based on feature fusion with CNN deep features. IEEE Access. 2019; 7: 105146-58.

18Li, X, Radulovic, M, Kanjer, K, Plataniotis, KN (2019). Discriminative pattern mining for breast cancer histopathology image classification via fully convolutional autoencoder. IEEE Access, 7, 36433-45.

19Qi Q, Li Y, Wang J, Zheng H, Huang Y, Ding X, Rohde GK. Label-efficient breast cancer histopathological image classification. IEEE J Biomed Health Inform. 2018; 23(5): 2108-16.

20Sebai, M, Wang, T, Al-Fadhli, SA (2020). PartMitosis: a partially supervised deep learning framework for mitosis detection in breast cancer histopathology images. IEEE Access, 8, 45133-47.

21Das, A, Mohanty, MN, Mallick, PK, Tiwari, P, Muhammad, K, Zhu, H (2021). Breast cancer detection using an ensemble deep learning method. Biomed Signal Process Control, 70, 103009.

22Sharma, A, Vans, E, Shigemizu, D, Boroevich, KA Tsunoda, T (2019). DeepInsight: a methodology to transform a non-image data to an image for convolution neural network architecture. Scientific Rep, 9(1), 1-7.

23Nanni L, Ghidoni S, Brahnam S. Ensemble of convolutional neural networks for bioimage classification. Appl Comput Inform. 2021; 17(1): 19-35.

24Gupta, V, Vasudev, M, Doegar, A, Sambyal, N (2021). Breast cancer detection from histopathology images using modified residual neural networks. Biocybernetics Biomed Eng, 41(4), 1272-87.

25Kashyap, R (2022). Breast cancer histopathological image classification using stochastic dilated residual Ghost model. Int J Inf Retrieval Res (IJIRR), 12(1), 1-24.

26Zou, Y, Zhang, J, Huang, S Liu, B (2022). Breast cancer histopathological image classification using attention high-order deep network. Int J Imaging Syst Technol, 32(1), 266-79.

27Sharma, S Kumar, S (2022). The Xception model: a potential feature extractor in breast cancer histology images classification. ICT Express, 8(1), 101-8.

28Singh, S, Kumar, R (2022). Breast cancer detection from histopathology images with deep inception and residual blocks. Multimedia Tools Appl, 81(4), 5849-65.

29Chali, Y, Hasan, SA, Mojahid, M (2014). Complex question answering: homogeneous or heterogeneous, which ensemble is better? in International Conference on Applications of Natural Language to Data Bases/Information Systems (pp. 160-3). Springer.

30Alshdaifat, E.A., Al-hassan, M., Aloqaily, A (2021). Effective heterogeneous ensemble classification: an alternative approach for selecting base classifiers. ICT Express, 7(3), 342-9.

31Zhong Y, Chen W, Wang Z, Chen Y, Wang K, Li Y, Yin X, Shi X, Yang J, Li K. HELAD: a novel network anomaly detection model based on heterogeneous ensemble learning. Computer Networks. 2020; 169: 107049.

32Schölkopf, B, Smola, A, Müller, K-R (1998). Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput, 10(5), 1299-319.

33Janowczyk A, Madabhushi A. Deep learning for digital pathology image analysis: a comprehensive tutorial with selected use cases. J Pathol Inform. 2016; 7(1): 29.

34Das A, Patra G R, Mohanty M N. A comparison study of recurrent neural networks in recognition of handwritten Odia numerals. Advances in electronics, communication and computing. Singapore: Springer; 2021. 251-260.

35Das A, Mohanty MN. Classification of magnetic resonance images of brain using concatenated deep neural network. Int J Model, Identification and Control, Inderscience. In press.

36Das, A, Mohapatra, SK, Mohanty, MN (2022). Design of deep ensemble classifier with fuzzy decision method for biomedical image classification. Appl Soft Comput, 115, 108178.

37Carpenter, GA, Grossberg, S, Markuzon, N, Reynolds, JH, Rosen, DB (1992). Fuzzy ARTMAP: a neural network architecture for incremental supervised learning of analog multidimensional maps. IEEE Trans Neural netw, 3(5), 698-713.

Acknowledgements

Conflict of interest: The authors declare that there is no conflict of interest for this work.

Corresponding author

Mihir Narayan Mohanty can be contacted at: mihir.n.mohanty@gmail.com

Related articles