Search results

1 – 10 of 25
Article
Publication date: 15 November 2021

Priyanka Yadlapalli, D. Bhavana and Suryanarayana Gunnam

Computed tomography (CT) scan can provide valuable information in the diagnosis of lung diseases. To detect the location of the cancerous lung nodules, this work uses novel deep…

Abstract

Purpose

Computed tomography (CT) scan can provide valuable information in the diagnosis of lung diseases. To detect the location of the cancerous lung nodules, this work uses novel deep learning methods. The majority of the early investigations used CT, magnetic resonance and mammography imaging. Using appropriate procedures, the professional doctor in this sector analyses these images to discover and diagnose the various degrees of lung cancer. All of the methods used to discover and detect cancer illnesses are time-consuming, expensive and stressful for the patients. To address all of these issues, appropriate deep learning approaches for analyzing these medical images, which included CT scan images, were utilized.

Design/methodology/approach

Radiologists currently employ chest CT scans to detect lung cancer at an early stage. In certain situations, radiologists' perception plays a critical role in identifying lung melanoma which is incorrectly detected. Deep learning is a new, capable and influential approach for predicting medical images. In this paper, the authors employed deep transfer learning algorithms for intelligent classification of lung nodules. Convolutional neural networks (VGG16, VGG19, MobileNet and DenseNet169) are used to constrain the input and output layers of a chest CT scan image dataset.

Findings

The collection includes normal chest CT scan pictures as well as images from two kinds of lung cancer, squamous and adenocarcinoma impacted chest CT scan images. According to the confusion matrix results, the VGG16 transfer learning technique has the highest accuracy in lung cancer classification with 91.28% accuracy, followed by VGG19 with 89.39%, MobileNet with 85.60% and DenseNet169 with 83.71% accuracy, which is analyzed using Google Collaborator.

Originality/value

The proposed approach using VGG16 maximizes the classification accuracy when compared to VGG19, MobileNet and DenseNet169. The results are validated by computing the confusion matrix for each network type.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 15 no. 3
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 9 April 2024

Shola Usharani, R. Gayathri, Uday Surya Deveswar Reddy Kovvuri, Maddukuri Nivas, Abdul Quadir Md, Kong Fah Tee and Arun Kumar Sivaraman

Automation of detecting cracked surfaces on buildings or in any industrially manufactured products is emerging nowadays. Detection of the cracked surface is a challenging task for…

Abstract

Purpose

Automation of detecting cracked surfaces on buildings or in any industrially manufactured products is emerging nowadays. Detection of the cracked surface is a challenging task for inspectors. Image-based automatic inspection of cracks can be very effective when compared to human eye inspection. With the advancement in deep learning techniques, by utilizing these methods the authors can create automation of work in a particular sector of various industries.

Design/methodology/approach

In this study, an upgraded convolutional neural network-based crack detection method has been proposed. The dataset consists of 3,886 images which include cracked and non-cracked images. Further, these data have been split into training and validation data. To inspect the cracks more accurately, data augmentation was performed on the dataset, and regularization techniques have been utilized to reduce the overfitting problems. In this work, VGG19, Xception and Inception V3, along with Resnet50 V2 CNN architectures to train the data.

Findings

A comparison between the trained models has been performed and from the obtained results, Xception performs better than other algorithms with 99.54% test accuracy. The results show detecting cracked regions and firm non-cracked regions is very efficient by the Xception algorithm.

Originality/value

The proposed method can be way better back to an automatic inspection of cracks in buildings with different design patterns such as decorated historical monuments.

Details

International Journal of Structural Integrity, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1757-9864

Keywords

Article
Publication date: 8 June 2021

Naga Swetha R, Vimal K. Shrivastava and K. Parvathi

The mortality rate due to skin cancers has been increasing over the past decades. Early detection and treatment of skin cancers can save lives. However, due to visual resemblance…

Abstract

Purpose

The mortality rate due to skin cancers has been increasing over the past decades. Early detection and treatment of skin cancers can save lives. However, due to visual resemblance of normal skin and lesion and blurred lesion borders, skin cancer diagnosis has become a challenging task even for skilled dermatologists. Hence, the purpose of this study is to present an image-based automatic approach for multiclass skin lesion classification and compare the performance of various models.

Design/methodology/approach

In this paper, the authors have presented a multiclass skin lesion classification approach based on transfer learning of deep convolutional neural network. The following pre-trained models have been used: VGG16, VGG19, ResNet50, ResNet101, ResNet152, Xception, MobileNet and compared their performances on skin cancer classification.

Findings

The experiments have been performed on HAM10000 dataset, which contains 10,015 dermoscopic images of seven skin lesion classes. The categorical accuracy of 83.69%, Top2 accuracy of 91.48% and Top3 accuracy of 96.19% has been obtained.

Originality/value

Early detection and treatment of skin cancer can save millions of lives. This work demonstrates that the transfer learning can be an effective way to classify skin cancer images, providing adequate performance with less computational complexity.

Details

International Journal of Intelligent Unmanned Systems, vol. 12 no. 2
Type: Research Article
ISSN: 2049-6427

Keywords

Article
Publication date: 25 March 2020

Wang Zhao and Long Lu

Facial expression provides abundant information for social interaction, and the analysis and utilization of facial expression data are playing a huge driving role in all areas of…

Abstract

Purpose

Facial expression provides abundant information for social interaction, and the analysis and utilization of facial expression data are playing a huge driving role in all areas of society. Facial expression data can reflect people's mental state. In health care, the analysis and processing of facial expression data can promote the improvement of people's health. This paper introduces several important public facial expression databases and describes the process of facial expression recognition. The standard facial expression database FER2013 and CK+ were used as the main training samples. At the same time, the facial expression image data of 16 Chinese children were collected as supplementary samples. With the help of VGG19 and Resnet18 algorithm models of deep convolution neural network, this paper studies and develops an information system for the diagnosis of autism by facial expression data.

Design/methodology/approach

The facial expression data of the training samples are based on the standard expression database FER2013 and CK+. FER2013 and CK+ databases are a common facial expression data set, which is suitable for the research of facial expression recognition. On the basis of FER2013 and CK+ facial expression database, this paper uses the machine learning model support vector machine (SVM) and deep convolution neural network model CNN, VGG19 and Resnet18 to complete the facial expression recognition.

Findings

In this study, ten normal children and ten autistic patients were recruited to test the accuracy of the information system and the diagnostic effect of autism. After testing, the accuracy rate of facial expression recognition is 81.4 percent. This information system can easily identify autistic children. The feasibility of recognizing autism through facial expression is verified.

Research limitations/implications

The CK+ facial expression database contains some adult facial expression images. In order to improve the accuracy of facial expression recognition for children, more facial expression data of children will be collected as training samples. Therefore, the recognition rate of the information system will be further improved.

Originality/value

This research uses facial expression data and the latest artificial intelligence technology, which is advanced in technology. The diagnostic accuracy of autism is higher than that of traditional systems, so this study is innovative. Research topics come from the actual needs of doctors, and the contents and methods of research have been discussed with doctors many times. The system can diagnose autism as early as possible, promote the early treatment and rehabilitation of patients, and then reduce the economic and mental burden of patients. Therefore, this information system has good social benefits and application value.

Details

Library Hi Tech, vol. 38 no. 4
Type: Research Article
ISSN: 0737-8831

Keywords

Article
Publication date: 18 October 2022

Hasnae Zerouaoui, Ali Idri and Omar El Alaoui

Hundreds of thousands of deaths each year in the world are caused by breast cancer (BC). An early-stage diagnosis of this disease can positively reduce the morbidity and mortality…

Abstract

Purpose

Hundreds of thousands of deaths each year in the world are caused by breast cancer (BC). An early-stage diagnosis of this disease can positively reduce the morbidity and mortality rate by helping to select the most appropriate treatment options, especially by using histological BC images for the diagnosis.

Design/methodology/approach

The present study proposes and evaluates a novel approach which consists of 24 deep hybrid heterogenous ensembles that combine the strength of seven deep learning techniques (DenseNet 201, Inception V3, VGG16, VGG19, Inception-ResNet-V3, MobileNet V2 and ResNet 50) for feature extraction and four well-known classifiers (multi-layer perceptron, support vector machines, K-nearest neighbors and decision tree) by means of hard and weighted voting combination methods for histological classification of BC medical image. Furthermore, the best deep hybrid heterogenous ensembles were compared to the deep stacked ensembles to determine the best strategy to design the deep ensemble methods. The empirical evaluations used four classification performance criteria (accuracy, sensitivity, precision and F1-score), fivefold cross-validation, Scott–Knott (SK) statistical test and Borda count voting method. All empirical evaluations were assessed using four performance measures, including accuracy, precision, recall and F1-score, and were over the histological BreakHis public dataset with four magnification factors (40×, 100×, 200× and 400×). SK statistical test and Borda count were also used to cluster the designed techniques and rank the techniques belonging to the best SK cluster, respectively.

Findings

Results showed that the deep hybrid heterogenous ensembles outperformed both their singles and the deep stacked ensembles and reached the accuracy values of 96.3, 95.6, 96.3 and 94 per cent across the four magnification factors 40×, 100×, 200× and 400×, respectively.

Originality/value

The proposed deep hybrid heterogenous ensembles can be applied for the BC diagnosis to assist pathologists in reducing the missed diagnoses and proposing adequate treatments for the patients.

Details

Data Technologies and Applications, vol. 57 no. 2
Type: Research Article
ISSN: 2514-9288

Keywords

Article
Publication date: 17 March 2021

Eslam Mohammed Abdelkader

Cracks on surface are often identified as one of the early indications of damage and possible future catastrophic structural failure. Thus, detection of cracks is vital for the…

Abstract

Purpose

Cracks on surface are often identified as one of the early indications of damage and possible future catastrophic structural failure. Thus, detection of cracks is vital for the timely inspection, health diagnosis and maintenance of infrastructures. However, conventional visual inspection-based methods are criticized for being subjective, greatly affected by inspector's expertise, labor-intensive and time-consuming.

Design/methodology/approach

This paper proposes a novel self-adaptive-based method for automated and semantic crack detection and recognition in various infrastructures using computer vision technologies. The developed method is envisioned on three main models that are structured to circumvent the shortcomings of visual inspection in detection of cracks in walls, pavement and deck. The first model deploys modified visual geometry group network (VGG19) for extraction of global contextual and local deep learning features in an attempt to alleviate the drawbacks of hand-crafted features. The second model is conceptualized on the integration of K-nearest neighbors (KNN) and differential evolution (DE) algorithm for the automated optimization of its structure. The third model is designated for validating the developed method through an extensive four layers of performance evaluation and statistical comparisons.

Findings

It was observed that the developed method significantly outperformed other crack and detection models. For instance, the developed wall crack detection method accomplished overall accuracy, F-measure, Kappa coefficient, area under the curve, balanced accuracy, Matthew's correlation coefficient and Youden's index of 99.62%, 99.16%, 0.998, 0.998, 99.17%, 0.989 and 0.983, respectively.

Originality/value

Literature review lacks an efficient method which can look at crack detection and recognition of an ensemble of infrastructures. Furthermore, there is absence of systematic and detailed comparisons between crack detection and recognition models.

Details

Smart and Sustainable Built Environment, vol. 11 no. 3
Type: Research Article
ISSN: 2046-6099

Keywords

Article
Publication date: 26 January 2022

K. Venkataravana Nayak, J.S. Arunalatha, G.U. Vasanthakumar and K.R. Venugopal

The analysis of multimedia content is being applied in various real-time computer vision applications. In multimedia content, digital images constitute a significant part. The…

Abstract

Purpose

The analysis of multimedia content is being applied in various real-time computer vision applications. In multimedia content, digital images constitute a significant part. The representation of digital images interpreted by humans is subjective in nature and complex. Hence, searching for relevant images from the archives is difficult. Thus, electronic image analysis strategies have become effective tools in the process of image interpretation.

Design/methodology/approach

The traditional approach used is text-based, i.e. searching images using textual annotations. It consumes time in the manual process of annotating images and is difficult to reduce the dependency in textual annotations if the archive consists of large number of samples. Therefore, content-based image retrieval (CBIR) is adopted in which the high-level visuals of images are represented in terms of feature vectors, which contain numerical values. It is a commonly used approach to understand the content of query images in retrieving relevant images. Still, the performance is less than optimal due to the presence of semantic gap among the image content representation and human visual understanding perspective because of the image content photometric, geometric variations and occlusions in search environments.

Findings

The authors proposed an image retrieval framework to generate semantic response through the feature extraction with convolution network and optimization of extracted features using adaptive moment estimation algorithm towards enhancing the retrieval performance.

Originality/value

The proposed framework is tested on Corel-1k and ImageNet datasets resulted in an accuracy of 98 and 96%, respectively, compared to the state-of-the-art approaches.

Details

International Journal of Intelligent Unmanned Systems, vol. 11 no. 1
Type: Research Article
ISSN: 2049-6427

Keywords

Article
Publication date: 16 August 2021

Faris Elghaish, Saeed Talebi, Essam Abdellatef, Sandra T. Matarneh, M. Reza Hosseini, Song Wu, Mohammad Mayouf, Aso Hajirasouli and The-Quan Nguyen

This paper aims to Test the capabilities/accuracies of four deep learning pre trained convolutional neural network (CNN) models to detect and classify types of highway cracks, as…

Abstract

Purpose

This paper aims to Test the capabilities/accuracies of four deep learning pre trained convolutional neural network (CNN) models to detect and classify types of highway cracks, as well as developing a new CNN model to maximize the accuracy at different learning rates.

Design/methodology/approach

A sample of 4,663 images of highway cracks were collected and classified into three categories of cracks, namely, “vertical cracks,” “horizontal and vertical cracks” and “diagonal cracks,” subsequently, using “Matlab” to classify the sample to training (70%) and testing (30%) to apply the four deep learning CNN models and compute their accuracies. After that, developing a new deep learning CNN model to maximize the accuracy of detecting and classifying highway cracks and testing the accuracy using three optimization algorithms at different learning rates.

Findings

The accuracies result of the four deep learning pre-trained models are above the averages between top-1 and top-5 and the accuracy of classifying and detecting the samples exceeded the top-5 accuracy for the pre-trained AlexNet model around 3% and by 0.2% for the GoogleNet model. The accurate model here is the GoogleNet model as the accuracy is 89.08% and it is higher than AlexNet by 1.26%. While the computed accuracy for the new created deep learning CNN model exceeded all pre-trained models by achieving 97.62% at a learning rate of 0.001 using Adam’s optimization algorithm.

Practical implications

The created deep learning CNN model will enable users (e.g. highway agencies) to scan a long highway and detect types of cracks accurately in a very short time compared to traditional approaches.

Originality/value

A new deep learning CNN-based highway cracks detection was developed based on testing four pre-trained CNN models and analyze the capabilities of each model to maximize the accuracy of the proposed CNN.

Details

Journal of Engineering, Design and Technology , vol. 20 no. 4
Type: Research Article
ISSN: 1726-0531

Keywords

Article
Publication date: 7 June 2021

Sixian Chan, Jian Tao, Xiaolong Zhou, Binghui Wu, Hongqiang Wang and Shengyong Chen

Visual tracking technology enables industrial robots interacting with human beings intelligently. However, due to the complexity of the tracking problem, the accuracy of visual…

Abstract

Purpose

Visual tracking technology enables industrial robots interacting with human beings intelligently. However, due to the complexity of the tracking problem, the accuracy of visual target tracking still has great space for improvement. This paper aims to propose an accurate visual target tracking method based on standard hedging and feature fusion.

Design/methodology/approach

For this study, the authors first learn the discriminative information between targets and similar objects in the histogram of oriented gradients by feature optimization method, and then use standard hedging algorithms to dynamically balance the weights between different feature optimization components. Moreover, they penalize the filter coefficients by incorporating spatial regularization coefficient and extend the Kernelized Correlation Filter for robust tracking. Finally, a model update mechanism to improve the effectiveness of the tracking is proposed.

Findings

Extensive experimental results demonstrate the superior performance of the proposed method comparing to the state-of-the-art tracking methods.

Originality/value

Improvements to existing visual target tracking algorithms are achieved through feature fusion and standard hedging algorithms to further improve the tracking accuracy of robots on targets in reality.

Details

Industrial Robot: the international journal of robotics research and application, vol. 48 no. 5
Type: Research Article
ISSN: 0143-991X

Keywords

Article
Publication date: 29 December 2023

Thanh-Nghi Do and Minh-Thu Tran-Nguyen

This study aims to propose novel edge device-tailored federated learning algorithms of local classifiers (stochastic gradient descent, support vector machines), namely, FL-lSGD…

Abstract

Purpose

This study aims to propose novel edge device-tailored federated learning algorithms of local classifiers (stochastic gradient descent, support vector machines), namely, FL-lSGD and FL-lSVM. These algorithms are designed to address the challenge of large-scale ImageNet classification.

Design/methodology/approach

The authors’ FL-lSGD and FL-lSVM trains in a parallel and incremental manner to build an ensemble local classifier on Raspberry Pis without requiring data exchange. The algorithms load small data blocks of the local training subset stored on the Raspberry Pi sequentially to train the local classifiers. The data block is split into k partitions using the k-means algorithm, and models are trained in parallel on each data partition to enable local data classification.

Findings

Empirical test results on the ImageNet data set show that the authors’ FL-lSGD and FL-lSVM algorithms with 4 Raspberry Pis (Quad core Cortex-A72, ARM v8, 64-bit SoC @ 1.5GHz, 4GB RAM) are faster than the state-of-the-art LIBLINEAR algorithm run on a PC (Intel(R) Core i7-4790 CPU, 3.6 GHz, 4 cores, 32GB RAM).

Originality/value

Efficiently addressing the challenge of large-scale ImageNet classification, the authors’ novel federated learning algorithms of local classifiers have been tailored to work on the Raspberry Pi. These algorithms can handle 1,281,167 images and 1,000 classes effectively.

Details

International Journal of Web Information Systems, vol. 20 no. 1
Type: Research Article
ISSN: 1744-0084

Keywords

1 – 10 of 25