Search results

1 – 10 of 593
Article
Publication date: 14 December 2018

Erion Çano and Maurizio Morisio

The fabulous results of convolution neural networks in image-related tasks attracted attention of text mining, sentiment analysis and other text analysis researchers. It is…

Abstract

Purpose

The fabulous results of convolution neural networks in image-related tasks attracted attention of text mining, sentiment analysis and other text analysis researchers. It is, however, difficult to find enough data for feeding such networks, optimize their parameters, and make the right design choices when constructing network architectures. The purpose of this paper is to present the creation steps of two big data sets of song emotions. The authors also explore usage of convolution and max-pooling neural layers on song lyrics, product and movie review text data sets. Three variants of a simple and flexible neural network architecture are also compared.

Design/methodology/approach

The intention was to spot any important patterns that can serve as guidelines for parameter optimization of similar models. The authors also wanted to identify architecture design choices which lead to high performing sentiment analysis models. To this end, the authors conducted a series of experiments with neural architectures of various configurations.

Findings

The results indicate that parallel convolutions of filter lengths up to 3 are usually enough for capturing relevant text features. Also, max-pooling region size should be adapted to the length of text documents for producing the best feature maps.

Originality/value

Top results the authors got are obtained with feature maps of lengths 6–18. An improvement on future neural network models for sentiment analysis could be generating sentiment polarity prediction of documents using aggregation of predictions on smaller excerpt of the entire text.

Details

Data Technologies and Applications, vol. 53 no. 1
Type: Research Article
ISSN: 2514-9288

Keywords

Article
Publication date: 29 August 2022

Jianbin Xiong, Jinji Nie and Jiehao Li

This paper primarily aims to focus on a review of convolutional neural network (CNN)-based eye control systems. The performance of CNNs in big data has led to the development of…

Abstract

Purpose

This paper primarily aims to focus on a review of convolutional neural network (CNN)-based eye control systems. The performance of CNNs in big data has led to the development of eye control systems. Therefore, a review of eye control systems based on CNNs is helpful for future research.

Design/methodology/approach

In this paper, first, it covers the fundamentals of the eye control system as well as the fundamentals of CNNs. Second, the standard CNN model and the target detection model are summarized. The eye control system’s CNN gaze estimation approach and model are next described and summarized. Finally, the progress of the gaze estimation of the eye control system is discussed and anticipated.

Findings

The eye control system accomplishes the control effect using gaze estimation technology, which focuses on the features and information of the eyeball, eye movement and gaze, among other things. The traditional eye control system adopts pupil monitoring, pupil positioning, Hough algorithm and other methods. This study will focus on a CNN-based eye control system. First of all, the authors present the CNN model, which is effective in image identification, target detection and tracking. Furthermore, the CNN-based eye control system is separated into three categories: semantic information, monocular/binocular and full-face. Finally, three challenges linked to the development of an eye control system based on a CNN are discussed, along with possible solutions.

Originality/value

This research can provide theoretical and engineering basis for the eye control system platform. In addition, it also summarizes the ideas of predecessors to support the development of future research.

Details

Assembly Automation, vol. 42 no. 5
Type: Research Article
ISSN: 0144-5154

Keywords

Article
Publication date: 3 July 2020

Azra Nazir, Roohie Naaz Mir and Shaima Qureshi

The trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud…

274

Abstract

Purpose

The trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud as their resource giant. But this picture leads to underutilization of ever-increasing device pool of IoT that has already passed 15 billion mark in 2015. Thus, it is high time to explore a different approach to tackle this issue, keeping in view the characteristics and needs of the two fields. Processing at the Edge can boost applications with real-time deadlines while complementing security.

Design/methodology/approach

This review paper contributes towards three cardinal directions of research in the field of DL for IoT. The first section covers the categories of IoT devices and how Fog can aid in overcoming the underutilization of millions of devices, forming the realm of the things for IoT. The second direction handles the issue of immense computational requirements of DL models by uncovering specific compression techniques. An appropriate combination of these techniques, including regularization, quantization, and pruning, can aid in building an effective compression pipeline for establishing DL models for IoT use-cases. The third direction incorporates both these views and introduces a novel approach of parallelization for setting up a distributed systems view of DL for IoT.

Findings

DL models are growing deeper with every passing year. Well-coordinated distributed execution of such models using Fog displays a promising future for the IoT application realm. It is realized that a vertically partitioned compressed deep model can handle the trade-off between size, accuracy, communication overhead, bandwidth utilization, and latency but at the expense of an additionally considerable memory footprint. To reduce the memory budget, we propose to exploit Hashed Nets as potentially favorable candidates for distributed frameworks. However, the critical point between accuracy and size for such models needs further investigation.

Originality/value

To the best of our knowledge, no study has explored the inherent parallelism in deep neural network architectures for their efficient distribution over the Edge-Fog continuum. Besides covering techniques and frameworks that have tried to bring inference to the Edge, the review uncovers significant issues and possible future directions for endorsing deep models as processing engines for real-time IoT. The study is directed to both researchers and industrialists to take on various applications to the Edge for better user experience.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 13 no. 3
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 4 March 2021

Defeng Lv, Huawei Wang and Changchang Che

The purpose of this study is to achieve an accurate intelligent fault diagnosis of rolling bearing.

Abstract

Purpose

The purpose of this study is to achieve an accurate intelligent fault diagnosis of rolling bearing.

Design/methodology/approach

To extract deep features of the original vibration signal and improve the generalization ability and robustness of the fault diagnosis model, this paper proposes a fault diagnosis method of rolling bearing based on multiscale convolutional neural network (MCNN) and decision fusion. The original vibration signals are normalized and matrixed to form grayscale image samples. In addition, multiscale samples can be achieved by convoluting these samples with different convolution kernels. Subsequently, MCNN is constructed for fault diagnosis. The results of MCNN are put into a data fusion model to obtain comprehensive fault diagnosis results.

Findings

The bearing data sets with multiple multivariate time series are used to testify the effectiveness of the proposed method. The proposed model can achieve 99.8% accuracy of fault diagnosis. Based on MCNN and decision fusion, the accuracy can be improved by 0.7%–3.4% compared with other models.

Originality/value

The proposed model can extract deep general features of vibration signals by MCNN and obtained robust fault diagnosis results based on the decision fusion model. For a long time series of vibration signals with noise, the proposed model can still achieve accurate fault diagnosis.

Details

Industrial Lubrication and Tribology, vol. 73 no. 3
Type: Research Article
ISSN: 0036-8792

Keywords

Article
Publication date: 15 November 2021

Priyanka Yadlapalli, D. Bhavana and Suryanarayana Gunnam

Computed tomography (CT) scan can provide valuable information in the diagnosis of lung diseases. To detect the location of the cancerous lung nodules, this work uses novel deep…

Abstract

Purpose

Computed tomography (CT) scan can provide valuable information in the diagnosis of lung diseases. To detect the location of the cancerous lung nodules, this work uses novel deep learning methods. The majority of the early investigations used CT, magnetic resonance and mammography imaging. Using appropriate procedures, the professional doctor in this sector analyses these images to discover and diagnose the various degrees of lung cancer. All of the methods used to discover and detect cancer illnesses are time-consuming, expensive and stressful for the patients. To address all of these issues, appropriate deep learning approaches for analyzing these medical images, which included CT scan images, were utilized.

Design/methodology/approach

Radiologists currently employ chest CT scans to detect lung cancer at an early stage. In certain situations, radiologists' perception plays a critical role in identifying lung melanoma which is incorrectly detected. Deep learning is a new, capable and influential approach for predicting medical images. In this paper, the authors employed deep transfer learning algorithms for intelligent classification of lung nodules. Convolutional neural networks (VGG16, VGG19, MobileNet and DenseNet169) are used to constrain the input and output layers of a chest CT scan image dataset.

Findings

The collection includes normal chest CT scan pictures as well as images from two kinds of lung cancer, squamous and adenocarcinoma impacted chest CT scan images. According to the confusion matrix results, the VGG16 transfer learning technique has the highest accuracy in lung cancer classification with 91.28% accuracy, followed by VGG19 with 89.39%, MobileNet with 85.60% and DenseNet169 with 83.71% accuracy, which is analyzed using Google Collaborator.

Originality/value

The proposed approach using VGG16 maximizes the classification accuracy when compared to VGG19, MobileNet and DenseNet169. The results are validated by computing the confusion matrix for each network type.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 15 no. 3
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 20 January 2021

Xueqing Zhao, Min Zhang and Junjun Zhang

Classifying the types of fabric defects in the textile industry requires a way to effectively detect. The traditional textile fabric defects detection method is human eyes, which…

Abstract

Purpose

Classifying the types of fabric defects in the textile industry requires a way to effectively detect. The traditional textile fabric defects detection method is human eyes, which performs very low efficiency and high cost. Therefore, how to improve the classification accuracy of textile fabric defects by using current artificial intelligence and to better meet the needs in the textile industry, the purpose of this article is to develop a method to improve the accuracy of textile fabric defects classification.

Design/methodology/approach

To improve the accuracy of textile fabric defects classification, an ensemble learning-based convolutional neural network (CNN) method in terms of textile fabric defects classification (short for ECTFDC) on an enhanced TILDA database is used. ECTFDC first adopts ensemble learning-based model to classify five types of fabric defects from TILDA. Subsequently, ECTFDC extracts features of fabric defects via an ensemble multiple convolutional neural network model and obtains parameters by using transfer learning method.

Findings

The authors applied ECTFDC on an enhanced TILDA database to improve the robustness and generalization ability of the proposed networks. Experimental results show that ECTFDC outperforms the other networks, the precision and recall rates are 97.8%, 97.68%, respectively.

Originality/value

The ensemble convolutional neural network textile fabric defect classification method in this paper can quickly and effectively classify textile fabric defect categories; it can reduce the production cost of textiles and it can alleviate the visual fatigue of inspectors working for a long time.

Details

International Journal of Clothing Science and Technology, vol. 33 no. 4
Type: Research Article
ISSN: 0955-6222

Keywords

Article
Publication date: 7 February 2023

Riju Bhattacharya, Naresh Kumar Nagwani and Sarsij Tripathi

A community demonstrates the unique qualities and relationships between its members that distinguish it from other communities within a network. Network analysis relies heavily on…

Abstract

Purpose

A community demonstrates the unique qualities and relationships between its members that distinguish it from other communities within a network. Network analysis relies heavily on community detection. Despite the traditional spectral clustering and statistical inference methods, deep learning techniques for community detection have grown in popularity due to their ease of processing high-dimensional network data. Graph convolutional neural networks (GCNNs) have received much attention recently and have developed into a potential and ubiquitous method for directly detecting communities on graphs. Inspired by the promising results of graph convolutional networks (GCNs) in analyzing graph structure data, a novel community graph convolutional network (CommunityGCN) as a semi-supervised node classification model has been proposed and compared with recent baseline methods graph attention network (GAT), GCN-based technique for unsupervised community detection and Markov random fields combined with graph convolutional network (MRFasGCN).

Design/methodology/approach

This work presents the method for identifying communities that combines the notion of node classification via message passing with the architecture of a semi-supervised graph neural network. Six benchmark datasets, namely, Cora, CiteSeer, ACM, Karate, IMDB and Facebook, have been used in the experimentation.

Findings

In the first set of experiments, the scaled normalized average matrix of all neighbor's features including the node itself was obtained, followed by obtaining the weighted average matrix of low-dimensional nodes. In the second set of experiments, the average weighted matrix was forwarded to the GCN with two layers and the activation function for predicting the node class was applied. The results demonstrate that node classification with GCN can improve the performance of identifying communities on graph datasets.

Originality/value

The experiment reveals that the CommunityGCN approach has given better results with accuracy, normalized mutual information, F1 and modularity scores of 91.26, 79.9, 92.58 and 70.5 per cent, respectively, for detecting communities in the graph network, which is much greater than the range of 55.7–87.07 per cent reported in previous literature. Thus, it has been concluded that the GCN with node classification models has improved the accuracy.

Details

Data Technologies and Applications, vol. 57 no. 4
Type: Research Article
ISSN: 2514-9288

Keywords

Article
Publication date: 11 January 2021

Rajit Nair, Santosh Vishwakarma, Mukesh Soni, Tejas Patel and Shubham Joshi

The latest 2019 coronavirus (COVID-2019), which first appeared in December 2019 in Wuhan's city in China, rapidly spread around the world and became a pandemic. It has had a…

Abstract

Purpose

The latest 2019 coronavirus (COVID-2019), which first appeared in December 2019 in Wuhan's city in China, rapidly spread around the world and became a pandemic. It has had a devastating impact on daily lives, the public's health and the global economy. The positive cases must be identified as soon as possible to avoid further dissemination of this disease and swift care of patients affected. The need for supportive diagnostic instruments increased, as no specific automated toolkits are available. The latest results from radiology imaging techniques indicate that these photos provide valuable details on the virus COVID-19. User advanced artificial intelligence (AI) technologies and radiological imagery can help diagnose this condition accurately and help resolve the lack of specialist doctors in isolated areas. In this research, a new paradigm for automatic detection of COVID-19 with bare chest X-ray images is displayed. Images are presented. The proposed model DarkCovidNet is designed to provide correct binary classification diagnostics (COVID vs no detection) and multi-class (COVID vs no results vs pneumonia) classification. The implemented model computed the average precision for the binary and multi-class classification of 98.46% and 91.352%, respectively, and an average accuracy of 98.97% and 87.868%. The DarkNet model was used in this research as a classifier for a real-time object detection method only once. A total of 17 convolutionary layers and different filters on each layer have been implemented. This platform can be used by the radiologists to verify their initial application screening and can also be used for screening patients through the cloud.

Design/methodology/approach

This study also uses the CNN-based model named Darknet-19 model, and this model will act as a platform for the real-time object detection system. The architecture of this system is designed in such a way that they can be able to detect real-time objects. This study has developed the DarkCovidNet model based on Darknet architecture with few layers and filters. So before discussing the DarkCovidNet model, look at the concept of Darknet architecture with their functionality. Typically, the DarkNet architecture consists of 5 pool layers though the max pool and 19 convolution layers. Assume as a convolution layer, and as a pooling layer.

Findings

The work discussed in this paper is used to diagnose the various radiology images and to develop a model that can accurately predict or classify the disease. The data set used in this work is the images bases on COVID-19 and non-COVID-19 taken from the various sources. The deep learning model named DarkCovidNet is applied to the data set, and these have shown signification performance in the case of binary classification and multi-class classification. During the multi-class classification, the model has shown an average accuracy 98.97% for the detection of COVID-19, whereas in a multi-class classification model has achieved an average accuracy of 87.868% during the classification of COVID-19, no detection and Pneumonia.

Research limitations/implications

One of the significant limitations of this work is that a limited number of chest X-ray images were used. It is observed that patients related to COVID-19 are increasing rapidly. In the future, the model on the larger data set which can be generated from the local hospitals will be implemented, and how the model is performing on the same will be checked.

Originality/value

Deep learning technology has made significant changes in the field of AI by generating good results, especially in pattern recognition. A conventional CNN structure includes a convolution layer that extracts characteristics from the input using the filters it applies, a pooling layer that reduces calculation efficiency and the neural network's completely connected layer. A CNN model is created by integrating one or more of these layers, and its internal parameters are modified to accomplish a specific mission, such as classification or object recognition. A typical CNN structure has a convolution layer that extracts features from the input with the filters it applies, a pooling layer to reduce the size for computational performance and a fully connected layer, which is a neural network. A CNN model is created by combining one or more such layers, and its internal parameters are adjusted to accomplish a particular task, such as classification or object recognition.

Details

World Journal of Engineering, vol. 19 no. 1
Type: Research Article
ISSN: 1708-5284

Keywords

Article
Publication date: 9 April 2024

Shola Usharani, R. Gayathri, Uday Surya Deveswar Reddy Kovvuri, Maddukuri Nivas, Abdul Quadir Md, Kong Fah Tee and Arun Kumar Sivaraman

Automation of detecting cracked surfaces on buildings or in any industrially manufactured products is emerging nowadays. Detection of the cracked surface is a challenging task for…

Abstract

Purpose

Automation of detecting cracked surfaces on buildings or in any industrially manufactured products is emerging nowadays. Detection of the cracked surface is a challenging task for inspectors. Image-based automatic inspection of cracks can be very effective when compared to human eye inspection. With the advancement in deep learning techniques, by utilizing these methods the authors can create automation of work in a particular sector of various industries.

Design/methodology/approach

In this study, an upgraded convolutional neural network-based crack detection method has been proposed. The dataset consists of 3,886 images which include cracked and non-cracked images. Further, these data have been split into training and validation data. To inspect the cracks more accurately, data augmentation was performed on the dataset, and regularization techniques have been utilized to reduce the overfitting problems. In this work, VGG19, Xception and Inception V3, along with Resnet50 V2 CNN architectures to train the data.

Findings

A comparison between the trained models has been performed and from the obtained results, Xception performs better than other algorithms with 99.54% test accuracy. The results show detecting cracked regions and firm non-cracked regions is very efficient by the Xception algorithm.

Originality/value

The proposed method can be way better back to an automatic inspection of cracks in buildings with different design patterns such as decorated historical monuments.

Details

International Journal of Structural Integrity, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1757-9864

Keywords

Article
Publication date: 25 March 2020

Wang Zhao and Long Lu

Facial expression provides abundant information for social interaction, and the analysis and utilization of facial expression data are playing a huge driving role in all areas of…

Abstract

Purpose

Facial expression provides abundant information for social interaction, and the analysis and utilization of facial expression data are playing a huge driving role in all areas of society. Facial expression data can reflect people's mental state. In health care, the analysis and processing of facial expression data can promote the improvement of people's health. This paper introduces several important public facial expression databases and describes the process of facial expression recognition. The standard facial expression database FER2013 and CK+ were used as the main training samples. At the same time, the facial expression image data of 16 Chinese children were collected as supplementary samples. With the help of VGG19 and Resnet18 algorithm models of deep convolution neural network, this paper studies and develops an information system for the diagnosis of autism by facial expression data.

Design/methodology/approach

The facial expression data of the training samples are based on the standard expression database FER2013 and CK+. FER2013 and CK+ databases are a common facial expression data set, which is suitable for the research of facial expression recognition. On the basis of FER2013 and CK+ facial expression database, this paper uses the machine learning model support vector machine (SVM) and deep convolution neural network model CNN, VGG19 and Resnet18 to complete the facial expression recognition.

Findings

In this study, ten normal children and ten autistic patients were recruited to test the accuracy of the information system and the diagnostic effect of autism. After testing, the accuracy rate of facial expression recognition is 81.4 percent. This information system can easily identify autistic children. The feasibility of recognizing autism through facial expression is verified.

Research limitations/implications

The CK+ facial expression database contains some adult facial expression images. In order to improve the accuracy of facial expression recognition for children, more facial expression data of children will be collected as training samples. Therefore, the recognition rate of the information system will be further improved.

Originality/value

This research uses facial expression data and the latest artificial intelligence technology, which is advanced in technology. The diagnostic accuracy of autism is higher than that of traditional systems, so this study is innovative. Research topics come from the actual needs of doctors, and the contents and methods of research have been discussed with doctors many times. The system can diagnose autism as early as possible, promote the early treatment and rehabilitation of patients, and then reduce the economic and mental burden of patients. Therefore, this information system has good social benefits and application value.

Details

Library Hi Tech, vol. 38 no. 4
Type: Research Article
ISSN: 0737-8831

Keywords

1 – 10 of 593