Search results

1 – 10 of over 19000
Article
Publication date: 14 June 2022

Rie Isshiki, Ryota Kawamata, Shinji Wakao and Noboru Murata

The density method is one of the powerful topology optimization methods of magnetic devices. The density method has the advantage that it has a high degree of freedom of shape…

Abstract

Purpose

The density method is one of the powerful topology optimization methods of magnetic devices. The density method has the advantage that it has a high degree of freedom of shape expression which results in a high-performance design. On the other hand, it has also the drawback that unsuitable shapes for actually manufacturing are likely to be generated, e.g. checkerboards or grayscale. The purpose of this paper is to develop a method that enables topology optimization suitable for fabrication while taking advantage of the density method.

Design/methodology/approach

This study proposes a novel topology optimization method that combines convolutional neural network (CNN) as an effective smoothing filter with the density method and apply the method to the shield design with magnetic nonlinearity.

Findings

This study demonstrated some numerical examples verifying that the proposed method enables to efficiently obtain a smooth and easy-to-manufacture shield shape with high shielding ability. A network architecture suitable as smoothing filter was also exemplified.

Originality/value

In the field of magnetic field analysis, very few studies have verified the usefulness of smoothing by using CNN in the topology optimization of magnetic devices. This paper develops a novel topology optimization method that skillfully combines CNN with the nonlinear magnetic field analysis and also clarifies a suitable network architecture that makes it possible to obtain a target device shape that is easy to manufacture while minimizing the objective function value.

Details

COMPEL - The international journal for computation and mathematics in electrical and electronic engineering , vol. 41 no. 6
Type: Research Article
ISSN: 0332-1649

Keywords

Article
Publication date: 27 February 2023

Dilawar Ali, Kenzo Milleville, Steven Verstockt, Nico Van de Weghe, Sally Chambers and Julie M. Birkholz

Historical newspaper collections provide a wealth of information about the past. Although the digitization of these collections significantly improves their accessibility, a large…

Abstract

Purpose

Historical newspaper collections provide a wealth of information about the past. Although the digitization of these collections significantly improves their accessibility, a large portion of digitized historical newspaper collections, such as those of KBR, the Royal Library of Belgium, are not yet searchable at article-level. However, recent developments in AI-based research methods, such as document layout analysis, have the potential for further enriching the metadata to improve the searchability of these historical newspaper collections. This paper aims to discuss the aforementioned issue.

Design/methodology/approach

In this paper, the authors explore how existing computer vision and machine learning approaches can be used to improve access to digitized historical newspapers. To do this, the authors propose a workflow, using computer vision and machine learning approaches to (1) provide article-level access to digitized historical newspaper collections using document layout analysis, (2) extract specific types of articles (e.g. feuilletons – literary supplements from Le Peuple from 1938), (3) conduct image similarity analysis using (un)supervised classification methods and (4) perform named entity recognition (NER) to link the extracted information to open data.

Findings

The results show that the proposed workflow improves the accessibility and searchability of digitized historical newspapers, and also contributes to the building of corpora for digital humanities research. The AI-based methods enable automatic extraction of feuilletons, clustering of similar images and dynamic linking of related articles.

Originality/value

The proposed workflow enables automatic extraction of articles, including detection of a specific type of article, such as a feuilleton or literary supplement. This is particularly valuable for humanities researchers as it improves the searchability of these collections and enables corpora to be built around specific themes. Article-level access to, and improved searchability of, KBR's digitized newspapers are demonstrated through the online tool (https://tw06v072.ugent.be/kbr/).

Article
Publication date: 21 April 2020

Bo Li, Jian ming Wang, Qi Wang, Xiu yan Li and Xiaojie Duan

The purpose of this paper is to explore gas/liquid two-phase flow is widely existed in industrial fields, especially in chemical engineering. Electrical resistance tomography…

Abstract

Purpose

The purpose of this paper is to explore gas/liquid two-phase flow is widely existed in industrial fields, especially in chemical engineering. Electrical resistance tomography (ERT) is considered to be one of the most promising techniques to monitor the transient flow process because of its advantages such as fast respond speed and cross-section imaging. However, maintaining high resolution in space together with low cost is still challenging for two-phase flow imaging because of the ill-conditioning of ERT inverse problem.

Design/methodology/approach

In this paper, a sparse reconstruction (SR) method based on the learned dictionary has been proposed for ERT, to accurately monitor the transient flow process of gas/liquid two-phase flow in a pipeline. The high-level representation of the conductivity distributions for typical flow regimes can be extracted based on denoising the deep extreme learning machine (DDELM) model, which is used as prior information for dictionary learning.

Findings

The results from simulation and dynamic experiments indicate that the proposed algorithm efficiently improves the quality of reconstructed images as compared to some typical algorithms such as Landweber and SR-discrete fourier transformation/discrete cosine transformation. Furthermore, the SR-DDELM has also used to estimate the important parameters of the chemical process, a case in point is the volume flow rate. Therefore, the SR-DDELM is considered an ideal candidate for online monitor the gas/liquid two-phase flow.

Originality/value

This paper fulfills a novel approach to effectively monitor the gas/liquid two-phase flow in pipelines. One deep learning model and one adaptive dictionary are trained via the same prior conductivity, respectively. The model is used to extract high-level representation. The dictionary is used to represent the features of the flow process. SR and extraction of high-level representation are performed iteratively. The new method can obviously improve the monitoring accuracy and save calculation time.

Details

Sensor Review, vol. 40 no. 4
Type: Research Article
ISSN: 0260-2288

Keywords

Article
Publication date: 16 August 2019

Shuangshuang Liu and Xiaoling Li

Conventional image super-resolution reconstruction by the conventional deep learning architectures suffers from the problems of hard training and gradient disappearing. In order…

Abstract

Purpose

Conventional image super-resolution reconstruction by the conventional deep learning architectures suffers from the problems of hard training and gradient disappearing. In order to solve such problems, the purpose of this paper is to propose a novel image super-resolution algorithm based on improved generative adversarial networks (GANs) with Wasserstein distance and gradient penalty.

Design/methodology/approach

The proposed algorithm first introduces the conventional GANs architecture, the Wasserstein distance and the gradient penalty for the task of image super-resolution reconstruction (SRWGANs-GP). In addition, a novel perceptual loss function is designed for the SRWGANs-GP to meet the task of image super-resolution reconstruction. The content loss is extracted from the deep model’s feature maps, and such features are introduced to calculate mean square error (MSE) for the loss calculation of generators.

Findings

To validate the effectiveness and feasibility of the proposed algorithm, a lot of compared experiments are applied on three common data sets, i.e. Set5, Set14 and BSD100. Experimental results have shown that the proposed SRWGANs-GP architecture has a stable error gradient and iteratively convergence. Compared with the baseline deep models, the proposed GANs models have a significant improvement on performance and efficiency for image super-resolution reconstruction. The MSE calculated by the deep model’s feature maps gives more advantages for constructing contour and texture.

Originality/value

Compared with the state-of-the-art algorithms, the proposed algorithm obtains a better performance on image super-resolution and better reconstruction results on contour and texture.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 12 no. 3
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 17 June 2021

Ambica Ghai, Pradeep Kumar and Samrat Gupta

Web users rely heavily on online content make decisions without assessing the veracity of the content. The online content comprising text, image, video or audio may be tampered…

1161

Abstract

Purpose

Web users rely heavily on online content make decisions without assessing the veracity of the content. The online content comprising text, image, video or audio may be tampered with to influence public opinion. Since the consumers of online information (misinformation) tend to trust the content when the image(s) supplement the text, image manipulation software is increasingly being used to forge the images. To address the crucial problem of image manipulation, this study focusses on developing a deep-learning-based image forgery detection framework.

Design/methodology/approach

The proposed deep-learning-based framework aims to detect images forged using copy-move and splicing techniques. The image transformation technique aids the identification of relevant features for the network to train effectively. After that, the pre-trained customized convolutional neural network is used to train on the public benchmark datasets, and the performance is evaluated on the test dataset using various parameters.

Findings

The comparative analysis of image transformation techniques and experiments conducted on benchmark datasets from a variety of socio-cultural domains establishes the effectiveness and viability of the proposed framework. These findings affirm the potential applicability of proposed framework in real-time image forgery detection.

Research limitations/implications

This study bears implications for several important aspects of research on image forgery detection. First this research adds to recent discussion on feature extraction and learning for image forgery detection. While prior research on image forgery detection, hand-crafted the features, the proposed solution contributes to stream of literature that automatically learns the features and classify the images. Second, this research contributes to ongoing effort in curtailing the spread of misinformation using images. The extant literature on spread of misinformation has prominently focussed on textual data shared over social media platforms. The study addresses the call for greater emphasis on the development of robust image transformation techniques.

Practical implications

This study carries important practical implications for various domains such as forensic sciences, media and journalism where image data is increasingly being used to make inferences. The integration of image forgery detection tools can be helpful in determining the credibility of the article or post before it is shared over the Internet. The content shared over the Internet by the users has become an important component of news reporting. The framework proposed in this paper can be further extended and trained on more annotated real-world data so as to function as a tool for fact-checkers.

Social implications

In the current scenario wherein most of the image forgery detection studies attempt to assess whether the image is real or forged in an offline mode, it is crucial to identify any trending or potential forged image as early as possible. By learning from historical data, the proposed framework can aid in early prediction of forged images to detect the newly emerging forged images even before they occur. In summary, the proposed framework has a potential to mitigate physical spreading and psychological impact of forged images on social media.

Originality/value

This study focusses on copy-move and splicing techniques while integrating transfer learning concepts to classify forged images with high accuracy. The synergistic use of hitherto little explored image transformation techniques and customized convolutional neural network helps design a robust image forgery detection framework. Experiments and findings establish that the proposed framework accurately classifies forged images, thus mitigating the negative socio-cultural spread of misinformation.

Details

Information Technology & People, vol. 37 no. 2
Type: Research Article
ISSN: 0959-3845

Keywords

Article
Publication date: 23 August 2019

Haiqing He, Ting Chen, Minqiang Chen, Dajun Li and Penggen Cheng

This paper aims to present a novel approach of image super-resolution based on deep–shallow cascaded convolutional neural networks for reconstructing a clear and high-resolution…

Abstract

Purpose

This paper aims to present a novel approach of image super-resolution based on deep–shallow cascaded convolutional neural networks for reconstructing a clear and high-resolution (HR) remote sensing image from a low-resolution (LR) input.

Design/methodology/approach

The proposed approach directly learns the residuals and mapping between simulated LR and their corresponding HR remote sensing images based on deep and shallow end-to-end convolutional networks instead of assuming any specific restored models. Extra max-pooling and up-sampling are used to achieve a multiscale space by concatenating low- and high-level feature maps, and an HR image is generated by combining LR input and the residual image. This model ensures a strong response to spatially local input patterns by using a large filter and cascaded small filters. The authors adopt a strategy based on epochs to update the learning rate for boosting convergence speed.

Findings

The proposed deep network is trained to reconstruct high-quality images for low-quality inputs through a simulated dataset, which is generated with Set5, Set14, Berkeley Segmentation Data set and remote sensing images. Experimental results demonstrate that this model considerably enhances remote sensing images in terms of spatial detail and spectral fidelity and outperforms state-of-the-art SR methods in terms of peak signal-to-noise ratio, structural similarity and visual assessment.

Originality/value

The proposed method can reconstruct an HR remote sensing image from an LR input and significantly improve the quality of remote sensing images in terms of spatial detail and fidelity.

Details

Sensor Review, vol. 39 no. 5
Type: Research Article
ISSN: 0260-2288

Keywords

Article
Publication date: 18 October 2021

Saurabh Kumar

Decision-making in human beings is affected by emotions and sentiments. The affective computing takes this into account, intending to tailor decision support to the emotional…

Abstract

Purpose

Decision-making in human beings is affected by emotions and sentiments. The affective computing takes this into account, intending to tailor decision support to the emotional states of people. However, the representation and classification of emotions is a very challenging task. The study used customized methods of deep learning models to aid in the accurate classification of emotions and sentiments.

Design/methodology/approach

The present study presents affective computing model using both text and image data. The text-based affective computing was conducted on four standard datasets using three deep learning customized models, namely LSTM, GRU and CNN. The study used four variants of deep learning including the LSTM model, LSTM model with GloVe embeddings, Bi-directional LSTM model and LSTM model with attention layer.

Findings

The result suggests that the proposed method outperforms the earlier methods. For image-based affective computing, the data was extracted from Instagram, and Facial emotion recognition was carried out using three deep learning models, namely CNN, transfer learning with VGG-19 model and transfer learning with ResNet-18 model. The results suggest that the proposed methods for both text and image can be used for affective computing and aid in decision-making.

Originality/value

The study used deep learning for affective computing. Earlier studies have used machine learning algorithms for affective computing. However, the present study uses deep learning for affective computing.

Details

Journal of Enterprise Information Management, vol. 34 no. 5
Type: Research Article
ISSN: 1741-0398

Keywords

Article
Publication date: 23 August 2022

Siyuan Huang, Limin Liu, Xiongjun Fu, Jian Dong, Fuyu Huang and Ping Lang

The purpose of this paper is to summarize the existing point cloud target detection algorithms based on deep learning, and provide reference for researchers in related fields. In…

Abstract

Purpose

The purpose of this paper is to summarize the existing point cloud target detection algorithms based on deep learning, and provide reference for researchers in related fields. In recent years, with its outstanding performance in target detection of 2D images, deep learning technology has been applied in light detection and ranging (LiDAR) point cloud data to improve the automation and intelligence level of target detection. However, there are still some difficulties and room for improvement in target detection from the 3D point cloud. In this paper, the vehicle LiDAR target detection method is chosen as the research subject.

Design/methodology/approach

Firstly, the challenges of applying deep learning to point cloud target detection are described; secondly, solutions in relevant research are combed in response to the above challenges. The currently popular target detection methods are classified, among which some are compared with illustrate advantages and disadvantages. Moreover, approaches to improve the accuracy of network target detection are introduced.

Findings

Finally, this paper also summarizes the shortcomings of existing methods and signals the prospective development trend.

Originality/value

This paper introduces some existing point cloud target detection methods based on deep learning, which can be applied to a driverless, digital map, traffic monitoring and other fields, and provides a reference for researchers in related fields.

Details

Sensor Review, vol. 42 no. 5
Type: Research Article
ISSN: 0260-2288

Keywords

Article
Publication date: 9 July 2020

Xin Liu, Junhui Wu, Yiyun Man, Xibao Xu and Jifeng Guo

With the continuous development of aerospace technology, space exploration missions have been increasing year by year, and higher requirements have been placed on the upper level…

Abstract

Purpose

With the continuous development of aerospace technology, space exploration missions have been increasing year by year, and higher requirements have been placed on the upper level rocket. The purpose of this paper is to improve the ability to identify and detect potential targets for upper level rocket.

Design/methodology/approach

Aiming at the upper-level recognition of space satellites and core components, this paper proposes a deep learning-based spatial multi-target recognition method, which can simultaneously recognize space satellites and core components. First, the implementation framework of spatial multi-target recognition is given. Second, by comparing and analyzing convolutional neural networks, a convolutional neural network model based on YOLOv3 is designed. Finally, seven satellite scale models are constructed based on systems tool kit (STK) and Solidworks. Multi targets, such as nozzle, star sensor, solar,etc., are selected as the recognition objects.

Findings

By labeling, training and testing the image data set, the accuracy of the proposed method for spatial multi-target recognition is 90.17%, which is improved compared with the recognition accuracy and rate based on the YOLOv1 model, thereby effectively verifying the correctness of the proposed method.

Research limitations/implications

This paper only recognizes space multi-targets under ideal simulation conditions, but has not fully considered the space multi-target recognition under the more complex space lighting environment, nutation, precession, roll and other motion laws. In the later period, training and detection can be performed by simulating more realistic space lighting environment images or multi-target images taken by upper-level rocket to further verify the feasibility of multi-target recognition algorithms in complex space environments.

Practical implications

The research in this paper validates that the deep learning-based algorithm to recognize multiple targets in the space environment is feasible in terms of accuracy and rate.

Originality/value

The paper helps to set up an image data set containing six satellite models in STK and one digital satellite model that simulates spatial illumination changes and spins in Solidworks, and use the characteristics of spatial targets (such as rectangles, circles and lines) to provide prior values to the network convolutional layer.

Details

Aircraft Engineering and Aerospace Technology, vol. 92 no. 8
Type: Research Article
ISSN: 1748-8842

Keywords

Article
Publication date: 20 April 2023

Vishva Payghode, Ayush Goyal, Anupama Bhan, Sailesh Suryanarayan Iyer and Ashwani Kumar Dubey

This paper aims to implement and extend the You Only Live Once (YOLO) algorithm for detection of objects and activities. The advantage of YOLO is that it only runs a neural…

Abstract

Purpose

This paper aims to implement and extend the You Only Live Once (YOLO) algorithm for detection of objects and activities. The advantage of YOLO is that it only runs a neural network once to detect the objects in an image, which is why it is powerful and fast. Cameras are found at many different crossroads and locations, but video processing of the feed through an object detection algorithm allows determining and tracking what is captured. Video Surveillance has many applications such as Car Tracking and tracking of people related to crime prevention. This paper provides exhaustive comparison between the existing methods and proposed method. Proposed method is found to have highest object detection accuracy.

Design/methodology/approach

The goal of this research is to develop a deep learning framework to automate the task of analyzing video footage through object detection in images. This framework processes video feed or image frames from CCTV, webcam or a DroidCam, which allows the camera in a mobile phone to be used as a webcam for a laptop. The object detection algorithm, with its model trained on a large data set of images, is able to load in each image given as an input, process the image and determine the categories of the matching objects that it finds. As a proof of concept, this research demonstrates the algorithm on images of several different objects. This research implements and extends the YOLO algorithm for detection of objects and activities. The advantage of YOLO is that it only runs a neural network once to detect the objects in an image, which is why it is powerful and fast. Cameras are found at many different crossroads and locations, but video processing of the feed through an object detection algorithm allows determining and tracking what is captured. For video surveillance of traffic cameras, this has many applications, such as car tracking and person tracking for crime prevention. In this research, the implemented algorithm with the proposed methodology is compared against several different prior existing methods in literature. The proposed method was found to have the highest object detection accuracy for object detection and activity recognition, better than other existing methods.

Findings

The results indicate that the proposed deep learning–based model can be implemented in real-time for object detection and activity recognition. The added features of car crash detection, fall detection and social distancing detection can be used to implement a real-time video surveillance system that can help save lives and protect people. Such a real-time video surveillance system could be installed at street and traffic cameras and in CCTV systems. When this system would detect a car crash or a fatal human or pedestrian fall with injury, it can be programmed to send automatic messages to the nearest local police, emergency and fire stations. When this system would detect a social distancing violation, it can be programmed to inform the local authorities or sound an alarm with a warning message to alert the public to maintain their distance and avoid spreading their aerosol particles that may cause the spread of viruses, including the COVID-19 virus.

Originality/value

This paper proposes an improved and augmented version of the YOLOv3 model that has been extended to perform activity recognition, such as car crash detection, human fall detection and social distancing detection. The proposed model is based on a deep learning convolutional neural network model used to detect objects in images. The model is trained using the widely used and publicly available Common Objects in Context data set. The proposed model, being an extension of YOLO, can be implemented for real-time object and activity recognition. The proposed model had higher accuracies for both large-scale and all-scale object detection. This proposed model also exceeded all the other previous methods that were compared in extending and augmenting the object detection to activity recognition. The proposed model resulted in the highest accuracy for car crash detection, fall detection and social distancing detection.

Details

International Journal of Web Information Systems, vol. 19 no. 3/4
Type: Research Article
ISSN: 1744-0084

Keywords

1 – 10 of over 19000