Search results

1 – 10 of 109
Article
Publication date: 30 April 2024

Baoxu Tu, Yuanfei Zhang, Kang Min, Fenglei Ni and Minghe Jin

This paper aims to estimate contact location from sparse and high-dimensional soft tactile array sensor data using the tactile image. The authors used three feature extraction…

Abstract

Purpose

This paper aims to estimate contact location from sparse and high-dimensional soft tactile array sensor data using the tactile image. The authors used three feature extraction methods: handcrafted features, convolutional features and autoencoder features. Subsequently, these features were mapped to contact locations through a contact location regression network. Finally, the network performance was evaluated using spherical fittings of three different radii to further determine the optimal feature extraction method.

Design/methodology/approach

This paper aims to estimate contact location from sparse and high-dimensional soft tactile array sensor data using the tactile image.

Findings

This research indicates that data collected by probes can be used for contact localization. Introducing a batch normalization layer after the feature extraction stage significantly enhances the model’s generalization performance. Through qualitative and quantitative analyses, the authors conclude that convolutional methods can more accurately estimate contact locations.

Originality/value

The paper provides both qualitative and quantitative analyses of the performance of three contact localization methods across different datasets. To address the challenge of obtaining accurate contact locations in quantitative analysis, an indirect measurement metric is proposed.

Details

Industrial Robot: the international journal of robotics research and application, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0143-991X

Keywords

Article
Publication date: 10 April 2024

Qihua Ma, Qilin Li, Wenchao Wang and Meng Zhu

This study aims to achieve superior localization and mapping performance in point cloud degradation scenarios through the effective removal of dynamic obstacles. With the…

Abstract

Purpose

This study aims to achieve superior localization and mapping performance in point cloud degradation scenarios through the effective removal of dynamic obstacles. With the continuous development of various technologies for autonomous vehicles, the LIDAR-based Simultaneous localization and mapping (SLAM) system is becoming increasingly important. However, in SLAM systems, effectively addressing the challenges of point cloud degradation scenarios is essential for accurate localization and mapping, with dynamic obstacle removal being a key component.

Design/methodology/approach

This paper proposes a method that combines adaptive feature extraction and loop closure detection algorithms to address this challenge. In the SLAM system, the ground point cloud and non-ground point cloud are separated to reduce the impact of noise. And based on the cylindrical projection image of the point cloud, the intensity features are adaptively extracted, the degradation direction is determined by the degradation factor and the intensity features are matched with the map to correct the degraded pose. Moreover, through the difference in raster distribution of the point clouds before and after two frames in the loop process, the dynamic point clouds are identified and removed, and the map is updated.

Findings

Experimental results show that the method has good performance. The absolute displacement accuracy of the laser odometer is improved by 27.1%, the relative displacement accuracy is improved by 33.5% and the relative angle accuracy is improved by 23.8% after using the adaptive intensity feature extraction method. The position error is reduced by 30% after removing the dynamic target.

Originality/value

Compared with LiDAR odometry and mapping algorithm, the method has greater robustness and accuracy in mapping and localization.

Details

Industrial Robot: the international journal of robotics research and application, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0143-991X

Keywords

Article
Publication date: 7 May 2024

Xinzhe Li, Qinglong Li, Dasom Jeong and Jaekyeong Kim

Most previous studies predicting review helpfulness ignored the significance of deep features embedded in review text and instead relied on hand-crafted features. Hand-crafted and…

Abstract

Purpose

Most previous studies predicting review helpfulness ignored the significance of deep features embedded in review text and instead relied on hand-crafted features. Hand-crafted and deep features have the advantages of high interpretability and predictive accuracy. This study aims to propose a novel review helpfulness prediction model that uses deep learning (DL) techniques to consider the complementarity between hand-crafted and deep features.

Design/methodology/approach

First, an advanced convolutional neural network was applied to extract deep features from unstructured review text. Second, this study used previous studies to extract hand-crafted features that impact the helpfulness of reviews and enhance their interpretability. Third, this study incorporated deep and hand-crafted features into a review helpfulness prediction model and evaluated its performance using the Yelp.com data set. To measure the performance of the proposed model, this study used 2,417,796 restaurant reviews.

Findings

Extensive experiments confirmed that the proposed methodology performs better than traditional machine learning methods. Moreover, this study confirms through an empirical analysis that combining hand-crafted and deep features demonstrates better prediction performance.

Originality/value

To the best of the authors’ knowledge, this is one of the first studies to apply DL techniques and use structured and unstructured data to predict review helpfulness in the restaurant context. In addition, an advanced feature-fusion method was adopted to better use the extracted feature information and identify the complementarity between features.

研究目的

大多数先前预测评论有用性的研究忽视了嵌入在评论文本中的深层特征的重要性, 而主要依赖手工制作的特征。手工制作和深层特征具有高解释性和预测准确性的优势。本研究提出了一种新颖的评论有用性预测模型, 利用深度学习技术来考虑手工制作特征和深层特征之间的互补性。

研究方法

首先, 采用先进的卷积神经网络从非结构化的评论文本中提取深层特征。其次, 本研究利用先前研究中提取的手工制作特征, 这些特征影响了评论的有用性并增强了其解释性。第三, 本研究将深层特征和手工制作特征结合到一个评论有用性预测模型中, 并使用Yelp.com数据集对其性能进行评估。为了衡量所提出模型的性能, 本研究使用了2,417,796条餐厅评论。

研究发现

广泛的实验验证了所提出的方法优于传统的机器学习方法。此外, 通过实证分析, 本研究证实了结合手工制作和深层特征可以展现出更好的预测性能。

研究创新

据我们所知, 这是首个在餐厅评论预测中应用深度学习技术, 并结合了结构化和非结构化数据来预测评论有用性的研究之一。此外, 本研究采用了先进的特征融合方法, 更好地利用了提取的特征信息, 并识别了特征之间的互补性。

Article
Publication date: 9 April 2024

Shola Usharani, R. Gayathri, Uday Surya Deveswar Reddy Kovvuri, Maddukuri Nivas, Abdul Quadir Md, Kong Fah Tee and Arun Kumar Sivaraman

Automation of detecting cracked surfaces on buildings or in any industrially manufactured products is emerging nowadays. Detection of the cracked surface is a challenging task for…

Abstract

Purpose

Automation of detecting cracked surfaces on buildings or in any industrially manufactured products is emerging nowadays. Detection of the cracked surface is a challenging task for inspectors. Image-based automatic inspection of cracks can be very effective when compared to human eye inspection. With the advancement in deep learning techniques, by utilizing these methods the authors can create automation of work in a particular sector of various industries.

Design/methodology/approach

In this study, an upgraded convolutional neural network-based crack detection method has been proposed. The dataset consists of 3,886 images which include cracked and non-cracked images. Further, these data have been split into training and validation data. To inspect the cracks more accurately, data augmentation was performed on the dataset, and regularization techniques have been utilized to reduce the overfitting problems. In this work, VGG19, Xception and Inception V3, along with Resnet50 V2 CNN architectures to train the data.

Findings

A comparison between the trained models has been performed and from the obtained results, Xception performs better than other algorithms with 99.54% test accuracy. The results show detecting cracked regions and firm non-cracked regions is very efficient by the Xception algorithm.

Originality/value

The proposed method can be way better back to an automatic inspection of cracks in buildings with different design patterns such as decorated historical monuments.

Details

International Journal of Structural Integrity, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1757-9864

Keywords

Open Access
Article
Publication date: 26 April 2024

Adela Sobotkova, Ross Deans Kristensen-McLachlan, Orla Mallon and Shawn Adrian Ross

This paper provides practical advice for archaeologists and heritage specialists wishing to use ML approaches to identify archaeological features in high-resolution satellite…

Abstract

Purpose

This paper provides practical advice for archaeologists and heritage specialists wishing to use ML approaches to identify archaeological features in high-resolution satellite imagery (or other remotely sensed data sources). We seek to balance the disproportionately optimistic literature related to the application of ML to archaeological prospection through a discussion of limitations, challenges and other difficulties. We further seek to raise awareness among researchers of the time, effort, expertise and resources necessary to implement ML successfully, so that they can make an informed choice between ML and manual inspection approaches.

Design/methodology/approach

Automated object detection has been the holy grail of archaeological remote sensing for the last two decades. Machine learning (ML) models have proven able to detect uniform features across a consistent background, but more variegated imagery remains a challenge. We set out to detect burial mounds in satellite imagery from a diverse landscape in Central Bulgaria using a pre-trained Convolutional Neural Network (CNN) plus additional but low-touch training to improve performance. Training was accomplished using MOUND/NOT MOUND cutouts, and the model assessed arbitrary tiles of the same size from the image. Results were assessed using field data.

Findings

Validation of results against field data showed that self-reported success rates were misleadingly high, and that the model was misidentifying most features. Setting an identification threshold at 60% probability, and noting that we used an approach where the CNN assessed tiles of a fixed size, tile-based false negative rates were 95–96%, false positive rates were 87–95% of tagged tiles, while true positives were only 5–13%. Counterintuitively, the model provided with training data selected for highly visible mounds (rather than all mounds) performed worse. Development of the model, meanwhile, required approximately 135 person-hours of work.

Research limitations/implications

Our attempt to deploy a pre-trained CNN demonstrates the limitations of this approach when it is used to detect varied features of different sizes within a heterogeneous landscape that contains confounding natural and modern features, such as roads, forests and field boundaries. The model has detected incidental features rather than the mounds themselves, making external validation with field data an essential part of CNN workflows. Correcting the model would require refining the training data as well as adopting different approaches to model choice and execution, raising the computational requirements beyond the level of most cultural heritage practitioners.

Practical implications

Improving the pre-trained model’s performance would require considerable time and resources, on top of the time already invested. The degree of manual intervention required – particularly around the subsetting and annotation of training data – is so significant that it raises the question of whether it would be more efficient to identify all of the mounds manually, either through brute-force inspection by experts or by crowdsourcing the analysis to trained – or even untrained – volunteers. Researchers and heritage specialists seeking efficient methods for extracting features from remotely sensed data should weigh the costs and benefits of ML versus manual approaches carefully.

Social implications

Our literature review indicates that use of artificial intelligence (AI) and ML approaches to archaeological prospection have grown exponentially in the past decade, approaching adoption levels associated with “crossing the chasm” from innovators and early adopters to the majority of researchers. The literature itself, however, is overwhelmingly positive, reflecting some combination of publication bias and a rhetoric of unconditional success. This paper presents the failure of a good-faith attempt to utilise these approaches as a counterbalance and cautionary tale to potential adopters of the technology. Early-majority adopters may find ML difficult to implement effectively in real-life scenarios.

Originality/value

Unlike many high-profile reports from well-funded projects, our paper represents a serious but modestly resourced attempt to apply an ML approach to archaeological remote sensing, using techniques like transfer learning that are promoted as solutions to time and cost problems associated with, e.g. annotating and manipulating training data. While the majority of articles uncritically promote ML, or only discuss how challenges were overcome, our paper investigates how – despite reasonable self-reported scores – the model failed to locate the target features when compared to field data. We also present time, expertise and resourcing requirements, a rarity in ML-for-archaeology publications.

Details

Journal of Documentation, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0022-0418

Keywords

Article
Publication date: 1 May 2024

Mamun Mishra and Bibhuti Bhusan Pati

Islanding detection has become a serious concern due to the extensive integration of renewable energy sources. The non-detection zone (NDZ) and system-specific applicability…

Abstract

Purpose

Islanding detection has become a serious concern due to the extensive integration of renewable energy sources. The non-detection zone (NDZ) and system-specific applicability, which are the two major issues with the islanding detection methods, are addressed here. The purpose of this paper is to devise an islanding detection method with zero NDZ and, which will be applicable to all types of renewable energy sources using the sequence components of the point of common coupling voltage.

Design/methodology/approach

Here, a parameter using the sequence components is derived to devise an islanding detection method. The parameter derived from the sequence components of point of common coupling voltage is analysed using wavelet transform. Various operating conditions, such as islanding and non-islanding, are considered for several test systems to evaluate the performance of the proposed method. All the simulations are carried out in Simulink/MATLAB environment.

Findings

The results showed that the proposed method has zero NDZ for both inverter- and synchronous generator-based renewable energy sources. In addition, the proposed method works satisfactorily as per the IEEE 1547 standards requirement.

Originality/value

Performance of the proposed method has been tested in several test systems and is found to be better than some conventional methods.

Details

World Journal of Engineering, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1708-5284

Keywords

Article
Publication date: 9 April 2024

Lu Wang, Jiahao Zheng, Jianrong Yao and Yuangao Chen

With the rapid growth of the domestic lending industry, assessing whether the borrower of each loan is at risk of default is a pressing issue for financial institutions. Although…

Abstract

Purpose

With the rapid growth of the domestic lending industry, assessing whether the borrower of each loan is at risk of default is a pressing issue for financial institutions. Although there are some models that can handle such problems well, there are still some shortcomings in some aspects. The purpose of this paper is to improve the accuracy of credit assessment models.

Design/methodology/approach

In this paper, three different stages are used to improve the classification performance of LSTM, so that financial institutions can more accurately identify borrowers at risk of default. The first approach is to use the K-Means-SMOTE algorithm to eliminate the imbalance within the class. In the second step, ResNet is used for feature extraction, and then two-layer LSTM is used for learning to strengthen the ability of neural networks to mine and utilize deep information. Finally, the model performance is improved by using the IDWPSO algorithm for optimization when debugging the neural network.

Findings

On two unbalanced datasets (category ratios of 700:1 and 3:1 respectively), the multi-stage improved model was compared with ten other models using accuracy, precision, specificity, recall, G-measure, F-measure and the nonparametric Wilcoxon test. It was demonstrated that the multi-stage improved model showed a more significant advantage in evaluating the imbalanced credit dataset.

Originality/value

In this paper, the parameters of the ResNet-LSTM hybrid neural network, which can fully mine and utilize the deep information, are tuned by an innovative intelligent optimization algorithm to strengthen the classification performance of the model.

Details

Kybernetes, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0368-492X

Keywords

Article
Publication date: 6 May 2024

Ahmed Taibi, Said Touati, Lyes Aomar and Nabil Ikhlef

Bearings play a critical role in the reliable operation of induction machines, and their failure can lead to significant operational challenges and downtime. Detecting and…

Abstract

Purpose

Bearings play a critical role in the reliable operation of induction machines, and their failure can lead to significant operational challenges and downtime. Detecting and diagnosing these defects is imperative to ensure the longevity of induction machines and preventing costly downtime. The purpose of this paper is to develop a novel approach for diagnosis of bearing faults in induction machine.

Design/methodology/approach

To identify the different fault states of the bearing with accurately and efficiently in this paper, the original bearing vibration signal is first decomposed into several intrinsic mode functions (IMFs) using variational mode decomposition (VMD). The IMFs that contain more noise information are selected using the Pearson correlation coefficient. Subsequently, discrete wavelet transform (DWT) is used to filter the noisy IMFs. Second, the composite multiscale weighted permutation entropy (CMWPE) of each component is calculated to form the features vector. Finally, the features vector is reduced using the locality-sensitive discriminant analysis algorithm, to be fed into the support vector machine model for training and classification.

Findings

The obtained results showed the ability of the VMD_DWT algorithm to reduce the noise of raw vibration signals. It also demonstrated that the proposed method can effectively extract different fault features from vibration signals.

Originality/value

This study suggested a new VMD_DWT method to reduce the noise of the bearing vibration signal. The proposed approach for bearing fault diagnosis of induction machine based on VMD-DWT and CMWPE is highly effective. Its effectiveness has been verified using experimental data.

Details

COMPEL - The international journal for computation and mathematics in electrical and electronic engineering , vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0332-1649

Keywords

Article
Publication date: 30 April 2024

Shiqing Wu, Jiahai Wang, Haibin Jiang and Weiye Xue

The purpose of this study is to explore a new assembly process planning and execution mode to realize rapid response, reduce the labor intensity of assembly workers and improve…

Abstract

Purpose

The purpose of this study is to explore a new assembly process planning and execution mode to realize rapid response, reduce the labor intensity of assembly workers and improve the assembly efficiency and quality.

Design/methodology/approach

Based on the related concepts of digital twin, this paper studies the product assembly planning in digital space, the process execution in physical space and the interaction between digital space and physical space. The assembly process planning is simulated and verified in the digital space to generate three-dimensional visual assembly process specification documents, the implementation of the assembly process specification documents in the physical space is monitored and feed back to revise the assembly process and improve the assembly quality.

Findings

Digital twin technology enhances the quality and efficiency of assembly process planning and execution system.

Originality/value

It provides a new perspective for assembly process planning and execution, the architecture, connections and data acquisition approaches of the digital twin-driven framework are proposed in this paper, which is of important theoretical values. What is more, a smart assembly workbench is developed, the specific image classification algorithms are presented in detail too, which is of some industrial application values.

Details

Robotic Intelligence and Automation, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2754-6969

Keywords

Article
Publication date: 12 April 2024

Ahmad Honarjoo and Ehsan Darvishan

This study aims to obtain methods to identify and find the place of damage, which is one of the topics that has always been discussed in structural engineering. The cost of…

Abstract

Purpose

This study aims to obtain methods to identify and find the place of damage, which is one of the topics that has always been discussed in structural engineering. The cost of repairing and rehabilitating massive bridges and buildings is very high, highlighting the need to monitor the structures continuously. One way to track the structure's health is to check the cracks in the concrete. Meanwhile, the current methods of concrete crack detection have complex and heavy calculations.

Design/methodology/approach

This paper presents a new lightweight architecture based on deep learning for crack classification in concrete structures. The proposed architecture was identified and classified in less time and with higher accuracy than other traditional and valid architectures in crack detection. This paper used a standard dataset to detect two-class and multi-class cracks.

Findings

Results show that two images were recognized with 99.53% accuracy based on the proposed method, and multi-class images were classified with 91% accuracy. The low execution time of the proposed architecture compared to other valid architectures in deep learning on the same hardware platform. The use of Adam's optimizer in this research had better performance than other optimizers.

Originality/value

This paper presents a framework based on a lightweight convolutional neural network for nondestructive monitoring of structural health to optimize the calculation costs and reduce execution time in processing.

Details

International Journal of Structural Integrity, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1757-9864

Keywords

Access

Year

Last month (109)

Content type

Earlycite article (109)
1 – 10 of 109