Search results

1 – 10 of 84
Article
Publication date: 29 August 2024

Yizhuo Zhang, Yunfei Zhang, Huiling Yu and Shen Shi

The anomaly detection task for oil and gas pipelines based on acoustic signals faces issues such as background noise coverage, lack of effective features, and small sample sizes…

Abstract

Purpose

The anomaly detection task for oil and gas pipelines based on acoustic signals faces issues such as background noise coverage, lack of effective features, and small sample sizes, resulting in low fault identification accuracy and slow efficiency. The purpose of this paper is to study an accurate and efficient method of pipeline anomaly detection.

Design/methodology/approach

First, to address the impact of background noise on the accuracy of anomaly signals, the adaptive multi-threshold center frequency variational mode decomposition method(AMTCF-VMD) method is used to eliminate strong noise in pipeline signals. Secondly, to address the strong data dependency and loss of local features in the Swin Transformer network, a Hybrid Pyramid ConvNet network with an Agent Attention mechanism is proposed. This compensates for the limitations of CNN’s receptive field and enhances the Swin Transformer’s global contextual feature representation capabilities. Thirdly, to address the sparsity and imbalance of anomaly samples, the SpecAugment and Scaper methods are integrated to enhance the model’s generalization ability.

Findings

In the pipeline anomaly audio and environmental datasets such as ESC-50, the AMTCF-VMD method shows more significant denoising effects compared to wavelet packet decomposition and EMD methods. Additionally, the model achieved 98.7% accuracy on the preprocessed anomaly audio dataset and 99.0% on the ESC-50 dataset.

Originality/value

This paper innovatively proposes and combines the AMTCF-VMD preprocessing method with the Agent-SwinPyramidNet model, addressing noise interference and low accuracy issues in pipeline anomaly detection, and providing strong support for oil and gas pipeline anomaly recognition tasks in high-noise environments.

Details

International Journal of Intelligent Computing and Cybernetics, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 26 August 2024

S. Punitha and K. Devaki

Predicting student performance is crucial in educational settings to identify and support students who may need additional help or resources. Understanding and predicting student…

Abstract

Purpose

Predicting student performance is crucial in educational settings to identify and support students who may need additional help or resources. Understanding and predicting student performance is essential for educators to provide targeted support and guidance to students. By analyzing various factors like attendance, study habits, grades, and participation, teachers can gain insights into each student’s academic progress. This information helps them tailor their teaching methods to meet the individual needs of students, ensuring a more personalized and effective learning experience. By identifying patterns and trends in student performance, educators can intervene early to address any challenges and help students acrhieve their full potential. However, the complexity of human behavior and learning patterns makes it difficult to accurately forecast how a student will perform. Additionally, the availability and quality of data can vary, impacting the accuracy of predictions. Despite these obstacles, continuous improvement in data collection methods and the development of more robust predictive models can help address these challenges and enhance the accuracy and effectiveness of student performance predictions. However, the scalability of the existing models to different educational settings and student populations can be a hurdle. Ensuring that the models are adaptable and effective across diverse environments is crucial for their widespread use and impact. To implement a student’s performance-based learning recommendation scheme for predicting the student’s capabilities and suggesting better materials like papers, books, videos, and hyperlinks according to their needs. It enhances the performance of higher education.

Design/methodology/approach

Thus, a predictive approach for student achievement is presented using deep learning. At the beginning, the data is accumulated from the standard database. Next, the collected data undergoes a stage where features are carefully selected using the Modified Red Deer Algorithm (MRDA). After that, the selected features are given to the Deep Ensemble Networks (DEnsNet), in which techniques such as Gated Recurrent Unit (GRU), Deep Conditional Random Field (DCRF), and Residual Long Short-Term Memory (Res-LSTM) are utilized for predicting the student performance. In this case, the parameters within the DEnsNet network are finely tuned by the MRDA algorithm. Finally, the results from the DEnsNet network are obtained using a superior method that delivers the final prediction outcome. Following that, the Adaptive Generative Adversarial Network (AGAN) is introduced for recommender systems, with these parameters optimally selected using the MRDA algorithm. Lastly, the method for predicting student performance is evaluated numerically and compared to traditional methods to demonstrate the effectiveness of the proposed approach.

Findings

The accuracy of the developed model is 7.66%, 9.91%, 5.3%, and 3.53% more than HHO-DEnsNet, ROA-DEnsNet, GTO-DEnsNet, and AOA-DEnsNet for dataset-1, and 7.18%, 7.54%, 5.43% and 3% enhanced than HHO-DEnsNet, ROA-DEnsNet, GTO-DEnsNet, and AOA-DEnsNet for dataset-2.

Originality/value

The developed model recommends the appropriate learning materials within a short period to improve student’s learning ability.

Article
Publication date: 12 September 2024

Zhanglin Peng, Tianci Yin, Xuhui Zhu, Xiaonong Lu and Xiaoyu Li

To predict the price of battery-grade lithium carbonate accurately and provide proper guidance to investors, a method called MFTBGAM is proposed in this study. This method…

Abstract

Purpose

To predict the price of battery-grade lithium carbonate accurately and provide proper guidance to investors, a method called MFTBGAM is proposed in this study. This method integrates textual and numerical information using TCN-BiGRU–Attention.

Design/methodology/approach

The Word2Vec model is initially employed to process the gathered textual data concerning battery-grade lithium carbonate. Subsequently, a dual-channel text-numerical extraction model, integrating TCN and BiGRU, is constructed to extract textual and numerical features separately. Following this, the attention mechanism is applied to extract fusion features from the textual and numerical data. Finally, the market price prediction results for battery-grade lithium carbonate are calculated and outputted using the fully connected layer.

Findings

Experiments in this study are carried out using datasets consisting of news and investor commentary. The findings reveal that the MFTBGAM model exhibits superior performance compared to alternative models, showing its efficacy in precisely forecasting the future market price of battery-grade lithium carbonate.

Research limitations/implications

The dataset analyzed in this study spans from 2020 to 2023, and thus, the forecast results are specifically relevant to this timeframe. Altering the sample data would necessitate repetition of the experimental process, resulting in different outcomes. Furthermore, recognizing that raw data might include noise and irrelevant information, future endeavors will explore efficient data preprocessing techniques to mitigate such issues, thereby enhancing the model’s predictive capabilities in long-term forecasting tasks.

Social implications

The price prediction model serves as a valuable tool for investors in the battery-grade lithium carbonate industry, facilitating informed investment decisions. By using the results of price prediction, investors can discern opportune moments for investment. Moreover, this study utilizes two distinct types of text information – news and investor comments – as independent sources of textual data input. This approach provides investors with a more precise and comprehensive understanding of market dynamics.

Originality/value

We propose a novel price prediction method based on TCN-BiGRU Attention for “text-numerical” information fusion. We separately use two types of textual information, news and investor comments, for prediction to enhance the model's effectiveness and generalization ability. Additionally, we utilize news datasets including both titles and content to improve the accuracy of battery-grade lithium carbonate market price predictions.

Details

Kybernetes, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0368-492X

Keywords

Open Access
Article
Publication date: 26 April 2024

Adela Sobotkova, Ross Deans Kristensen-McLachlan, Orla Mallon and Shawn Adrian Ross

This paper provides practical advice for archaeologists and heritage specialists wishing to use ML approaches to identify archaeological features in high-resolution satellite…

Abstract

Purpose

This paper provides practical advice for archaeologists and heritage specialists wishing to use ML approaches to identify archaeological features in high-resolution satellite imagery (or other remotely sensed data sources). We seek to balance the disproportionately optimistic literature related to the application of ML to archaeological prospection through a discussion of limitations, challenges and other difficulties. We further seek to raise awareness among researchers of the time, effort, expertise and resources necessary to implement ML successfully, so that they can make an informed choice between ML and manual inspection approaches.

Design/methodology/approach

Automated object detection has been the holy grail of archaeological remote sensing for the last two decades. Machine learning (ML) models have proven able to detect uniform features across a consistent background, but more variegated imagery remains a challenge. We set out to detect burial mounds in satellite imagery from a diverse landscape in Central Bulgaria using a pre-trained Convolutional Neural Network (CNN) plus additional but low-touch training to improve performance. Training was accomplished using MOUND/NOT MOUND cutouts, and the model assessed arbitrary tiles of the same size from the image. Results were assessed using field data.

Findings

Validation of results against field data showed that self-reported success rates were misleadingly high, and that the model was misidentifying most features. Setting an identification threshold at 60% probability, and noting that we used an approach where the CNN assessed tiles of a fixed size, tile-based false negative rates were 95–96%, false positive rates were 87–95% of tagged tiles, while true positives were only 5–13%. Counterintuitively, the model provided with training data selected for highly visible mounds (rather than all mounds) performed worse. Development of the model, meanwhile, required approximately 135 person-hours of work.

Research limitations/implications

Our attempt to deploy a pre-trained CNN demonstrates the limitations of this approach when it is used to detect varied features of different sizes within a heterogeneous landscape that contains confounding natural and modern features, such as roads, forests and field boundaries. The model has detected incidental features rather than the mounds themselves, making external validation with field data an essential part of CNN workflows. Correcting the model would require refining the training data as well as adopting different approaches to model choice and execution, raising the computational requirements beyond the level of most cultural heritage practitioners.

Practical implications

Improving the pre-trained model’s performance would require considerable time and resources, on top of the time already invested. The degree of manual intervention required – particularly around the subsetting and annotation of training data – is so significant that it raises the question of whether it would be more efficient to identify all of the mounds manually, either through brute-force inspection by experts or by crowdsourcing the analysis to trained – or even untrained – volunteers. Researchers and heritage specialists seeking efficient methods for extracting features from remotely sensed data should weigh the costs and benefits of ML versus manual approaches carefully.

Social implications

Our literature review indicates that use of artificial intelligence (AI) and ML approaches to archaeological prospection have grown exponentially in the past decade, approaching adoption levels associated with “crossing the chasm” from innovators and early adopters to the majority of researchers. The literature itself, however, is overwhelmingly positive, reflecting some combination of publication bias and a rhetoric of unconditional success. This paper presents the failure of a good-faith attempt to utilise these approaches as a counterbalance and cautionary tale to potential adopters of the technology. Early-majority adopters may find ML difficult to implement effectively in real-life scenarios.

Originality/value

Unlike many high-profile reports from well-funded projects, our paper represents a serious but modestly resourced attempt to apply an ML approach to archaeological remote sensing, using techniques like transfer learning that are promoted as solutions to time and cost problems associated with, e.g. annotating and manipulating training data. While the majority of articles uncritically promote ML, or only discuss how challenges were overcome, our paper investigates how – despite reasonable self-reported scores – the model failed to locate the target features when compared to field data. We also present time, expertise and resourcing requirements, a rarity in ML-for-archaeology publications.

Details

Journal of Documentation, vol. 80 no. 5
Type: Research Article
ISSN: 0022-0418

Keywords

Article
Publication date: 9 September 2024

Weixing Wang, Yixia Chen and Mingwei Lin

Based on the strong feature representation ability of the convolutional neural network (CNN), generous object detection methods in remote sensing (RS) have been proposed one after…

Abstract

Purpose

Based on the strong feature representation ability of the convolutional neural network (CNN), generous object detection methods in remote sensing (RS) have been proposed one after another. However, due to the large variation in scale and the omission of relevant relationships between objects, there are still great challenges for object detection in RS. Most object detection methods fail to take the difficulties of detecting small and medium-sized objects and global context into account. Moreover, inference time and lightness are also major pain points in the field of RS.

Design/methodology/approach

To alleviate the aforementioned problems, this study proposes a novel method for object detection in RS, which is called lightweight object detection with a multi-receptive field and long-range dependency in RS images (MFLD). The multi-receptive field extraction (MRFE) and long-range dependency information extraction (LDIE) modules are put forward.

Findings

To concentrate on the variability of objects in RS, MRFE effectively expands the receptive field by a combination of atrous separable convolutions with different dilated rates. Considering the shortcomings of CNN in extracting global information, LDIE is designed to capture the relationships between objects. Extensive experiments over public datasets in RS images demonstrate that our MFLD method surpasses the state-of-the-art methods. Most of all, on the NWPU VHR-10 dataset, our MFLD method achieves 94.6% mean average precision with 4.08 M model volume.

Originality/value

This paper proposed a method called lightweight object detection with multi-receptive field and long-range dependency in RS images.

Details

International Journal of Intelligent Computing and Cybernetics, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 30 April 2024

Baoxu Tu, Yuanfei Zhang, Kang Min, Fenglei Ni and Minghe Jin

This paper aims to estimate contact location from sparse and high-dimensional soft tactile array sensor data using the tactile image. The authors used three feature extraction…

Abstract

Purpose

This paper aims to estimate contact location from sparse and high-dimensional soft tactile array sensor data using the tactile image. The authors used three feature extraction methods: handcrafted features, convolutional features and autoencoder features. Subsequently, these features were mapped to contact locations through a contact location regression network. Finally, the network performance was evaluated using spherical fittings of three different radii to further determine the optimal feature extraction method.

Design/methodology/approach

This paper aims to estimate contact location from sparse and high-dimensional soft tactile array sensor data using the tactile image.

Findings

This research indicates that data collected by probes can be used for contact localization. Introducing a batch normalization layer after the feature extraction stage significantly enhances the model’s generalization performance. Through qualitative and quantitative analyses, the authors conclude that convolutional methods can more accurately estimate contact locations.

Originality/value

The paper provides both qualitative and quantitative analyses of the performance of three contact localization methods across different datasets. To address the challenge of obtaining accurate contact locations in quantitative analysis, an indirect measurement metric is proposed.

Details

Industrial Robot: the international journal of robotics research and application, vol. 51 no. 5
Type: Research Article
ISSN: 0143-991X

Keywords

Article
Publication date: 12 July 2024

Peng Guo, Weiyong Si and Chenguang Yang

The purpose of this paper is to enhance the performance of robots in peg-in-hole assembly tasks, enabling them to swiftly and robustly accomplish the task. It also focuses on the…

73

Abstract

Purpose

The purpose of this paper is to enhance the performance of robots in peg-in-hole assembly tasks, enabling them to swiftly and robustly accomplish the task. It also focuses on the robot’s ability to generalize across assemblies with different hole sizes.

Design/methodology/approach

Human behavior in peg-in-hole assembly serves as inspiration, where individuals visually locate the hole firstly and then continuously adjust the peg pose based on force/torque feedback during the insertion process. This paper proposes a novel framework that integrate visual servo and adjustment based on force/torque feedback, the authors use deep neural network (DNN) and image processing techniques to determine the pose of hole, then an incremental learning approach based on a broad learning system (BLS) is used to simulate human learning ability, the number of adjustments required for insertion process is continuously reduced.

Findings

The author conducted experiments on visual servo, adjustment based on force/torque feedback, and the proposed framework. Visual servo inferred the pixel position and orientation of the target hole in only about 0.12 s, and the robot achieved peg insertion with 1–3 adjustments based on force/torque feedback. The success rate for peg-in-hole assembly using the proposed framework was 100%. These results proved the effectiveness of the proposed framework.

Originality/value

This paper proposes a framework for peg-in-hole assembly that combines visual servo and adjustment based on force/torque feedback. The assembly tasks are accomplished using DNN, image processing and BLS. To the best of the authors’ knowledge, no similar methods were found in other people’s work. Therefore, the authors believe that this work is original.

Details

Robotic Intelligence and Automation, vol. 44 no. 5
Type: Research Article
ISSN: 2754-6969

Keywords

Open Access
Article
Publication date: 2 September 2024

Siddhartha S. Bora and Ani L. Katchova

Long-term forecasts about commodity market indicators play an important role in informing policy and investment decisions by governments and market participants. Our study…

Abstract

Purpose

Long-term forecasts about commodity market indicators play an important role in informing policy and investment decisions by governments and market participants. Our study examines whether the accuracy of the multi-step forecasts can be improved using deep learning methods.

Design/methodology/approach

We first formulate a supervised learning problem and set benchmarks for forecast accuracy using traditional econometric models. We then train a set of deep neural networks and measure their performance against the benchmark.

Findings

We find that while the United States Department of Agriculture (USDA) baseline projections perform better for shorter forecast horizons, the performance of the deep neural networks improves for longer horizons. The findings may inform future revisions of the forecasting process.

Originality/value

This study demonstrates an application of deep learning methods to multi-horizon forecasts of agri-cultural commodities, which is a departure from the current methods used in producing these types of forecasts.

Details

Agricultural Finance Review, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0002-1466

Keywords

Article
Publication date: 9 September 2024

Andry Alamsyah, Fadiah Nadhila and Nabila Kalvina Izumi

Technology serves as a key catalyst in shaping society and the economy, significantly altering customer dynamics. Through a deep understanding of these evolving behaviors, a…

Abstract

Purpose

Technology serves as a key catalyst in shaping society and the economy, significantly altering customer dynamics. Through a deep understanding of these evolving behaviors, a service can be tailored to address each customer's unique needs and personality. We introduce a strategy to integrate customer complaints with their personality traits, enabling responses that resonate with the customer’s unique personality.

Design/methodology/approach

We propose a strategy to incorporate customer complaints with their personality traits, enabling responses that reflect the customer’s unique personality. Our approach is twofold: firstly, we employ the customer complaints ontology (CCOntology) framework enforced with multi-class classification based on a machine learning algorithm, to classify complaints. Secondly, we leverage the personality measurement platform (PMP), powered by the big five personality model to predict customer’s personalities. We develop the framework for the Indonesian language by extracting tweets containing customer complaints directed towards Indonesia's three biggest e-commerce services.

Findings

By mapping customer complaints and their personality type, we can identify specific personality traits associated with customer dissatisfaction. Thus, personalizing how we offer the solution based on specific characteristics.

Originality/value

The research enriches the state-of-the-art personalizing service research based on captured customer behavior. Thus, our research fills the research gap in considering customer personalities. We provide comprehensive insights by aligning customer feedback with corresponding personality traits extracted from social media data. The result is a highly customized response mechanism attuned to individual customer preferences and requirements.

Details

Data Technologies and Applications, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2514-9288

Keywords

Article
Publication date: 13 September 2024

Ahmad Honarjoo, Ehsan Darvishan, Hassan Rezazadeh and Amir Homayoon Kosarieh

This article introduces SigBERT, a novel approach that fine-tunes bidirectional encoder representations from transformers (BERT) for the purpose of distinguishing between intact…

Abstract

Purpose

This article introduces SigBERT, a novel approach that fine-tunes bidirectional encoder representations from transformers (BERT) for the purpose of distinguishing between intact and impaired structures by analyzing vibration signals. Structural health monitoring (SHM) systems are crucial for identifying and locating damage in civil engineering structures. The proposed method aims to improve upon existing methods in terms of cost-effectiveness, accuracy and operational reliability.

Design/methodology/approach

SigBERT employs a fine-tuning process on the BERT model, leveraging its capabilities to effectively analyze time-series data from vibration signals to detect structural damage. This study compares SigBERT's performance with baseline models to demonstrate its superior accuracy and efficiency.

Findings

The experimental results, obtained through the Qatar University grandstand simulator, show that SigBERT outperforms existing models in terms of damage detection accuracy. The method is capable of handling environmental fluctuations and offers high reliability for non-destructive monitoring of structural health. The study mentions the quantifiable results of the study, such as achieving a 99% accuracy rate and an F-1 score of 0.99, to underline the effectiveness of the proposed model.

Originality/value

SigBERT presents a significant advancement in SHM by integrating deep learning with a robust transformer model. The method offers improved performance in both computational efficiency and diagnostic accuracy, making it suitable for real-world operational environments.

Details

International Journal of Structural Integrity, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1757-9864

Keywords

1 – 10 of 84