Search results

1 – 9 of 9
Article
Publication date: 29 August 2023

Hei-Chia Wang, Martinus Maslim and Hung-Yu Liu

A clickbait is a deceptive headline designed to boost ad revenue without presenting closely relevant content. There are numerous negative repercussions of clickbait, such as…

Abstract

Purpose

A clickbait is a deceptive headline designed to boost ad revenue without presenting closely relevant content. There are numerous negative repercussions of clickbait, such as causing viewers to feel tricked and unhappy, causing long-term confusion, and even attracting cyber criminals. Automatic detection algorithms for clickbait have been developed to address this issue. The fact that there is only one semantic representation for the same term and a limited dataset in Chinese is a need for the existing technologies for detecting clickbait. This study aims to solve the limitations of automated clickbait detection in the Chinese dataset.

Design/methodology/approach

This study combines both to train the model to capture the probable relationship between clickbait news headlines and news content. In addition, part-of-speech elements are used to generate the most appropriate semantic representation for clickbait detection, improving clickbait detection performance.

Findings

This research successfully compiled a dataset containing up to 20,896 Chinese clickbait news articles. This collection contains news headlines, articles, categories and supplementary metadata. The suggested context-aware clickbait detection (CA-CD) model outperforms existing clickbait detection approaches on many criteria, demonstrating the proposed strategy's efficacy.

Originality/value

The originality of this study resides in the newly compiled Chinese clickbait dataset and contextual semantic representation-based clickbait detection approach employing transfer learning. This method can modify the semantic representation of each word based on context and assist the model in more precisely interpreting the original meaning of news articles.

Details

Data Technologies and Applications, vol. 58 no. 2
Type: Research Article
ISSN: 2514-9288

Keywords

Content available
Article
Publication date: 4 January 2023

Shilpa Sonawani and Kailas Patil

Indoor air quality monitoring is extremely important in urban, industrial areas. Considering the devastating effect of declining quality of air in major part of the countries like…

Abstract

Purpose

Indoor air quality monitoring is extremely important in urban, industrial areas. Considering the devastating effect of declining quality of air in major part of the countries like India and China, it is highly recommended to monitor the quality of air which can help people with respiratory diseases, children and elderly people to take necessary precautions and stay safe at their homes. The purpose of this study is to detect air quality and perform predictions which could be part of smart home automation with the use of newer technology.

Design/methodology/approach

This study proposes an Internet-of-Things (IoT)-based air quality measurement, warning and prediction system for ambient assisted living. The proposed ambient assisted living system consists of low-cost air quality sensors and ESP32 controller with new generation embedded system architecture. It can detect Indoor Air Quality parameters like CO, PM2.5, NO2, O3, NH3, temperature, pressure, humidity, etc. The low cost sensor data are calibrated using machine learning techniques for performance improvement. The system has a novel prediction model, multiheaded convolutional neural networks-gated recurrent unit which can detect next hour pollution concentration. The model uses a transfer learning (TL) approach for prediction when the system is new and less data available for prediction. Any neighboring site data can be used to transfer knowledge for early predictions for the new system. It can have a mobile-based application which can send warning notifications to users if the Indoor Air Quality parameters exceed the specified threshold values. This is all required to take necessary measures against bad air quality.

Findings

The IoT-based system has implemented the TL framework, and the results of this study showed that the system works efficiently with performance improvement of 55.42% in RMSE scores for prediction at new target system with insufficient data.

Originality/value

This study demonstrates the implementation of an IoT system which uses low-cost sensors and deep learning model for predicting pollution concentration. The system is tackling the issues of the low-cost sensors for better performance. The novel approach of pretrained models and TL work very well at the new system having data insufficiency issues. This study contributes significantly with the usage of low-cost sensors, open-source advanced technology and performance improvement in prediction ability at new systems. Experimental results and findings are disclosed in this study. This will help install multiple new cost-effective monitoring stations in smart city for pollution forecasting.

Details

International Journal of Pervasive Computing and Communications, vol. 20 no. 1
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 7 July 2023

Wuyan Liang and Xiaolong Xu

In the COVID-19 era, sign language (SL) translation has gained attention in online learning, which evaluates the physical gestures of each student and bridges the communication…

Abstract

Purpose

In the COVID-19 era, sign language (SL) translation has gained attention in online learning, which evaluates the physical gestures of each student and bridges the communication gap between dysphonia and hearing people. The purpose of this paper is to devote the alignment between SL sequence and nature language sequence with high translation performance.

Design/methodology/approach

SL can be characterized as joint/bone location information in two-dimensional space over time, forming skeleton sequences. To encode joint, bone and their motion information, we propose a multistream hierarchy network (MHN) along with a vocab prediction network (VPN) and a joint network (JN) with the recurrent neural network transducer. The JN is used to concatenate the sequences encoded by the MHN and VPN and learn their sequence alignments.

Findings

We verify the effectiveness of the proposed approach and provide experimental results on three large-scale datasets, which show that translation accuracy is 94.96, 54.52, and 92.88 per cent, and the inference time is 18 and 1.7 times faster than listen-attend-spell network (LAS) and visual hierarchy to lexical sequence network (H2SNet) , respectively.

Originality/value

In this paper, we propose a novel framework that can fuse multimodal input (i.e. joint, bone and their motion stream) and align input streams with nature language. Moreover, the provided framework is improved by the different properties of MHN, VPN and JN. Experimental results on the three datasets demonstrate that our approaches outperform the state-of-the-art methods in terms of translation accuracy and speed.

Details

Data Technologies and Applications, vol. 58 no. 2
Type: Research Article
ISSN: 2514-9288

Keywords

Article
Publication date: 3 November 2022

Vinod Nistane

Rolling element bearings (REBs) are commonly used in rotating machinery such as pumps, motors, fans and other machineries. The REBs deteriorate over life cycle time. To know the…

Abstract

Purpose

Rolling element bearings (REBs) are commonly used in rotating machinery such as pumps, motors, fans and other machineries. The REBs deteriorate over life cycle time. To know the amount of deteriorate at any time, this paper aims to present a prognostics approach based on integrating optimize health indicator (OHI) and machine learning algorithm.

Design/methodology/approach

Proposed optimum prediction model would be used to evaluate the remaining useful life (RUL) of REBs. Initially, signal raw data are preprocessing through mother wavelet transform; after that, the primary fault features are extracted. Further, these features process to elevate the clarity of features using the random forest algorithm. Based on variable importance of features, the best representation of fault features is selected. Optimize the selected feature by adjusting weight vector using optimization techniques such as genetic algorithm (GA), sequential quadratic optimization (SQO) and multiobjective optimization (MOO). New OHIs are determined and apply to train the network. Finally, optimum predictive models are developed by integrating OHI and artificial neural network (ANN), K-mean clustering (KMC) (i.e. OHI–GA–ANN, OHI–SQO–ANN, OHI–MOO–ANN, OHI–GA–KMC, OHI–SQO–KMC and OHI–MOO–KMC).

Findings

Optimum prediction models performance are recorded and compared with the actual value. Finally, based on error term values best optimum prediction model is proposed for evaluation of RUL of REBs.

Originality/value

Proposed OHI–GA–KMC model is compared in terms of error values with previously published work. RUL predicted by OHI–GA–KMC model is smaller, giving the advantage of this method.

Article
Publication date: 26 May 2022

Ismail Abiodun Sulaimon, Hafiz Alaka, Razak Olu-Ajayi, Mubashir Ahmad, Saheed Ajayi and Abdul Hye

Road traffic emissions are generally believed to contribute immensely to air pollution, but the effect of road traffic data sets on air quality (AQ) predictions has not been fully…

262

Abstract

Purpose

Road traffic emissions are generally believed to contribute immensely to air pollution, but the effect of road traffic data sets on air quality (AQ) predictions has not been fully investigated. This paper aims to investigate the effects traffic data set have on the performance of machine learning (ML) predictive models in AQ prediction.

Design/methodology/approach

To achieve this, the authors have set up an experiment with the control data set having only the AQ data set and meteorological (Met) data set, while the experimental data set is made up of the AQ data set, Met data set and traffic data set. Several ML models (such as extra trees regressor, eXtreme gradient boosting regressor, random forest regressor, K-neighbors regressor and two others) were trained, tested and compared on these individual combinations of data sets to predict the volume of PM2.5, PM10, NO2 and O3 in the atmosphere at various times of the day.

Findings

The result obtained showed that various ML algorithms react differently to the traffic data set despite generally contributing to the performance improvement of all the ML algorithms considered in this study by at least 20% and an error reduction of at least 18.97%.

Research limitations/implications

This research is limited in terms of the study area, and the result cannot be generalized outside of the UK as some of the inherent conditions may not be similar elsewhere. Additionally, only the ML algorithms commonly used in literature are considered in this research, therefore, leaving out a few other ML algorithms.

Practical implications

This study reinforces the belief that the traffic data set has a significant effect on improving the performance of air pollution ML prediction models. Hence, there is an indication that ML algorithms behave differently when trained with a form of traffic data set in the development of an AQ prediction model. This implies that developers and researchers in AQ prediction need to identify the ML algorithms that behave in their best interest before implementation.

Originality/value

The result of this study will enable researchers to focus more on algorithms of benefit when using traffic data sets in AQ prediction.

Details

Journal of Engineering, Design and Technology , vol. 22 no. 3
Type: Research Article
ISSN: 1726-0531

Keywords

Article
Publication date: 15 February 2024

Xinyu Liu, Kun Ma, Ke Ji, Zhenxiang Chen and Bo Yang

Propaganda is a prevalent technique used in social media to intentionally express opinions or actions with the aim of manipulating or deceiving users. Existing methods for…

Abstract

Purpose

Propaganda is a prevalent technique used in social media to intentionally express opinions or actions with the aim of manipulating or deceiving users. Existing methods for propaganda detection primarily focus on capturing language features within its content. However, these methods tend to overlook the information presented within the external news environment from which propaganda news originated and spread. This news environment reflects recent mainstream media opinions and public attention and contains language characteristics of non-propaganda news. Therefore, the authors have proposed a graph-based multi-information integration network with an external news environment (abbreviated as G-MINE) for propaganda detection.

Design/methodology/approach

G-MINE is proposed to comprise four parts: textual information extraction module, external news environment perception module, multi-information integration module and classifier. Specifically, the external news environment perception module and multi-information integration module extract and integrate the popularity and novelty into the textual information and capture the high-order complementary information between them.

Findings

G-MINE achieves state-of-the-art performance on both the TSHP-17, Qprop and the PTC data sets, with an accuracy of 98.24%, 90.59% and 97.44%, respectively.

Originality/value

An external news environment perception module is proposed to capture the popularity and novelty information, and a multi-information integration module is proposed to effectively fuse them with the textual information.

Details

International Journal of Web Information Systems, vol. 20 no. 2
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 22 February 2024

Yuzhuo Wang, Chengzhi Zhang, Min Song, Seongdeok Kim, Youngsoo Ko and Juhee Lee

In the era of artificial intelligence (AI), algorithms have gained unprecedented importance. Scientific studies have shown that algorithms are frequently mentioned in papers…

88

Abstract

Purpose

In the era of artificial intelligence (AI), algorithms have gained unprecedented importance. Scientific studies have shown that algorithms are frequently mentioned in papers, making mention frequency a classical indicator of their popularity and influence. However, contemporary methods for evaluating influence tend to focus solely on individual algorithms, disregarding the collective impact resulting from the interconnectedness of these algorithms, which can provide a new way to reveal their roles and importance within algorithm clusters. This paper aims to build the co-occurrence network of algorithms in the natural language processing field based on the full-text content of academic papers and analyze the academic influence of algorithms in the group based on the features of the network.

Design/methodology/approach

We use deep learning models to extract algorithm entities from articles and construct the whole, cumulative and annual co-occurrence networks. We first analyze the characteristics of algorithm networks and then use various centrality metrics to obtain the score and ranking of group influence for each algorithm in the whole domain and each year. Finally, we analyze the influence evolution of different representative algorithms.

Findings

The results indicate that algorithm networks also have the characteristics of complex networks, with tight connections between nodes developing over approximately four decades. For different algorithms, algorithms that are classic, high-performing and appear at the junctions of different eras can possess high popularity, control, central position and balanced influence in the network. As an algorithm gradually diminishes its sway within the group, it typically loses its core position first, followed by a dwindling association with other algorithms.

Originality/value

To the best of the authors’ knowledge, this paper is the first large-scale analysis of algorithm networks. The extensive temporal coverage, spanning over four decades of academic publications, ensures the depth and integrity of the network. Our results serve as a cornerstone for constructing multifaceted networks interlinking algorithms, scholars and tasks, facilitating future exploration of their scientific roles and semantic relations.

Details

Aslib Journal of Information Management, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2050-3806

Keywords

Article
Publication date: 27 February 2024

Shefali Arora, Ruchi Mittal, Avinash K. Shrivastava and Shivani Bali

Deep learning (DL) is on the rise because it can make predictions and judgments based on data that is unseen. Blockchain technologies are being combined with DL frameworks in…

Abstract

Purpose

Deep learning (DL) is on the rise because it can make predictions and judgments based on data that is unseen. Blockchain technologies are being combined with DL frameworks in various industries to provide a safe and effective infrastructure. The review comprises literature that lists the most recent techniques used in the aforementioned application sectors. We examine the current research trends across several fields and evaluate the literature in terms of its advantages and disadvantages.

Design/methodology/approach

The integration of blockchain and DL has been explored in several application domains for the past five years (2018–2023). Our research is guided by five research questions, and based on these questions, we concentrate on key application domains such as the usage of Internet of Things (IoT) in several applications, healthcare and cryptocurrency price prediction. We have analyzed the main challenges and possibilities concerning blockchain technologies. We have discussed the methodologies used in the pertinent publications in these areas and contrasted the research trends during the previous five years. Additionally, we provide a comparison of the widely used blockchain frameworks that are used to create blockchain-based DL frameworks.

Findings

By responding to five research objectives, the study highlights and assesses the effectiveness of already published works using blockchain and DL. Our findings indicate that IoT applications, such as their use in smart cities and cars, healthcare and cryptocurrency, are the key areas of research. The primary focus of current research is the enhancement of existing systems, with data analysis, storage and sharing via decentralized systems being the main motivation for this integration. Amongst the various frameworks employed, Ethereum and Hyperledger are popular among researchers in the domain of IoT and healthcare, whereas Bitcoin is popular for research on cryptocurrency.

Originality/value

There is a lack of literature that summarizes the state-of-the-art methods incorporating blockchain and DL in popular domains such as healthcare, IoT and cryptocurrency price prediction. We analyze the existing research done in the past five years (2018–2023) to review the issues and emerging trends.

Details

International Journal of Quality & Reliability Management, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0265-671X

Keywords

Article
Publication date: 16 February 2024

Mengyang Gao, Jun Wang and Ou Liu

Given the critical role of user-generated content (UGC) in e-commerce, exploring various aspects of UGC can aid in understanding user purchase intention and commodity…

Abstract

Purpose

Given the critical role of user-generated content (UGC) in e-commerce, exploring various aspects of UGC can aid in understanding user purchase intention and commodity recommendation. Therefore, this study investigates the impact of UGC on purchase decisions and proposes new recommendation models based on sentiment analysis, which are verified in Douban, one of the most popular UGC websites in China.

Design/methodology/approach

After verifying the relationship between various factors and product sales, this study proposes two models, collaborative filtering recommendation model based on sentiment (SCF) and hidden factors topics recommendation model based on sentiment (SHFT), by combining traditional collaborative filtering model (CF) and hidden factors topics model (HFT) with sentiment analysis.

Findings

The results indicate that sentiment significantly influences purchase intention. Furthermore, the proposed sentiment-based recommendation models outperform traditional CF and HFT in terms of mean absolute error (MAE) and root mean square error (RMSE). Moreover, the two models yield different outcomes for various product categories, providing actionable insights for organizers to implement more precise recommendation strategies.

Practical implications

The findings of this study advocate the incorporation of UGC sentimental factors into websites to heighten recommendation accuracy. Additionally, different recommendation strategies can be employed for different products types.

Originality/value

This study introduces a novel perspective to the recommendation algorithm field. It not only validates the impact of UGC sentiment on purchase intention but also evaluates the proposed models with real-world data. The study provides valuable insights for managerial decision-making aimed at enhancing recommendation systems.

Details

Industrial Management & Data Systems, vol. 124 no. 4
Type: Research Article
ISSN: 0263-5577

Keywords

1 – 9 of 9