Search results

21 – 30 of 212
Article
Publication date: 6 February 2023

Xiaobo Tang, Heshen Zhou and Shixuan Li

Predicting highly cited papers can enable an evaluation of the potential of papers and the early detection and determination of academic achievement value. However, most highly…

Abstract

Purpose

Predicting highly cited papers can enable an evaluation of the potential of papers and the early detection and determination of academic achievement value. However, most highly cited paper prediction studies consider early citation information, so predicting highly cited papers by publication is challenging. Therefore, the authors propose a method for predicting early highly cited papers based on their own features.

Design/methodology/approach

This research analyzed academic papers published in the Journal of the Association for Computing Machinery (ACM) from 2000 to 2013. Five types of features were extracted: paper features, journal features, author features, reference features and semantic features. Subsequently, the authors applied a deep neural network (DNN), support vector machine (SVM), decision tree (DT) and logistic regression (LGR), and they predicted highly cited papers 1–3 years after publication.

Findings

Experimental results showed that early highly cited academic papers are predictable when they are first published. The authors’ prediction models showed considerable performance. This study further confirmed that the features of references and authors play an important role in predicting early highly cited papers. In addition, the proportion of high-quality journal references has a more significant impact on prediction.

Originality/value

Based on the available information at the time of publication, this study proposed an effective early highly cited paper prediction model. This study facilitates the early discovery and realization of the value of scientific and technological achievements.

Details

Library Hi Tech, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0737-8831

Keywords

Article
Publication date: 26 April 2023

Fawaz Qasem

This purpose of this study is to examine future fears and reassurances about the nature of the recent artificial intelligence (AI) language model-based application, ChatGPT, use…

2183

Abstract

Purpose

This purpose of this study is to examine future fears and reassurances about the nature of the recent artificial intelligence (AI) language model-based application, ChatGPT, use in the fields of scientific research and academic works and assignments. This study aims at exploring the positive and negative aspects of the use of ChatGPT by researchers and students. This paper recommends some practical academic steps and suggestions that help the researchers and publishers curtail the percentage of spread of unethical works such as plagiarism.

Design/methodology/approach

The emergence of OpenAI’s Generative Pre-Trained Transformer 3 (GPT-3) has recently sparked controversy and heated debate among academics worldwide about its use and application. The concern of experts and researchers about the GPT-3 platform entails how it would be of much support to the researchers and academic staff and how it might be used and misused to negatively affect academic and scholarly works. This research explored future fears and reassurances about the nature of Chat GPT-3 use at academic and scientific levels. The data for this research was obtained through the qualitative interviews of seven experts in AI, scientific research and academic works. The findings of the study showed that ChatGPT-3 has significant potential and is helpful if used wisely and ethically at scientific and academic levels. On the other hand, the results reported the experts' fears of the frequent use of ChatGPT including the misuse of ChatGPT as a tool to plagiarize and make the researchers dependent, not self-reliant and lazy. The widespread concern of many scholars is that ChatGPT would lead to an increase in the possibility of plagiarism and provide less control over research and writing ethics. This study proposed some stages and suggested that AI language model programs, including ChatGPT, should be integrated with widespread publishers and academic platforms to curtail the percentage of plagiarism and organize the process of publishing and writing scientific research and academic works to save the rights of researchers and writers.

Findings

The findings of the research presented that ChatGPT can act as a potential and useful tool if used wisely and ethically at scientific and academic fields. On contrast, the results also reported the negative aspects of the extensive ChatGPT's that leads to the spread of plagiarism and making the researchers and the students machine-dependent, not self-reliant and lazy. This study proposed some stages and suggested that AI language model programs, including ChatGPT, should be integrated with widespread publishers and academic platforms to curtail the percentage of plagiarism and organize the process of publishing and writing scientific research and academic works to save rights of researchers and writers.

Originality/value

To the best of the authors’ knowledge, this paper is the first of its kind to highlight the relationship between using ChatGPT and the spread of both positive and negative aspects of its extensive use in scientific research and academic work. The importance of this study lies in the fact that it presents the concerns and future fears of people in academia as they cope with and deal with the inevitable reality of AI language models such as ChatGPT.

Open Access
Article
Publication date: 19 July 2022

Shreyesh Doppalapudi, Tingyan Wang and Robin Qiu

Clinical notes typically contain medical jargons and specialized words and phrases that are complicated and technical to most people, which is one of the most challenging…

1060

Abstract

Purpose

Clinical notes typically contain medical jargons and specialized words and phrases that are complicated and technical to most people, which is one of the most challenging obstacles in health information dissemination to consumers by healthcare providers. The authors aim to investigate how to leverage machine learning techniques to transform clinical notes of interest into understandable expressions.

Design/methodology/approach

The authors propose a natural language processing pipeline that is capable of extracting relevant information from long unstructured clinical notes and simplifying lexicons by replacing medical jargons and technical terms. Particularly, the authors develop an unsupervised keywords matching method to extract relevant information from clinical notes. To automatically evaluate completeness of the extracted information, the authors perform a multi-label classification task on the relevant texts. To simplify lexicons in the relevant text, the authors identify complex words using a sequence labeler and leverage transformer models to generate candidate words for substitution. The authors validate the proposed pipeline using 58,167 discharge summaries from critical care services.

Findings

The results show that the proposed pipeline can identify relevant information with high completeness and simplify complex expressions in clinical notes so that the converted notes have a high level of readability but a low degree of meaning change.

Social implications

The proposed pipeline can help healthcare consumers well understand their medical information and therefore strengthen communications between healthcare providers and consumers for better care.

Originality/value

An innovative pipeline approach is developed to address the health literacy problem confronted by healthcare providers and consumers in the ongoing digital transformation process in the healthcare industry.

Article
Publication date: 17 May 2023

Tong Yang, Jie Wu and Junming Zhang

This study aims to establish a comprehensive satisfaction analysis framework by mining online restaurant reviews, which can not only accurately reveal consumer satisfaction but…

Abstract

Purpose

This study aims to establish a comprehensive satisfaction analysis framework by mining online restaurant reviews, which can not only accurately reveal consumer satisfaction but also identify factors leading to dissatisfaction and further quantify improvement opportunity levels.

Design/methodology/approach

Adopting deep learning, Cross-Bidirectional Encoder Representations Transformers (BERT) model is developed to measure customer satisfaction. Furthermore, opinion mining technique is used to extract consumers’ opinions and obtain dissatisfaction factors. Furthermore, the opportunity algorithm is introduced to quantify attributes’ improvement opportunity levels. A total of 19,133 online reviews of 31 restaurants in Universal Beijing Resort are crawled to validate the framework.

Findings

Results demonstrate the superiority of Cross-BERT model compared to existing models such as sentiment lexicon-based model and Naïve Bayes. More importantly, after effectively unveiling customer dissatisfaction factors (e.g. long queuing time and taste salty), “Dish taste,” “Waiters’ attitude” and “Decoration” are identified as the three secondary attributes with the greatest improvement opportunities.

Practical implications

The proposed framework helps managers, especially in the restaurant industry, accurately understand customer satisfaction and reasons behind dissatisfaction, thereby generating efficient countermeasures. Especially, the improvement opportunity levels also benefit practitioners in efficiently allocating limited business resources.

Originality/value

This work contributes to hospitality and tourism literature by developing a comprehensive customer satisfaction analysis framework in the big data era. Moreover, to the best of the authors’ knowledge, this work is among the first to introduce opportunity algorithm to quantify service improvement benefits. The proposed Cross-BERT model also advances the methodological literature on measuring customer satisfaction.

Details

International Journal of Contemporary Hospitality Management, vol. 36 no. 3
Type: Research Article
ISSN: 0959-6119

Keywords

Article
Publication date: 13 December 2022

Chengxi Yan, Xuemei Tang, Hao Yang and Jun Wang

The majority of existing studies about named entity recognition (NER) concentrate on the prediction enhancement of deep neural network (DNN)-based models themselves, but the…

Abstract

Purpose

The majority of existing studies about named entity recognition (NER) concentrate on the prediction enhancement of deep neural network (DNN)-based models themselves, but the issues about the scarcity of training corpus and the difficulty of annotation quality control are not fully solved, especially for Chinese ancient corpora. Therefore, designing a new integrated solution for Chinese historical NER, including automatic entity extraction and man-machine cooperative annotation, is quite valuable for improving the effectiveness of Chinese historical NER and fostering the development of low-resource information extraction.

Design/methodology/approach

The research provides a systematic approach for Chinese historical NER with a three-stage framework. In addition to the stage of basic preprocessing, the authors create, retrain and yield a high-performance NER model only using limited labeled resources during the stage of augmented deep active learning (ADAL), which entails three steps—DNN-based NER modeling, hybrid pool-based sampling (HPS) based on the active learning (AL), and NER-oriented data augmentation (DA). ADAL is thought to have the capacity to maintain the performance of DNN as high as possible under the few-shot constraint. Then, to realize machine-aided quality control in crowdsourcing settings, the authors design a stage of globally-optimized automatic label consolidation (GALC). The core of GALC is a newly-designed label consolidation model called simulated annealing-based automatic label aggregation (“SA-ALC”), which incorporates the factors of worker reliability and global label estimation. The model can assure the annotation quality of those data from a crowdsourcing annotation system.

Findings

Extensive experiments on two types of Chinese classical historical datasets show that the authors’ solution can effectively reduce the corpus dependency of a DNN-based NER model and alleviate the problem of label quality. Moreover, the results also show the superior performance of the authors’ pipeline approaches (i.e. HPS + DA and SA-ALC) compared to equivalent baselines in each stage.

Originality/value

The study sheds new light on the automatic extraction of Chinese historical entities in an all-technological-process integration. The solution is helpful to effectively reducing the annotation cost and controlling the labeling quality for the NER task. It can be further applied to similar tasks of information extraction and other low-resource fields in theoretical and practical ways.

Details

Aslib Journal of Information Management, vol. 75 no. 3
Type: Research Article
ISSN: 2050-3806

Keywords

Article
Publication date: 29 December 2023

B. Vasavi, P. Dileep and Ulligaddala Srinivasarao

Aspect-based sentiment analysis (ASA) is a task of sentiment analysis that requires predicting aspect sentiment polarity for a given sentence. Many traditional techniques use…

Abstract

Purpose

Aspect-based sentiment analysis (ASA) is a task of sentiment analysis that requires predicting aspect sentiment polarity for a given sentence. Many traditional techniques use graph-based mechanisms, which reduce prediction accuracy and introduce large amounts of noise. The other problem with graph-based mechanisms is that for some context words, the feelings change depending on the aspect, and therefore it is impossible to draw conclusions on their own. ASA is challenging because a given sentence can reveal complicated feelings about multiple aspects.

Design/methodology/approach

This research proposed an optimized attention-based DL model known as optimized aspect and self-attention aware long short-term memory for target-based semantic analysis (OAS-LSTM-TSA). The proposed model goes through three phases: preprocessing, aspect extraction and classification. Aspect extraction is done using a double-layered convolutional neural network (DL-CNN). The optimized aspect and self-attention embedded LSTM (OAS-LSTM) is used to classify aspect sentiment into three classes: positive, neutral and negative.

Findings

To detect and classify sentiment polarity of the aspect using the optimized aspect and self-attention embedded LSTM (OAS-LSTM) model. The results of the proposed method revealed that it achieves a high accuracy of 95.3 per cent for the restaurant dataset and 96.7 per cent for the laptop dataset.

Originality/value

The novelty of the research work is the addition of two effective attention layers in the network model, loss function reduction and accuracy enhancement, using a recent efficient optimization algorithm. The loss function in OAS-LSTM is minimized using the adaptive pelican optimization algorithm, thus increasing the accuracy rate. The performance of the proposed method is validated on four real-time datasets, Rest14, Lap14, Rest15 and Rest16, for various performance metrics.

Details

Data Technologies and Applications, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2514-9288

Keywords

Article
Publication date: 28 November 2023

Tingting Tian, Hongjian Shi, Ruhui Ma and Yuan Liu

For privacy protection, federated learning based on data separation allows machine learning models to be trained on remote devices or in isolated data devices. However, due to the…

Abstract

Purpose

For privacy protection, federated learning based on data separation allows machine learning models to be trained on remote devices or in isolated data devices. However, due to the limited resources such as bandwidth and power of local devices, communication in federated learning can be much slower than in local computing. This study aims to improve communication efficiency by reducing the number of communication rounds and the size of information transmitted in each round.

Design/methodology/approach

This paper allows each user node to perform multiple local trainings, then upload the local model parameters to a central server. The central server updates the global model parameters by weighted averaging the parameter information. Based on this aggregation, user nodes first cluster the parameter information to be uploaded and then replace each value with the mean value of its cluster. Considering the asymmetry of the federated learning framework, adaptively select the optimal number of clusters required to compress the model information.

Findings

While maintaining the loss convergence rate similar to that of federated averaging, the test accuracy did not decrease significantly.

Originality/value

By compressing uplink traffic, the work can improve communication efficiency on dynamic networks with limited resources.

Details

International Journal of Web Information Systems, vol. 20 no. 1
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 14 July 2023

Yang Gao, Wanqi Zheng and Yaojun Wang

This study aims to explore the risk spillover effects among different sectors of the Chinese stock market after the outbreak of COVID-19 from both Internet sentiment and price…

132

Abstract

Purpose

This study aims to explore the risk spillover effects among different sectors of the Chinese stock market after the outbreak of COVID-19 from both Internet sentiment and price fluctuations.

Design/methodology/approach

The authors develop four indicators used for risk contagion analysis, including Internet investors and news sentiments constructed by the FinBERT model, together with realized and jump volatilities yielded by high-frequency data. The authors also apply the time-varying parameter vector autoregressive (TVP-VAR) model-based and the tail-based connectedness framework to investigate the interdependence of tail risk during catastrophic events.

Findings

The empirical analysis provides meaningful results related to the COVID-19 pandemic, stock market conditions and tail behavior. The results show that after the outbreak of COVID-19, the connectivity between risk spillovers in China's stock market has grown, indicating the increased instability of the connected system and enhanced connectivity in the tail. The changes in network structure during COVID-19 pandemic are not only reflected by the increased spillover connectivity but also by the closer relationships between some industries. The authors also found that major public events could significantly impact total connectedness. In addition, spillovers and network structures vary with market conditions and tend to exhibit a highly connected network structure during extreme market status.

Originality/value

The results confirm the connectivity between sentiments and volatilities spillovers in China's stock market, especially in the tails. The conclusion further expands the practical application and theoretical framework of behavioral finance and also lays a theoretical basis for investors to focus on the practical application of volatility prediction and risk management across stock sectors.

Details

China Finance Review International, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2044-1398

Keywords

Article
Publication date: 19 May 2023

Jie Meng

This paper aims to quantify the quality of peer reviews, evaluate them from different perspectives and develop a model to predict the review quality. In addition, this paper…

Abstract

Purpose

This paper aims to quantify the quality of peer reviews, evaluate them from different perspectives and develop a model to predict the review quality. In addition, this paper investigates effective features to distinguish the reviews' quality. 

Design/methodology/approach

First, a fine-grained data set including peer review data, citations and review conformity scores was constructed. Second, metrics were proposed to evaluate the quality of peer reviews from three aspects. Third, five categories of features were proposed in terms of reviews, submissions and responses using natural language processing (NLP) techniques. Finally, different machine learning models were applied to predict the review quality, and feature analysis was performed to understand effective features.

Findings

The analysis results revealed that reviewers become more conservative and the review quality becomes worse over time in terms of these indicators. Among the three models, random forest model achieves the best performance on all three tasks. Sentiment polarity, review length, response length and readability are important factors that distinguish peer reviews’ quality, which can help meta-reviewers value more worthy reviews when making final decisions.

Originality/value

This study provides a new perspective for assessing review quality. Another originality of the research lies in the proposal of a novelty task that predict review quality. To address this task, a novel model was proposed which incorporated various of feature sets, thereby deepening the understanding of peer reviews.

Details

The Electronic Library , vol. 41 no. 2/3
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 4 August 2023

Jingxuan Huang, Qinyi Dong, Jiaxing Li and Lele Kang

While the growth of emerging technologies like Blockchain has created significant market opportunities and economic incentives for firms, it is valuable for both researchers and…

Abstract

Purpose

While the growth of emerging technologies like Blockchain has created significant market opportunities and economic incentives for firms, it is valuable for both researchers and practitioners to understand their creation mechanisms. This paper aims to discuss the aforementioned objective.

Design/methodology/approach

Based on the knowledge search perspective, this study examines the impact of search boundary on innovation novelty and quality. Additionally, innovation targets, namely R&D innovation and application innovation, are proposed as the moderator of the knowledge search effect. Using a combination of machine learning algorithms such as natural language processing and classification models, the authors propose new methods to measure the identified concepts.

Findings

The empirical results of 3,614 Blockchain patents indicate that search boundary enhances both innovation novelty and innovation quality. For R&D innovation, the positive impact of search boundary on innovation quality is enhanced, whereas for application innovation, the positive effect of search boundary on innovation novelty is improved.

Originality/value

This study mainly contributes to the growing literature on emerging technologies by describing their creation mechanisms. Specifically, the exploration of R&D and application taxonomy enriches researchers' understanding of knowledge search in the context of Blockchain invention.

Details

Industrial Management & Data Systems, vol. 123 no. 9
Type: Research Article
ISSN: 0263-5577

Keywords

21 – 30 of 212