Search results

1 – 10 of over 5000
Open Access
Article
Publication date: 19 December 2023

Qinxu Ding, Ding Ding, Yue Wang, Chong Guan and Bosheng Ding

The rapid rise of large language models (LLMs) has propelled them to the forefront of applications in natural language processing (NLP). This paper aims to present a comprehensive…

1483

Abstract

Purpose

The rapid rise of large language models (LLMs) has propelled them to the forefront of applications in natural language processing (NLP). This paper aims to present a comprehensive examination of the research landscape in LLMs, providing an overview of the prevailing themes and topics within this dynamic domain.

Design/methodology/approach

Drawing from an extensive corpus of 198 records published between 1996 to 2023 from the relevant academic database encompassing journal articles, books, book chapters, conference papers and selected working papers, this study delves deep into the multifaceted world of LLM research. In this study, the authors employed the BERTopic algorithm, a recent advancement in topic modeling, to conduct a comprehensive analysis of the data after it had been meticulously cleaned and preprocessed. BERTopic leverages the power of transformer-based language models like bidirectional encoder representations from transformers (BERT) to generate more meaningful and coherent topics. This approach facilitates the identification of hidden patterns within the data, enabling authors to uncover valuable insights that might otherwise have remained obscure. The analysis revealed four distinct clusters of topics in LLM research: “language and NLP”, “education and teaching”, “clinical and medical applications” and “speech and recognition techniques”. Each cluster embodies a unique aspect of LLM application and showcases the breadth of possibilities that LLM technology has to offer. In addition to presenting the research findings, this paper identifies key challenges and opportunities in the realm of LLMs. It underscores the necessity for further investigation in specific areas, including the paramount importance of addressing potential biases, transparency and explainability, data privacy and security, and responsible deployment of LLM technology.

Findings

The analysis revealed four distinct clusters of topics in LLM research: “language and NLP”, “education and teaching”, “clinical and medical applications” and “speech and recognition techniques”. Each cluster embodies a unique aspect of LLM application and showcases the breadth of possibilities that LLM technology has to offer. In addition to presenting the research findings, this paper identifies key challenges and opportunities in the realm of LLMs. It underscores the necessity for further investigation in specific areas, including the paramount importance of addressing potential biases, transparency and explainability, data privacy and security, and responsible deployment of LLM technology.

Practical implications

This classification offers practical guidance for researchers, developers, educators, and policymakers to focus efforts and resources. The study underscores the importance of addressing challenges in LLMs, including potential biases, transparency, data privacy, and responsible deployment. Policymakers can utilize this information to shape regulations, while developers can tailor technology development based on the diverse applications identified. The findings also emphasize the need for interdisciplinary collaboration and highlight ethical considerations, providing a roadmap for navigating the complex landscape of LLM research and applications.

Originality/value

This study stands out as the first to examine the evolution of LLMs across such a long time frame and across such diversified disciplines. It provides a unique perspective on the key areas of LLM research, highlighting the breadth and depth of LLM’s evolution.

Details

Journal of Electronic Business & Digital Economics, vol. 3 no. 1
Type: Research Article
ISSN: 2754-4214

Keywords

Article
Publication date: 5 May 2023

Subhajit Panda and Navkiran Kaur

The purpose of this research paper is to explore the significance of language processing in library systems and evaluate the effectiveness of integrating artificial intelligence…

Abstract

Purpose

The purpose of this research paper is to explore the significance of language processing in library systems and evaluate the effectiveness of integrating artificial intelligence and generative pre-trained transformer (GPT) technology in modern libraries. Specifically, the paper focuses on SheetGPT, a Google Sheet and GPT Plugin and its impact on language processing in library systems.

Design/methodology/approach

This paper adopts a comprehensive analysis approach to evaluate the integration of SheetGPT in library systems. The authors outlined a user-friendly approach for installation and use of SheetGPT using its “beginner plan”, appropriate for personal/student use or extended experimentation. The study includes a quantitative analysis to provide a thorough understanding of the benefits and limitations of SheetGPT in library systems.

Findings

The findings of this research paper suggest that SheetGPT is a highly effective language-processing tool for library systems. Additionally, ChatGPT’s integration with Google Sheets and easy accessibility over Google Marketplace makes it an efficient and user-friendly tool for library professionals. Overall, this study highlights the potential of SheetGPT to enhance language processing in library systems

Originality/value

This research paper contributes to the existing literature by providing a comprehensive analysis of the effectiveness of SheetGPT in library systems. The study’s approach is unique in that it evaluates SheetGPT’s impact on language processing and provides insights into its benefits and limitations. The study’s findings are original and provide a valuable resource for library professionals and researchers interested in exploring the potential of SheetGPT to enhance language processing in library systems.

Details

Library Hi Tech News, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0741-9058

Keywords

Article
Publication date: 31 October 2023

Hong Zhou, Binwei Gao, Shilong Tang, Bing Li and Shuyu Wang

The number of construction dispute cases has maintained a high growth trend in recent years. The effective exploration and management of construction contract risk can directly…

Abstract

Purpose

The number of construction dispute cases has maintained a high growth trend in recent years. The effective exploration and management of construction contract risk can directly promote the overall performance of the project life cycle. The miss of clauses may result in a failure to match with standard contracts. If the contract, modified by the owner, omits key clauses, potential disputes may lead to contractors paying substantial compensation. Therefore, the identification of construction project contract missing clauses has heavily relied on the manual review technique, which is inefficient and highly restricted by personnel experience. The existing intelligent means only work for the contract query and storage. It is urgent to raise the level of intelligence for contract clause management. Therefore, this paper aims to propose an intelligent method to detect construction project contract missing clauses based on Natural Language Processing (NLP) and deep learning technology.

Design/methodology/approach

A complete classification scheme of contract clauses is designed based on NLP. First, construction contract texts are pre-processed and converted from unstructured natural language into structured digital vector form. Following the initial categorization, a multi-label classification of long text construction contract clauses is designed to preliminary identify whether the clause labels are missing. After the multi-label clause missing detection, the authors implement a clause similarity algorithm by creatively integrating the image detection thought, MatchPyramid model, with BERT to identify missing substantial content in the contract clauses.

Findings

1,322 construction project contracts were tested. Results showed that the accuracy of multi-label classification could reach 93%, the accuracy of similarity matching can reach 83%, and the recall rate and F1 mean of both can reach more than 0.7. The experimental results verify the feasibility of intelligently detecting contract risk through the NLP-based method to some extent.

Originality/value

NLP is adept at recognizing textual content and has shown promising results in some contract processing applications. However, the mostly used approaches of its utilization for risk detection in construction contract clauses predominantly are rule-based, which encounter challenges when handling intricate and lengthy engineering contracts. This paper introduces an NLP technique based on deep learning which reduces manual intervention and can autonomously identify and tag types of contractual deficiencies, aligning with the evolving complexities anticipated in future construction contracts. Moreover, this method achieves the recognition of extended contract clause texts. Ultimately, this approach boasts versatility; users simply need to adjust parameters such as segmentation based on language categories to detect omissions in contract clauses of diverse languages.

Details

Engineering, Construction and Architectural Management, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0969-9988

Keywords

Article
Publication date: 1 April 2024

Xiaoxian Yang, Zhifeng Wang, Qi Wang, Ke Wei, Kaiqi Zhang and Jiangang Shi

This study aims to adopt a systematic review approach to examine the existing literature on law and LLMs.It involves analyzing and synthesizing relevant research papers, reports…

Abstract

Purpose

This study aims to adopt a systematic review approach to examine the existing literature on law and LLMs.It involves analyzing and synthesizing relevant research papers, reports and scholarly articles that discuss the use of LLMs in the legal domain. The review encompasses various aspects, including an analysis of LLMs, legal natural language processing (NLP), model tuning techniques, data processing strategies and frameworks for addressing the challenges associated with legal question-and-answer (Q&A) systems. Additionally, the study explores potential applications and services that can benefit from the integration of LLMs in the field of intelligent justice.

Design/methodology/approach

This paper surveys the state-of-the-art research on law LLMs and their application in the field of intelligent justice. The study aims to identify the challenges associated with developing Q&A systems based on LLMs and explores potential directions for future research and development. The ultimate goal is to contribute to the advancement of intelligent justice by effectively leveraging LLMs.

Findings

To effectively apply a law LLM, systematic research on LLM, legal NLP and model adjustment technology is required.

Originality/value

This study contributes to the field of intelligent justice by providing a comprehensive review of the current state of research on law LLMs.

Details

International Journal of Web Information Systems, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 7 December 2023

Leanne Bowler, Irene Lopatovska and Mark S. Rosin

The purpose of this study is to explore teen-adult dialogic interactions during the co-design of data literacy activities in order to determine the nature of teen thinking, their…

Abstract

Purpose

The purpose of this study is to explore teen-adult dialogic interactions during the co-design of data literacy activities in order to determine the nature of teen thinking, their emotions, level of engagement, and the power of relationships between teens and adults in the context of data literacy. This study conceives of co-design as a learning space for data literacy. It investigates the teen–adult dialogic interactions and what these interactions say about the nature of teen thinking, their emotions, level of engagement and the power relationships between teens and adults.

Design/methodology/approach

The study conceives of co-design as a learning space for teens. Linguistic Inquiry and Word Count (LIWC-22), a natural language processing (NLP) software tool, was used to examine the linguistic measures of Analytic Thinking, Clout, Authenticity, and Emotional Tone using transcriptions of recorded Data Labs with teens and adults. Linguistic Inquiry and Word Count (LIWC-22), a natural language processing (NLP) software tool, was used to examine the linguistic measures of Analytic Thinking, Clout, Authenticity and Emotional Tone using transcriptions of recorded Data Labs with teens and adults.

Findings

LIWC-22 scores on the linguistic measures Analytic Thinking, Clout, Authenticity and Emotional Tone indicate that teens had a high level of friendly engagement, a relatively low sense of power compared with the adult co-designers, medium levels of spontaneity and honesty and the prevalence of positive emotions during the co-design sessions.

Practical implications

This study provides a concrete example of how to apply NLP in the context of data literacy in the public library, mapping the LIWC-22 findings to STEM-focused informal learning. It adds to the understanding of assessment/measurement tools and methods for designing data literacy education, stimulating further research and discussion on the ways to empower youth to engage more actively in informal learning about data.

Originality/value

This study applies a novel approach for exploring teen engagement within a co-design project tasked with the creation of youth-oriented data literacy activities.

Details

Information and Learning Sciences, vol. 125 no. 3/4
Type: Research Article
ISSN: 2398-5348

Keywords

Article
Publication date: 30 May 2023

R.V. ShabbirHusain, Atul Arun Pathak, Shabana Chandrasekaran and Balamurugan Annamalai

This study aims to explore the role of the linguistic style used in the brand-posted social media content on consumer engagement in the Fintech domain.

Abstract

Purpose

This study aims to explore the role of the linguistic style used in the brand-posted social media content on consumer engagement in the Fintech domain.

Design/methodology/approach

A total of 3,286 tweets (registering nearly 1.35 million impressions) published by 10 leading Fintech unicorns in India were extracted using the Twitter API. The Linguistic Inquiry and Word Count (LIWC) dictionary was used to analyse the linguistic characteristics of the shared tweets. Negative Binomial Regression (NBR) was used for testing the hypotheses.

Findings

This study finds that using drive words and cognitive language increases consumer engagement with Fintech messages via the central route of information processing. Further, affective words and conversational language drive consumer engagement through the peripheral route of information processing.

Research limitations/implications

The study extends the literature on brand engagement by unveiling the effect of linguistic features used to design social media messages.

Practical implications

The study provides guidance to social media marketers of Fintech brands regarding what content strategies best enhance consumer engagement. The linguistic style to improve online consumer engagement (OCE) is detailed.

Originality/value

The study’s findings contribute to the growing stream of Fintech literature by exploring the role of linguistic style on consumer engagement in social media communication. The study’s findings indicate the relevance of the dual processing mechanism of elaboration likelihood model (ELM) as an explanatory theory for evaluating consumer engagement with messages posted by Fintech brands.

Details

International Journal of Bank Marketing, vol. 42 no. 2
Type: Research Article
ISSN: 0265-2323

Keywords

Article
Publication date: 2 January 2024

Tiara Kusumaningtiyas, Prasetyo Adi Nugroho and Nurul Aida Noor Azizi

The purpose of this paper is to explore the use of artificial intelligence (AI) in libraries, especially university libraries, which are faced with users from various countries…

Abstract

Purpose

The purpose of this paper is to explore the use of artificial intelligence (AI) in libraries, especially university libraries, which are faced with users from various countries who have different languages and cultures. Seamless M4T, which is being developed, has great potential for helping university librarians maximize library services by providing ease of communication.

Design/methodology/approach

Analyzing the possibility of developing Seamless M4T using natural language processing techniques and how to train language models to be smarter AI tools and can be used to break down language barriers between librarians and users.

Findings

The implementation of AI-based application Seamless M4T can help university librarians provide maximum service to users who are hampered by language and culture with advanced communication skills. Seamless M4T has an automatic speech recognition feature for dozens of languages, so it can translate speech-to-text, text-to-speech or both text and speech. To convert written words into verbal forms, this AI can also translate and transcribe text and speech in real-time without significant delays.

Originality/value

This paper emphasizes the use of AI in university libraries to improve services, especially in communication due to language differences between librarians and users. Advantages in using AI in libraries can support the collaboration and scholarly communication process.

Details

Library Hi Tech News, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0741-9058

Keywords

Open Access
Article
Publication date: 31 July 2023

Daniel Šandor and Marina Bagić Babac

Sarcasm is a linguistic expression that usually carries the opposite meaning of what is being said by words, thus making it difficult for machines to discover the actual meaning…

2941

Abstract

Purpose

Sarcasm is a linguistic expression that usually carries the opposite meaning of what is being said by words, thus making it difficult for machines to discover the actual meaning. It is mainly distinguished by the inflection with which it is spoken, with an undercurrent of irony, and is largely dependent on context, which makes it a difficult task for computational analysis. Moreover, sarcasm expresses negative sentiments using positive words, allowing it to easily confuse sentiment analysis models. This paper aims to demonstrate the task of sarcasm detection using the approach of machine and deep learning.

Design/methodology/approach

For the purpose of sarcasm detection, machine and deep learning models were used on a data set consisting of 1.3 million social media comments, including both sarcastic and non-sarcastic comments. The data set was pre-processed using natural language processing methods, and additional features were extracted and analysed. Several machine learning models, including logistic regression, ridge regression, linear support vector and support vector machines, along with two deep learning models based on bidirectional long short-term memory and one bidirectional encoder representations from transformers (BERT)-based model, were implemented, evaluated and compared.

Findings

The performance of machine and deep learning models was compared in the task of sarcasm detection, and possible ways of improvement were discussed. Deep learning models showed more promise, performance-wise, for this type of task. Specifically, a state-of-the-art model in natural language processing, namely, BERT-based model, outperformed other machine and deep learning models.

Originality/value

This study compared the performance of the various machine and deep learning models in the task of sarcasm detection using the data set of 1.3 million comments from social media.

Details

Information Discovery and Delivery, vol. 52 no. 2
Type: Research Article
ISSN: 2398-6247

Keywords

Open Access
Article
Publication date: 6 April 2023

Karlo Puh and Marina Bagić Babac

Predicting the stock market's prices has always been an interesting topic since its closely related to making money. Recently, the advances in natural language processing (NLP…

4394

Abstract

Purpose

Predicting the stock market's prices has always been an interesting topic since its closely related to making money. Recently, the advances in natural language processing (NLP) have opened new perspectives for solving this task. The purpose of this paper is to show a state-of-the-art natural language approach to using language in predicting the stock market.

Design/methodology/approach

In this paper, the conventional statistical models for time-series prediction are implemented as a benchmark. Then, for methodological comparison, various state-of-the-art natural language models ranging from the baseline convolutional and recurrent neural network models to the most advanced transformer-based models are developed, implemented and tested.

Findings

Experimental results show that there is a correlation between the textual information in the news headlines and stock price prediction. The model based on the GRU (gated recurrent unit) cell with one linear layer, which takes pairs of the historical prices and the sentiment score calculated using transformer-based models, achieved the best result.

Originality/value

This study provides an insight into how to use NLP to improve stock price prediction and shows that there is a correlation between news headlines and stock price prediction.

Details

American Journal of Business, vol. 38 no. 2
Type: Research Article
ISSN: 1935-5181

Keywords

Article
Publication date: 25 July 2023

Aida Khakimova, Oleg Zolotarev and Sanjay Kaushal

Effective communication is crucial in the medical field where different stakeholders use various terminologies to describe and classify healthcare concepts such as ICD, SNOMED CT…

Abstract

Purpose

Effective communication is crucial in the medical field where different stakeholders use various terminologies to describe and classify healthcare concepts such as ICD, SNOMED CT, UMLS and MeSH, but the problem of polysemy can make natural language processing difficult. This study explores the contextual meanings of the term “pattern” in the biomedical literature, compares them to existing definitions, annotates a corpus for use in machine learning and proposes new definitions of terms such as “Syndrome, feature” and “pattern recognition.”

Design/methodology/approach

Entrez API was used to retrieve articles form PubMed for the study which assembled a corpus of 398 articles using a search query for the ambiguous term “pattern” in the titles or abstracts. The python NLTK library was used to extract the terms and their contexts, and an expert check was carried out. To understand the various meanings of the term, the contextual environment was analyzed by extracting the surrounding words of the term. The expert determined the appropriate size of the context for analysis to gain a more nuanced understanding of the different meanings of the term pattern.

Findings

The study found that the categories of meanings of the term “pattern” are broader in biomedical publications than in common definitions, and new categories have been emerging from the term's use in the biomedical field. The study highlights the importance of annotated corpora in advancing natural language processing techniques and provides valuable insights into the nuances of biomedical language.

Originality/value

The study's findings demonstrate the importance of exploring contextual meanings and proposing new definitions of terms in the biomedical field to improve natural language processing techniques.

Details

Kybernetes, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0368-492X

Keywords

1 – 10 of over 5000