Search results
1 – 10 of over 3000Nurlaila, Syahron Lubis, Tengku Sylvana Sinar and Muhizar Muchtar
Purpose – This paper is aimed at describing semantics equivalence of cultural terms in meurukon texts translated from Acehnese into Indonesian. A qualitative descriptive approach…
Abstract
Purpose – This paper is aimed at describing semantics equivalence of cultural terms in meurukon texts translated from Acehnese into Indonesian. A qualitative descriptive approach is used to analyze the context of semantics equivalence in these texts: varied semantics structure, especially the ones caused by the cultural gap between the two languages.
Design/Methodology/Approach – This research is designed to be of qualitative descriptive nature, wherein data are documented and analyzed using various methods proposed by Miles, Huberman, and Saldana (2014), such as data condensation, data display, drawing and verifying conclusions. The researcher is considered the key instrument in the whole process. The source of the data collected is from meurukon texts and its translation that consists of 623 sentences: they mainly comprise words and phrases that contain semantics equivalence of cultural terms.
Findings – The result of the research shows that there are 129 cultural terms found in 623 sentences. Of the analyzed data, it is seen that only 16.66% of the data is not equivalent with the target text, while 83.34% words and phrases of meurukon text are equivalent. This suggests that as a result of translation, the meurukon text has high semantics or lexical equivalences with the target text.
Research Limitations/Implications – This research is focused on semantics equivalence found in meurukon texts. The semantic equivalence here only pertains to lexical meaning of nouns and adjectives by using componential analysis.
Practical Implications – The result can be used in a sample of ways for the analysis of semantics equivalence of cultural terms in meurukon text translated from Acehnese into Indonesian using componential analysis.
Originality/Value – This research identifies meurukon as an oral tradition of Acehnese culture, which is in the question and answer format about Islamic law in Aceh, specifically North Aceh.
Details
Keywords
Frida Nyqvist and Eva-Lena Lundgren-Henriksson
The purpose of this research is to explore how an industry is represented in multimodal public media narratives and to explore how this representation subsequently affects the…
Abstract
Purpose
The purpose of this research is to explore how an industry is represented in multimodal public media narratives and to explore how this representation subsequently affects the formation of public sense-giving space during a persisting crisis, such as a pandemic. The question asked is: how do the use of multimodality by public service media dynamically shape representations of industry identity during a persisting crisis?
Design/methodology/approach
This study made use of a multimodal approach. The verbal and visual media text on the restaurant industry during the COVID-19 pandemic that were published in Finland by the public service media distributor Yle were studied. Data published between March 2020 and March 2022 were analysed. The data consisted of 236 verbal texts, including 263 visuals.
Findings
Three narratives were identified– victim, servant and survivor – that construct power relations and depict the identity of the restaurant industry differently. It was argued that multimodal media narratives hold three meaning making functions: sentimentalizing, juxtaposing and nuancing industry characteristics. It was also argued that multimodal public service media narratives have wider implications in possibly shaping the future attractiveness of the industry and organizational members' understanding of their identity.
Originality/value
This research contributes to sensemaking literature in that it explores the role of power – explicitly or implicitly constructed through media narratives during crisis. Furthermore, this research contributes to sensemaking literature in that it shows how narratives take shape multimodally during a continuous crisis, and how this impacts the construction of industry identity.
Details
Keywords
Carolin Ischen, Theo B. Araujo, Hilde A.M. Voorveld, Guda Van Noort and Edith G. Smit
Virtual assistants are increasingly used for persuasive purposes, employing the different modalities of voice and text (or a combination of the two). In this study, the authors…
Abstract
Purpose
Virtual assistants are increasingly used for persuasive purposes, employing the different modalities of voice and text (or a combination of the two). In this study, the authors compare the persuasiveness of voice-and text-based virtual assistants. The authors argue for perceived human-likeness and cognitive load as underlying mechanisms that can explain why voice- and text-based assistants differ in their persuasive potential by suppressing the activation of consumers' persuasion knowledge.
Design/methodology/approach
A pre-registered online-experiment (n = 450) implemented a text-based and two voice-based (with and without interaction history displayed in text) virtual assistants.
Findings
Findings show that, contrary to expectations, a text-based assistant is perceived as more human-like compared to a voice-based assistant (regardless of whether the interaction history is displayed), which in turn positively influences brand attitudes and purchase intention. The authors also find that voice as a communication modality can increase persuasion knowledge by being cognitively more demanding in comparison to text.
Practical implications
Simply using voice as a presumably human cue might not suffice to give virtual assistants a human-like appeal. For the development of virtual assistants, it might be beneficial to actively engage consumers to increase awareness of persuasion.
Originality/value
The current study adds to the emergent research stream considering virtual assistants in explicitly exploring modality differences between voice and text (and a combination of the two) and provides insights into the effects of persuasion coming from virtual assistants.
Details
Keywords
Alenka Kavčič Čolić and Andreja Hari
The current predominant delivery format resulting from digitization is PDF, which is not appropriate for the blind, partially sighted and people who read on mobile devices. To…
Abstract
Purpose
The current predominant delivery format resulting from digitization is PDF, which is not appropriate for the blind, partially sighted and people who read on mobile devices. To meet the needs of both communities, as well as broader ones, alternative file formats are required. With the findings of the eBooks-On-Demand-Network Opening Publications for European Netizens project research, this study aims to improve access to digitized content for these communities.
Design/methodology/approach
In 2022, the authors conducted research on the digitization experiences of 13 EODOPEN partners at their organizations. The authors distributed the same sample of scans in English with different characteristics, and in accordance with Web content accessibility guidelines, the authors created 24 criteria to analyze their digitization workflows, output formats and optical character recognition (OCR) quality.
Findings
In this contribution, the authors present the results of a trial implementation among EODOPEN partners regarding their digitization workflows, used delivery file formats and the resulting quality of OCR results, depending on the type of digitization output file format. It was shown that partners using the OCR tool ABBYY FineReader Professional and producing scanning outputs in tagged PDF and PDF/UA formats achieved better results according to set criteria.
Research limitations/implications
The trial implementations were limited to 13 project partners’ organizations only.
Originality/value
This research paper can be a valuable contribution to the field of massive digitization practices, particularly in terms of improving the accessibility of the output delivery file formats.
Details
Keywords
Xuan Ji, Jiachen Wang and Zhijun Yan
Stock price prediction is a hot topic and traditional prediction methods are usually based on statistical and econometric models. However, these models are difficult to deal with…
Abstract
Purpose
Stock price prediction is a hot topic and traditional prediction methods are usually based on statistical and econometric models. However, these models are difficult to deal with nonstationary time series data. With the rapid development of the internet and the increasing popularity of social media, online news and comments often reflect investors’ emotions and attitudes toward stocks, which contains a lot of important information for predicting stock price. This paper aims to develop a stock price prediction method by taking full advantage of social media data.
Design/methodology/approach
This study proposes a new prediction method based on deep learning technology, which integrates traditional stock financial index variables and social media text features as inputs of the prediction model. This study uses Doc2Vec to build long text feature vectors from social media and then reduce the dimensions of the text feature vectors by stacked auto-encoder to balance the dimensions between text feature variables and stock financial index variables. Meanwhile, based on wavelet transform, the time series data of stock price is decomposed to eliminate the random noise caused by stock market fluctuation. Finally, this study uses long short-term memory model to predict the stock price.
Findings
The experiment results show that the method performs better than all three benchmark models in all kinds of evaluation indicators and can effectively predict stock price.
Originality/value
In this paper, this study proposes a new stock price prediction model that incorporates traditional financial features and social media text features which are derived from social media based on deep learning technology.
Details
Keywords
Matjaž Kragelj and Mirjana Kljajić Borštnar
The purpose of this study is to develop a model for automated classification of old digitised texts to the Universal Decimal Classification (UDC), using machine-learning methods.
Abstract
Purpose
The purpose of this study is to develop a model for automated classification of old digitised texts to the Universal Decimal Classification (UDC), using machine-learning methods.
Design/methodology/approach
The general research approach is inherent to design science research, in which the problem of UDC assignment of the old, digitised texts is addressed by developing a machine-learning classification model. A corpus of 70,000 scholarly texts, fully bibliographically processed by librarians, was used to train and test the model, which was used for classification of old texts on a corpus of 200,000 items. Human experts evaluated the performance of the model.
Findings
Results suggest that machine-learning models can correctly assign the UDC at some level for almost any scholarly text. Furthermore, the model can be recommended for the UDC assignment of older texts. Ten librarians corroborated this on 150 randomly selected texts.
Research limitations/implications
The main limitations of this study were unavailability of labelled older texts and the limited availability of librarians.
Practical implications
The classification model can provide a recommendation to the librarians during their classification work; furthermore, it can be implemented as an add-on to full-text search in the library databases.
Social implications
The proposed methodology supports librarians by recommending UDC classifiers, thus saving time in their daily work. By automatically classifying older texts, digital libraries can provide a better user experience by enabling structured searches. These contribute to making knowledge more widely available and useable.
Originality/value
These findings contribute to the field of automated classification of bibliographical information with the usage of full texts, especially in cases in which the texts are old, unstructured and in which archaic language and vocabulary are used.
Details
Keywords
Sara Lafia, David A. Bleckley and J. Trent Alexander
Many libraries and archives maintain collections of research documents, such as administrative records, with paper-based formats that limit the documents' access to in-person use…
Abstract
Purpose
Many libraries and archives maintain collections of research documents, such as administrative records, with paper-based formats that limit the documents' access to in-person use. Digitization transforms paper-based collections into more accessible and analyzable formats. As collections are digitized, there is an opportunity to incorporate deep learning techniques, such as Document Image Analysis (DIA), into workflows to increase the usability of information extracted from archival documents. This paper describes the authors' approach using digital scanning, optical character recognition (OCR) and deep learning to create a digital archive of administrative records related to the mortgage guarantee program of the Servicemen's Readjustment Act of 1944, also known as the G.I. Bill.
Design/methodology/approach
The authors used a collection of 25,744 semi-structured paper-based records from the administration of G.I. Bill Mortgages from 1946 to 1954 to develop a digitization and processing workflow. These records include the name and city of the mortgagor, the amount of the mortgage, the location of the Reconstruction Finance Corporation agent, one or more identification numbers and the name and location of the bank handling the loan. The authors extracted structured information from these scanned historical records in order to create a tabular data file and link them to other authoritative individual-level data sources.
Findings
The authors compared the flexible character accuracy of five OCR methods. The authors then compared the character error rate (CER) of three text extraction approaches (regular expressions, DIA and named entity recognition (NER)). The authors were able to obtain the highest quality structured text output using DIA with the Layout Parser toolkit by post-processing with regular expressions. Through this project, the authors demonstrate how DIA can improve the digitization of administrative records to automatically produce a structured data resource for researchers and the public.
Originality/value
The authors' workflow is readily transferable to other archival digitization projects. Through the use of digital scanning, OCR and DIA processes, the authors created the first digital microdata file of administrative records related to the G.I. Bill mortgage guarantee program available to researchers and the general public. These records offer research insights into the lives of veterans who benefited from loans, the impacts on the communities built by the loans and the institutions that implemented them.
Details
Keywords
Alexandra Kirkby, Carsten Baumgarth and Jörg Henseler
This paper aims to explore consumer perception of “brand voice” authenticity, brand authenticity and brand attitude when the source of text is disclosed as either artificial…
Abstract
Purpose
This paper aims to explore consumer perception of “brand voice” authenticity, brand authenticity and brand attitude when the source of text is disclosed as either artificial intelligence (AI)-generated or human-written.
Design/methodology/approach
A 3 × 3 experimental design using Adidas marketing texts disclosed as either “AI” or “human”, or not disclosed was applied to data gathered online from 624 English-speaking students.
Findings
Text disclosed as AI-generated is not perceived as less authentic than that disclosed as human-written. No negative effect on brand voice authenticity and brand attitude results if an AI-source is disclosed.
Practical implications
Findings offer brand managers the potential for cost and time savings but emphasise the strong effect of AI technology on perceived brand authenticity and brand attitude.
Originality/value
Results show that brands can afford to be transparent in disclosing the use of AI to support brand voice as communicated in product description or specification or in chatbot text.
Details
Keywords
The Internet has changed consumer decision-making and influenced business behaviour. User-generated product information is abundant and readily available. This paper argues that…
Abstract
Purpose
The Internet has changed consumer decision-making and influenced business behaviour. User-generated product information is abundant and readily available. This paper argues that user-generated content can be efficiently utilised for business intelligence using data science and develops an approach to demonstrate the methods and benefits of the different techniques.
Design/methodology/approach
Using Python Selenium, Beautiful Soup and various text mining approaches in R to access, retrieve and analyse user-generated content, we argue that (1) companies can extract information about the product attributes that matter most to consumers and (2) user-generated reviews enable the use of text mining results in combination with other demographic and statistical information (e.g. ratings) as an efficient input for competitive analysis.
Findings
The paper shows that combining different types of data (textual and numerical data) and applying and combining different methods can provide organisations with important business information and improve business performance.
Research limitations/implications
The paper shows that combining different types of data (textual and numerical data) and applying and combining different methods can provide organisations with important business information and improve business performance.
Originality/value
The study makes several contributions to the marketing and management literature, mainly by illustrating the methodological advantages of text mining and accompanying statistical analysis, the different types of distilled information and their use in decision-making.
Details
Keywords
Shreyesh Doppalapudi, Tingyan Wang and Robin Qiu
Clinical notes typically contain medical jargons and specialized words and phrases that are complicated and technical to most people, which is one of the most challenging…
Abstract
Purpose
Clinical notes typically contain medical jargons and specialized words and phrases that are complicated and technical to most people, which is one of the most challenging obstacles in health information dissemination to consumers by healthcare providers. The authors aim to investigate how to leverage machine learning techniques to transform clinical notes of interest into understandable expressions.
Design/methodology/approach
The authors propose a natural language processing pipeline that is capable of extracting relevant information from long unstructured clinical notes and simplifying lexicons by replacing medical jargons and technical terms. Particularly, the authors develop an unsupervised keywords matching method to extract relevant information from clinical notes. To automatically evaluate completeness of the extracted information, the authors perform a multi-label classification task on the relevant texts. To simplify lexicons in the relevant text, the authors identify complex words using a sequence labeler and leverage transformer models to generate candidate words for substitution. The authors validate the proposed pipeline using 58,167 discharge summaries from critical care services.
Findings
The results show that the proposed pipeline can identify relevant information with high completeness and simplify complex expressions in clinical notes so that the converted notes have a high level of readability but a low degree of meaning change.
Social implications
The proposed pipeline can help healthcare consumers well understand their medical information and therefore strengthen communications between healthcare providers and consumers for better care.
Originality/value
An innovative pipeline approach is developed to address the health literacy problem confronted by healthcare providers and consumers in the ongoing digital transformation process in the healthcare industry.
Details