Search results

1 – 10 of over 9000
Article
Publication date: 24 September 2021

Nina Rizun, Aleksandra Revina and Vera G. Meister

This study aims to draw the attention of business process management (BPM) research and practice to the textual data generated in the processes and the potential of…

Abstract

Purpose

This study aims to draw the attention of business process management (BPM) research and practice to the textual data generated in the processes and the potential of meaningful insights extraction. The authors apply standard natural language processing (NLP) approaches to gain valuable knowledge in the form of business process (BP) complexity concept suggested in the study. It is built on the objective, subjective and meta-knowledge extracted from the BP textual data and encompassing semantics, syntax and stylistics. As a result, the authors aim to create awareness about cognitive, attention and reading efforts forming the textual data-based BP complexity. The concept serves as a basis for the development of various decision-support solutions for BP workers.

Design/methodology/approach

The starting point is an investigation of the complexity concept in the BPM literature to develop an understanding of the related complexity research and to put the textual data-based BP complexity in its context. Afterward, utilizing the linguistic foundations and the theory of situation awareness (SA), the concept is empirically developed and evaluated in a real-world application case using qualitative interview-based and quantitative data-based methods.

Findings

In the practical, real-world application, the authors confirmed that BP textual data could be used to predict BP complexity from the semantic, syntactic and stylistic viewpoints. The authors were able to prove the value of this knowledge about the BP complexity formed based on the (1) professional contextual experience of the BP worker enriched by the awareness of cognitive efforts required for BP execution (objective knowledge), (2) business emotions enriched by attention efforts (subjective knowledge) and (3) quality of the text, i.e. professionalism, expertise and stress level of the text author, enriched by reading efforts (meta-knowledge). In particular, the BP complexity concept has been applied to an industrial example of Information Technology Infrastructure Library (ITIL) change management (CHM) Information Technology (IT) ticket processing. The authors used IT ticket texts from two samples of 28,157 and 4,625 tickets as the basis for the analysis. The authors evaluated the concept with the help of manually labeled tickets and a rule-based approach using historical ticket execution data. Having a recommendation character, the results showed to be useful in creating awareness regarding cognitive, attention and reading efforts for ITIL CHM BP workers coordinating the IT ticket processing.

Originality/value

While aiming to draw attention to those valuable insights inherent in BP textual data, the authors propose an unconventional approach to BP complexity definition through the lens of textual data. Hereby, the authors address the challenges specified by BPM researchers, i.e. focus on semantics in the development of vocabularies and organization- and sector-specific adaptation of standard NLP techniques.

Details

Business Process Management Journal, vol. 27 no. 7
Type: Research Article
ISSN: 1463-7154

Keywords

Article
Publication date: 13 May 2014

Margit Raich, Julia Müller and Dagmar Abfalter

The purpose of this paper is to provide insightful evidence of phenomena in organization and management theory. Textual data sets consist of two different elements, namely…

2219

Abstract

Purpose

The purpose of this paper is to provide insightful evidence of phenomena in organization and management theory. Textual data sets consist of two different elements, namely qualitative and quantitative aspects. Researchers often combine methods to harness both aspects. However, they frequently do this in a comparative, convergent, or sequential way.

Design/methodology/approach

The paper illustrates and discusses a hybrid textual data analysis approach employing the qualitative software application GABEK-WinRelan in a case study of an Austrian retail bank.

Findings

The paper argues that a hybrid analysis method, fully intertwining qualitative and quantitative analysis simultaneously on the same textual data set, can deliver new insight into more facets of a data set.

Originality/value

A hybrid approach is not a universally applicable solution to approaching research and management problems. Rather, this paper aims at triggering and intensifying scientific discussion about stronger integration of qualitative and quantitative data and analysis methods in management research.

Details

Management Decision, vol. 52 no. 4
Type: Research Article
ISSN: 0025-1747

Keywords

Article
Publication date: 1 March 2005

Rudolf R. Sinkovics, Elfriede Penz and Pervez N. Ghauri

To provide guidance for the formalised analysis of qualitative data and observations, to raise awareness about systematic analysis and illustrate promising avenues for the…

11512

Abstract

Purpose

To provide guidance for the formalised analysis of qualitative data and observations, to raise awareness about systematic analysis and illustrate promising avenues for the application of qualitative methodologies in international marketing research.

Design/methodology/approach

Conceptually, the nature of qualitative research, globalisation and its implications for the research landscape, text‐data as a source for international research and equivalence issues in international qualitative research are discussed. The methodology section applies these concepts and analysis challenges to a real‐world example using N*Vivo software.

Findings

A 14‐step analytic design is developed, introducing procedures of data analysis and interpretation which help to formalise qualitative research of textual data.

Research limitations/implications

The use of software programs (e.g. N*Vivo) helps to substantiate the analysis and interpretation process of textual data.

Practical implications

Step‐by‐step guidance on performing qualitative analysis of textual data and documenting findings.

Originality/value

The paper is valuable for researchers and practitioners looking for guidance in analysing and interpreting textual data from interviews. Specific support is given for N*Vivo software and its application.

Details

Qualitative Market Research: An International Journal, vol. 8 no. 1
Type: Research Article
ISSN: 1352-2752

Keywords

Article
Publication date: 17 August 2015

Takahiro Komamizu, Toshiyuki Amagasa and Hiroyuki Kitagawa

The purpose of this paper is to extract appropriate terms to summarize the current results in terms of the contents of textual facets. Faceted search on XML data helps…

Abstract

Purpose

The purpose of this paper is to extract appropriate terms to summarize the current results in terms of the contents of textual facets. Faceted search on XML data helps users find necessary information from XML data by giving attribute–content pairs (called facet-value pair) about the current search results. However, if most of the contents of a facet have longer texts in average (such facets are called textual facets), it is not easy to overview the current results.

Design/methodology/approach

The proposed approach is based upon subsumption relationships of terms among the contents of a facet. The subsumption relationship can be extracted using co-occurrences of terms among a number of documents (in this paper, a content of a facet is considered as a document). Subsumption relationships compose hierarchies, and the authors utilize the hierarchies to extract facet-values from textual facets. In the faceted search context, users have ambiguous search demands, they expect broader terms. Thus, we extract high-level terms in the hierarchies as facet-values.

Findings

The main findings of this paper are the extracted terms improve users’ search experiences, especially in cases when the search demands are ambiguous.

Originality/value

An originality of this paper is the way to utilize the textual contents of XML data for improving users’ search experiences on faceted search. The other originality is how to design the tasks to evaluate exploratory search like faceted search.

Details

International Journal of Web Information Systems, vol. 11 no. 3
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 1 June 2010

Jan Brase and Ina Blümel

The purpose of this paper is to describe the work being done at The German National Library of Science and Technology (TIB) to make non‐textual information, for example…

3604

Abstract

Purpose

The purpose of this paper is to describe the work being done at The German National Library of Science and Technology (TIB) to make non‐textual information, for example three dimensional objects, more easily accessible. The goal is to create workflows and develop tools that allow academic libraries to treat this data in the same way as textual documents within the library processing chain. This implies content‐based indexing and the offering of new kinds of interfaces for searching and displaying results.

Design/methodology/approach

The work of TIB on non textual information is described as well as DataCite and its launch in December 2009.

Findings

That the launch of Datacite ensures that this agency will take global leadership for promoting the use of persistent identifiers for datasets, to satisfy the needs of scientists. It will, through its members, establish and promote common methods, best practices, and guidance.

Practical implications

The work of TIB and the launch of Datacite will ensure that non textual data will become more easily available to researchers.

Originality/value

The value of this work has been underlined recently with the controversy over the accessibility of climate change data.

Details

Interlending & Document Supply, vol. 38 no. 2
Type: Research Article
ISSN: 0264-1615

Keywords

Article
Publication date: 14 May 2019

Georgia Boskou, Efstathios Kirkos and Charalambos Spathis

This paper aims to assess internal audit quality (IAQ) by using automated textual analysis of disclosures of internal audit mechanisms in annual reports.

1815

Abstract

Purpose

This paper aims to assess internal audit quality (IAQ) by using automated textual analysis of disclosures of internal audit mechanisms in annual reports.

Design/methodology/approach

This paper uses seven text mining techniques to construct classification models that predict whether companies listed on the Athens Stock Exchange are audited by a Big 4 firm, an auditor selection that prior research finds is associated with higher IAQ. The classification accuracy of the models is compared to predictions based on financial indicators.

Findings

The results show that classification models developed using text analysis can be a promising alternative proxy in assessing IAQ. Terms, N-Grams and financial indicators of a company, as they are presented in the annual reports, can provide information on the IAQ.

Practical implications

This study offers a novel approach to assessing the IAQ by applying textual analysis techniques. These findings are important for those who oversee internal audit activities, assess internal audit performance or want to improve or evaluate internal audit systems, such as managers or audit committees. Practitioners, regulators and investors may also extract useful information on internal audit and internal auditors by using textual analysis. The insights are also relevant for external auditors who are required to consider various aspects of corporate governance, including IAQ.

Originality/value

IAQ has been the subject of thorough examination. However, this study is the first attempt, to the authors’ knowledge, to introduce an innovative text mining approach utilizing unstructured textual disclosure from annual reports to develop a proxy for IAQ. It contributes to the internal audit field literature by further exploring concerns relevant to IAQ.

Details

Managerial Auditing Journal, vol. 34 no. 8
Type: Research Article
ISSN: 0268-6902

Keywords

Article
Publication date: 19 May 2020

Jui-Long Hung, Kerry Rice, Jennifer Kepka and Juan Yang

For studies in educational data mining or learning Analytics, the prediction of student’s performance or early warning is one of the most popular research topics. However…

Abstract

Purpose

For studies in educational data mining or learning Analytics, the prediction of student’s performance or early warning is one of the most popular research topics. However, research gaps indicate a paucity of research using machine learning and deep learning (DL) models in predictive analytics that include both behaviors and text analysis.

Design/methodology/approach

This study combined behavioral data and discussion board content to construct early warning models with machine learning and DL algorithms. In total, 680 course sections, 12,869 students and 14,951,368 logs were collected from a K-12 virtual school in the USA. Three rounds of experiments were conducted to demonstrate the effectiveness of the proposed approach.

Findings

The DL model performed better than machine learning models and was able to capture 51% of at-risk students in the eighth week with 86.8% overall accuracy. The combination of behavioral and textual data further improved the model’s performance in both recall and accuracy rates. The total word count is a more general indicator than the textual content feature. Successful students showed more words in analytic, and at-risk students showed more words in authentic when text was imported into a linguistic function word analysis tool. The balanced threshold was 0.315, which can capture up to 59% of at-risk students.

Originality/value

The results of this exploratory study indicate that the use of student behaviors and text in a DL approach may improve the predictive power of identifying at-risk learners early enough in the learning process to allow for interventions that can change the course of their trajectory.

Details

Information Discovery and Delivery, vol. 48 no. 4
Type: Research Article
ISSN: 2398-6247

Keywords

Article
Publication date: 2 December 2019

Fuli Zhou, Ming K. Lim, Yandong He and Saurabh Pratap

The increasingly booming e-commerce development has stimulated vehicle consumers to express individual reviews through online forum. The purpose of this paper is to probe…

Abstract

Purpose

The increasingly booming e-commerce development has stimulated vehicle consumers to express individual reviews through online forum. The purpose of this paper is to probe into the vehicle consumer consumption behavior and make recommendations for potential consumers from textual comments viewpoint.

Design/methodology/approach

A big data analytic-based approach is designed to discover vehicle consumer consumption behavior from online perspective. To reduce subjectivity of expert-based approaches, a parallel Naïve Bayes approach is designed to analyze the sentiment analysis, and the Saaty scale-based (SSC) scoring rule is employed to obtain specific sentimental value of attribute class, contributing to the multi-grade sentiment classification. To achieve the intelligent recommendation for potential vehicle customers, a novel SSC-VIKOR approach is developed to prioritize vehicle brand candidates from a big data analytical viewpoint.

Findings

The big data analytics argue that “cost-effectiveness” characteristic is the most important factor that vehicle consumers care, and the data mining results enable automakers to better understand consumer consumption behavior.

Research limitations/implications

The case study illustrates the effectiveness of the integrated method, contributing to much more precise operations management on marketing strategy, quality improvement and intelligent recommendation.

Originality/value

Researches of consumer consumption behavior are usually based on survey-based methods, and mostly previous studies about comments analysis focus on binary analysis. The hybrid SSC-VIKOR approach is developed to fill the gap from the big data perspective.

Details

Industrial Management & Data Systems, vol. 120 no. 1
Type: Research Article
ISSN: 0263-5577

Keywords

Article
Publication date: 23 March 2021

Hendri Murfi

The aim of this research is to develop an eigenspace-based fuzzy c-means method for scalable topic detection.

Abstract

Purpose

The aim of this research is to develop an eigenspace-based fuzzy c-means method for scalable topic detection.

Design/methodology/approach

The eigenspace-based fuzzy c-means (EFCM) combines representation learning and clustering. The textual data are transformed into a lower-dimensional eigenspace using truncated singular value decomposition. Fuzzy c-means is performed on the eigenspace to identify the centroids of each cluster. The topics are provided by transforming back the centroids into the nonnegative subspace of the original space. In this paper, we extend the EFCM method for scalability by using the two approaches, i.e. single-pass and online. We call the developed topic detection methods as oEFCM and spEFCM.

Findings

Our simulation shows that both oEFCM and spEFCM methods provide faster running times than EFCM for data sets that do not fit in memory. However, there is a decrease in the average coherence score. For both data sets that fit and do not fit into memory, the oEFCM method provides a tradeoff between running time and coherence score, which is better than spEFCM.

Originality/value

This research produces a scalable topic detection method. Besides this scalability capability, the developed method also provides a faster running time for the data set that fits in memory.

Details

Data Technologies and Applications, vol. 55 no. 4
Type: Research Article
ISSN: 2514-9288

Keywords

Book part
Publication date: 31 December 2003

Robert P Gephart

This comment provides a discovery-oriented reading of grounded theory which supplements the verification-oriented approach to level specification. I address how grounded…

Abstract

This comment provides a discovery-oriented reading of grounded theory which supplements the verification-oriented approach to level specification. I address how grounded theory can be used to discover new constructs, to surface properties of constructs, to validate constructs, and to enhance understanding of levels of analysis of constructs. Two approaches to the integration of qualitative and quantitative data in grounded theory are discussed: the linking approach and computer-aided interpretive textual analysis. The comment shows how multi-level constructs can be developed from real life interaction using discovery oriented grounded theory.

Details

Multi-Level Issues in Organizational Behavior and Strategy
Type: Book
ISBN: 978-0-76231-039-5

1 – 10 of over 9000