Search results
1 – 10 of over 5000Ming Li, Lisheng Chen and Yingcheng Xu
A large number of questions are posted on community question answering (CQA) websites every day. Providing a set of core questions will ease the question overload problem. These…
Abstract
Purpose
A large number of questions are posted on community question answering (CQA) websites every day. Providing a set of core questions will ease the question overload problem. These core questions should cover the main content of the original question set. There should be low redundancy within the core questions and a consistent distribution with the original question set. The paper aims to discuss these issues.
Design/methodology/approach
In the paper, a method named QueExt method for extracting core questions is proposed. First, questions are modeled using a biterm topic model. Then, these questions are clustered based on particle swarm optimization (PSO). With the clustering results, the number of core questions to be extracted from each cluster can be determined. Afterwards, the multi-objective PSO algorithm is proposed to extract the core questions. Both PSO algorithms are integrated with operators in genetic algorithms to avoid the local optimum.
Findings
Extensive experiments on real data collected from the famous CQA website Zhihu have been conducted and the experimental results demonstrate the superior performance over other benchmark methods.
Research limitations/implications
The proposed method provides new insight into and enriches research on information overload in CQA. It performs better than other methods in extracting core short text documents, and thus provides a better way to extract core data. The PSO is a novel method used for selecting core questions. The research on the application of the PSO model is expanded. The study also contributes to research on PSO-based clustering. With the integration of K-means++, the key parameter number of clusters is optimized.
Originality/value
The novel core question extraction method in CQA is proposed, which provides a novel and efficient way to alleviate the question overload. The PSO model is extended and novelty used in selecting core questions. The PSO model is integrated with K-means++ method to optimize the number of clusters, which is just the key parameter in text clustering based on PSO. It provides a new way to cluster texts.
Details
Keywords
Debasis Majhi and Bhaskar Mukherjee
The purpose of this study is to identify the research fronts by analysing highly cited core papers adjusted with the age of a paper in library and information science (LIS) where…
Abstract
Purpose
The purpose of this study is to identify the research fronts by analysing highly cited core papers adjusted with the age of a paper in library and information science (LIS) where natural language processing (NLP) is being applied significantly.
Design/methodology/approach
By excavating international databases, 3,087 core papers that received at least 5% of the total citations have been identified. By calculating the average mean years of these core papers, and total citations received, a CPT (citation/publication/time) value was calculated in all 20 fronts to understand how a front is relatively receiving greater attention among peers within a course of time. One theme article has been finally identified from each of these 20 fronts.
Findings
Bidirectional encoder representations from transformers with CPT value 1.608 followed by sentiment analysis with CPT 1.292 received highest attention in NLP research. Columbia University New York, in terms of University, Journal of the American Medical Informatics Association, in terms of journals, USA followed by People Republic of China, in terms of country and Xu, H., University of Texas, in terms of author are the top in these fronts. It is identified that the NLP applications boost the performance of digital libraries and automated library systems in the digital environment.
Practical implications
Any research fronts that are identified in the findings of this paper may be used as a base for researchers who intended to perform extensive research on NLP.
Originality/value
To the best of the authors’ knowledge, the methodology adopted in this paper is the first of its kind where meta-analysis approach has been used for understanding the research fronts in sub field like NLP for a broad domain like LIS.
Details
Keywords
Pompeu Casanovas, Marta Poblet, Núria Casellas, Jesus Contreras, V. Richard Benjamins and Mercedes Blazquez
In this paper we describe the process of developing and implementing a knowledge management system for the Spanish judicial domain. Spanish judges, especially newly‐recruited…
Abstract
Purpose
In this paper we describe the process of developing and implementing a knowledge management system for the Spanish judicial domain. Spanish judges, especially newly‐recruited ones, hold a solid background of theoretical legal knowledge, but are much less familiar with the judicial knowledge of the more senior judges acquired from everyday practice and case resolution. The aim of this development is to capture and model these two aspects of judicial knowledge – theoretical and practical – for knowledge browsing and retrieving.
Design/methodology/approach
Semantic web technologies are applied to feed a question‐answering system based on ontologies of professional legal knowledge (OPLK).
Findings
There is a kind of specific legal knowledge, which belongs properly to the expert domain, not being captured by current legal core ontologies, i.e. Judges require clues, hints or well‐grounded practical guidelines that refer to the problem they have before them when they put a question or start the query. A scalable and useful frequently‐asked questions system should have a simple, natural language interface, work in a real time environment, and the questions included in the system should be of high quality and reflect the current situation.
Originality/value
The final system will enable the users to ask queries in natural language and obtain answers, which are supported by legal documents stored in specialized legal databases. Special care is taken regarding usability issues, in order to ensure the highest user satisfaction.
Details
Keywords
Jinxiang Zeng, Shujin Cao, Yijin Chen, Pei Pan and Yafang Cai
This study analyzed the interdisciplinary characteristics of Chinese research studies in library and information science (LIS) measured by knowledge elements extracted through the…
Abstract
Purpose
This study analyzed the interdisciplinary characteristics of Chinese research studies in library and information science (LIS) measured by knowledge elements extracted through the Lexicon-LSTM model.
Design/methodology/approach
Eight research themes were selected for experiment, with a large-scale (N = 11,625) dataset of research papers from the China National Knowledge Infrastructure (CNKI) database constructed. And it is complemented with multiple corpora. Knowledge elements were extracted through a Lexicon-LSTM model. A subject knowledge graph is constructed to support the searching and classification of knowledge elements. An interdisciplinary-weighted average citation index space was constructed for measuring the interdisciplinary characteristics and contributions based on knowledge elements.
Findings
The empirical research shows that the Lexicon-LSTM model has superiority in the accuracy of extracting knowledge elements. In the field of LIS, the interdisciplinary diversity indicator showed an upward trend from 2011 to 2021, while the disciplinary balance and difference indicators showed a downward trend. The knowledge elements of theory and methodology could be used to detect and measure the interdisciplinary characteristics and contributions.
Originality/value
The extraction of knowledge elements facilitates the discovery of semantic information embedded in academic papers. The knowledge elements were proved feasible for measuring the interdisciplinary characteristics and exploring the changes in the time sequence, which helps for overview the state of the arts and future development trend of the interdisciplinary of research theme in LIS.
Details
Keywords
The purpose of this paper is to interpret organizational change from a co-evolutionary perspective. It examines the co-evolution between institutional environments and…
Abstract
Purpose
The purpose of this paper is to interpret organizational change from a co-evolutionary perspective. It examines the co-evolution between institutional environments and organizational change with the mediating role of uncertainty as perceived by managers.
Design/methodology/approach
The author employed an inductive case study to explore how institutional environments interact with organizational change in a novel context: a Chinese state-owned enterprise.
Findings
The author developed a co-evolutionary model of organizational change that emphasizes the interaction between institutional-level factors and organizational-level change as bridged by top management perceptions of uncertainty. The model also illustrates the dynamics of organizational uncertainty and its effects on organizational change.
Practical implications
The study implies that uncertainty may not be an inevitable negative influence on organizational development, and tell managers how to manages the dynamics of uncertainty through two principles.
Originality/value
This study contributes to the organizational change literature by interpreting organizational change as the results of interaction between multi-level factors from institutional, organizational, and team levels. The author also expand the understanding of uncertainty from a dynamic perspective.
Details
Keywords
Richard Entlich, Lorrin Garson, Michael Lesk, Lorraine Normore, Jan Olsen and Stuart Weibel
The Chemistry Online Retrieval Experiment (CORE), a five‐year R&D project, was one of the earliest attempts to make a substantial volume of the text and graphics from previously…
Abstract
The Chemistry Online Retrieval Experiment (CORE), a five‐year R&D project, was one of the earliest attempts to make a substantial volume of the text and graphics from previously published scholarly journals available to end‐users in electronic form, across a computer network. Since CORE dealt with material that had already gone through traditional print publication, its emphasis was on the process (and limitations) of conversion, the optimization of presentation, and use of the converted contents for readers. This article focuses on the user response to the system.
Strategic alliances among organizations are some of the central drivers of innovation and economic growth. However, the discovery of alliances has relied on pure manual search and…
Abstract
Purpose
Strategic alliances among organizations are some of the central drivers of innovation and economic growth. However, the discovery of alliances has relied on pure manual search and has limited scope. This paper proposes a text-mining framework, ACRank, that automatically extracts alliances from news articles. ACRank aims to provide human analysts with a higher coverage of strategic alliances compared to existing databases, yet maintain a reasonable extraction precision. It has the potential to discover alliances involving less well-known companies, a situation often neglected by commercial databases.
Design/methodology/approach
The proposed framework is a systematic process of alliance extraction and validation using natural language processing techniques and alliance domain knowledge. The process integrates news article search, entity extraction, and syntactic and semantic linguistic parsing techniques. In particular, Alliance Discovery Template (ADT) identifies a number of linguistic templates expanded from expert domain knowledge and extract potential alliances at sentence-level. Alliance Confidence Ranking (ACRank)further validates each unique alliance based on multiple features at document-level. The framework is designed to deal with extremely skewed, noisy data from news articles.
Findings
In evaluating the performance of ACRank on a gold standard data set of IBM alliances (2006–2008) showed that: Sentence-level ADT-based extraction achieved 78.1% recall and 44.7% precision and eliminated over 99% of the noise in news articles. ACRank further improved precision to 97% with the top20% of extracted alliance instances. Further comparison with Thomson Reuters SDC database showed that SDC covered less than 20% of total alliances, while ACRank covered 67%. When applying ACRank to Dow 30 company news articles, ACRank is estimated to achieve a recall between 0.48 and 0.95, and only 15% of the alliances appeared in SDC.
Originality/value
The research framework proposed in this paper indicates a promising direction of building a comprehensive alliance database using automatic approaches. It adds value to academic studies and business analyses that require in-depth knowledge of strategic alliances. It also encourages other innovative studies that use text mining and data analytics to study business relations.
Details
Keywords
Rafael Batista Duarte, Denis Silva da Silveira, Vinícius de Albuquerque Brito and Charlie Silva Lopes
Business process modeling can involve multiple stakeholders, so it is natural that problems may occur during the designing and understanding processes. One way to perceive these…
Abstract
Purpose
Business process modeling can involve multiple stakeholders, so it is natural that problems may occur during the designing and understanding processes. One way to perceive these problems is to evaluate the comprehension of business process models through the collection of data related to the readers' eye movement via an eye-tracking device. The purpose of this paper is to provide a comprehensive overview of the use of eye-trackers in understanding process models and to offer a research roadmap to challenge the community to address the identified limitations and open issues that require further investigation.
Design/methodology/approach
To achieve this goal, Systematic Literature Review (SLR) was performed following good practices from the Evidence-Based Software Engineering's (EBSE) field.
Findings
This study resulted in 10 primary studies selected for analysis and data extraction, from the 1,482 initially retrieved. The major findings indicate that the business process community still benefits little from the use of eye-tracking, e.g. not offering sufficient support for inexperienced designers and not having an explicit standardization in its use. These and other findings are synthesized in a research roadmap which results would benefit researchers and practitioners.
Originality/value
In the studies found, the methods used to explore eyes' movement in process models' comprehension analysis were presented as an advantage of the current study. Additionally, another aspect presented in this SRL as an originality is presenting a set of open questions, suggesting valuable topics for future research through a research script (research roadmap).
Details
Keywords
This paper aims to consider the potential implications of the layering of regulation in relation to hydraulic fracturing (fracking) at the borders between the nations of the UK.
Abstract
Purpose
This paper aims to consider the potential implications of the layering of regulation in relation to hydraulic fracturing (fracking) at the borders between the nations of the UK.
Design/methodology/approach
This paper uses a qualitative research method grounded in particular in legal geography to examine the existing approaches to regulating hydraulic fracturing and identify the places and their features that are constructed as a result of their intersection at the borders of the nations comprising the UK.
Findings
The current regulatory framework concerning hydraulic fracturing risks restricts the places in which the practice can occur in such a manner as to potentially cause greater environmental harm should the process be used. The regulations governing the process are not aligned in relation to the surface and subsurface aspects of the process to enable their management, once operational, as a singularly constructed place of extraction. Strong regulation at the surface can have the effect of influencing placement of the site only in relation to the place at which the resource sought reaches the surface, whilst having little to no impact on the environmental harms, which will result at the subsurface or relative to other potential surface site positions, and potentially even increasing them.
Research limitations/implications
This paper is limited by uncertainty as to the future use of hydraulic fracturing to extract oil and gas within the UK. The issues raised within it would also be applicable to other extractive industries where a surface site might be placed within a radius of the subsurface point of extraction, rather than having to be located at a fixed point relative to that in the subsurface. This paper therefore raises concerns that might be explored more generally in relation to the regulation of the place of resource extraction, particularly at legal borders between jurisdictions, and the impact of regulation, which does not account for the misalignment of regulation of spaces above and below the surface that form a single place at which extraction occurs.
Social implications
This paper considers the potential impacts of misaligned positions held by nations in the UK in relation to environmentally harmful practices undertaken by extractive industries, which are highlighted by an analysis of the extant regulatory framework for hydraulic fracturing.
Originality/value
Whilst the potential for cross internal border extraction of gas within the UK via hydraulic fracturing and the regulatory consequences of this has been highlighted in academic literature, this paper examines the implications of regulation for the least environmentally harmful placement of the process.
Details
Keywords
The purpose of this paper is to consider the ethical and environmental implications of allowing space resource extraction to disrupt existing fuel economies, including how…
Abstract
Purpose
The purpose of this paper is to consider the ethical and environmental implications of allowing space resource extraction to disrupt existing fuel economies, including how companies can be held accountable for ensuring the responsible use of their space assets. It will also briefly consider how such assets should be taxed, and the cost/benefit analyses required to justify the considerable expense of supporting this emerging space industry.
Design/methodology/approach
This paper adopts theoretical bioethics methodologies to explore issues of normative ethics and the formulation of moral rules to govern individual, collective and institutional behaviour. Specifically, it considers social justice and social contract theory, consequentialist and deontological accounts of ethical evaluation. It also draws on sociological and organisational literature to discuss Dowling and Pfeffer’s (1975) and Suchman’s (1995) theories of pragmatic, cognitive and moral legitimacy as they may be applied to off-world mining regulations and the handling of space assets.
Findings
The findings of this conceptual paper indicate there is both a growing appetite for tighter resource extraction regulations to address climate change and wealth concentration globally, and an opportunity to establish and legitimise new ethical norms for commercial activity in space that can avoid some of the challenges currently facing fossil fuel divestment movements on Earth.
Originality/value
By adopting methodologies from theoretical bioethics, sociology and business studies, including applying a legitimacy lens to the issue of off-world mining, this paper synthesises existing knowledges from these fields and brings them to the new context of the future space resource economy.
Details