Search results

1 – 10 of over 10000
Open Access
Article
Publication date: 14 August 2017

Xiu Susie Fang, Quan Z. Sheng, Xianzhi Wang, Anne H.H. Ngu and Yihong Zhang

This paper aims to propose a system for generating actionable knowledge from Big Data and use this system to construct a comprehensive knowledge base (KB), called GrandBase.

2053

Abstract

Purpose

This paper aims to propose a system for generating actionable knowledge from Big Data and use this system to construct a comprehensive knowledge base (KB), called GrandBase.

Design/methodology/approach

In particular, this study extracts new predicates from four types of data sources, namely, Web texts, Document Object Model (DOM) trees, existing KBs and query stream to augment the ontology of the existing KB (i.e. Freebase). In addition, a graph-based approach to conduct better truth discovery for multi-valued predicates is also proposed.

Findings

Empirical studies demonstrate the effectiveness of the approaches presented in this study and the potential of GrandBase. The future research directions regarding GrandBase construction and extension has also been discussed.

Originality/value

To revolutionize our modern society by using the wisdom of Big Data, considerable KBs have been constructed to feed the massive knowledge-driven applications with Resource Description Framework triples. The important challenges for KB construction include extracting information from large-scale, possibly conflicting and different-structured data sources (i.e. the knowledge extraction problem) and reconciling the conflicts that reside in the sources (i.e. the truth discovery problem). Tremendous research efforts have been contributed on both problems. However, the existing KBs are far from being comprehensive and accurate: first, existing knowledge extraction systems retrieve data from limited types of Web sources; second, existing truth discovery approaches commonly assume each predicate has only one true value. In this paper, the focus is on the problem of generating actionable knowledge from Big Data. A system is proposed, which consists of two phases, namely, knowledge extraction and truth discovery, to construct a broader KB, called GrandBase.

Details

PSU Research Review, vol. 1 no. 2
Type: Research Article
ISSN: 2399-1747

Keywords

Open Access
Article
Publication date: 19 November 2021

Cass Shum, Jaimi Garlington, Ankita Ghosh and Seyhmus Baloglu

This study aims to describe the development of hospitality research in terms of research methods and data sources used in the 2010s.

2150

Abstract

Purpose

This study aims to describe the development of hospitality research in terms of research methods and data sources used in the 2010s.

Design/methodology/approach

Content analyses of the research methods and data sources used in original hospitality research published in the 2010s in the Cornell Hospitality Quarterly (CQ), International Journal of Hospitality Management (IJHM), International Journal of Contemporary Hospitality Management (IJCHM), Journal of Hospitality and Tourism Research (JHTR) and International Hospitality Review (IHR) were conducted. It describes whether the time span, functional areas and geographic regions of data sources were related to the research methods and data sources.

Findings

Results from 2,759 original hospitality empirical articles showed that marketing research used various research methods and data sources. Most finance articles used archival data, while most human resources articles used survey designs with organizational data. In addition, only a small amount of research used data from Oceania, Africa and Latin America.

Research limitations/implications

This study sheds some light on the development of hospitality research in terms of research method and data source usage. However, it only focused on five English-based journals from 2010–2019. Therefore, future studies may seek to understand the impact of the COVID-19 pandemic on research methods and data source usage in hospitality research.

Originality/value

This is the first study to examine five hospitality journals' research methods and data sources used in the last decade. It sheds light on the development of hospitality research in the previous decade and identifies new hospitality research avenues.

Details

International Hospitality Review, vol. 37 no. 2
Type: Research Article
ISSN: 2516-8142

Keywords

Open Access
Article
Publication date: 16 November 2022

Laura Korkeamäki, Heikki Keskustalo and Sanna Kumpulainen

The purpose of this paper is to examine what types of task information media scholars need while gathering research data to create new knowledge.

Abstract

Purpose

The purpose of this paper is to examine what types of task information media scholars need while gathering research data to create new knowledge.

Design/methodology/approach

The research design is qualitative and user-oriented. A total of 25 media scholars were interviewed about their research processes and interactions with their research data. The interviews were semi-structured, complemented by critical incident interviews. The analysis focused on the activity of gathering research data. A typology of information (task, domain and task-solving information) guided the analysis of information types related to data gathering, with further analysis focusing only on task information types.

Findings

Media scholars needed the following task information types while gathering research data to create new knowledge: (1) information about research data (aboutness of data, characteristics of data, metadata and secondary information about data), (2) information about sources of research data (characteristics of sources, local media landscapes) and (3) information about cases and their contexts (case information, contextual information). All the task information types should be considered when building data services and tools to support media scholars' work.

Originality/value

The paper increases understanding of the concept of task information in the context of gathering research data to create new knowledge and thereby informs the providers of research data services about the task information types that researchers need.

Details

Journal of Documentation, vol. 78 no. 7
Type: Research Article
ISSN: 0022-0418

Keywords

Open Access
Article
Publication date: 20 July 2020

Abdelghani Bakhtouchi

With the progress of new technologies of information and communication, more and more producers of data exist. On the other hand, the web forms a huge support of all these kinds…

1853

Abstract

With the progress of new technologies of information and communication, more and more producers of data exist. On the other hand, the web forms a huge support of all these kinds of data. Unfortunately, existing data is not proper due to the existence of the same information in different sources, as well as erroneous and incomplete data. The aim of data integration systems is to offer to a user a unique interface to query a number of sources. A key challenge of such systems is to deal with conflicting information from the same source or from different sources. We present, in this paper, the resolution of conflict at the instance level into two stages: references reconciliation and data fusion. The reference reconciliation methods seek to decide if two data descriptions are references to the same entity in reality. We define the principles of reconciliation method then we distinguish the methods of reference reconciliation, first on how to use the descriptions of references, then the way to acquire knowledge. We finish this section by discussing some current data reconciliation issues that are the subject of current research. Data fusion in turn, has the objective to merge duplicates into a single representation while resolving conflicts between the data. We define first the conflicts classification, the strategies for dealing with conflicts and the implementing conflict management strategies. We present then, the relational operators and data fusion techniques. Likewise, we finish this section by discussing some current data fusion issues that are the subject of current research.

Details

Applied Computing and Informatics, vol. 18 no. 3/4
Type: Research Article
ISSN: 2634-1964

Keywords

Open Access
Article
Publication date: 15 September 2021

Elina Late and Sanna Kumpulainen

The paper examines academic historians' information interactions with material from digital historical-newspaper collections as the research process unfolds.

3194

Abstract

Purpose

The paper examines academic historians' information interactions with material from digital historical-newspaper collections as the research process unfolds.

Design/methodology/approach

The study employed qualitative analysis from in-depth interviews with Finnish history scholars who use digitised historical newspapers as primary sources for their research. A model for task-based information interaction guided the collection and analysis of data.

Findings

The study revealed numerous information interactions within activities related to task-planning, the search process, selecting and working with the items and synthesis and reporting. The information interactions differ with the activities involved, which call for system support mechanisms specific to each activity type. Various activities feature information search, which is an essential research method for those using digital collections in the compilation and analysis of data. Furthermore, application of quantitative methods and multidisciplinary collaboration may be shaping culture in history research toward convergence with the research culture of the natural sciences.

Originality/value

For sustainable digital humanities infrastructure and digital collections, it is of great importance that system designers understand how the collections are accessed, why and their use in the real-world context. The study enriches understanding of the collections' utilisation and advances a theoretical framework for explicating task-based information interaction.

Details

Journal of Documentation, vol. 78 no. 7
Type: Research Article
ISSN: 0022-0418

Keywords

Open Access
Article
Publication date: 29 June 2020

Paolo Manghi, Claudio Atzori, Michele De Bonis and Alessia Bardi

Several online services offer functionalities to access information from “big research graphs” (e.g. Google Scholar, OpenAIRE, Microsoft Academic Graph), which correlate…

4540

Abstract

Purpose

Several online services offer functionalities to access information from “big research graphs” (e.g. Google Scholar, OpenAIRE, Microsoft Academic Graph), which correlate scholarly/scientific communication entities such as publications, authors, datasets, organizations, projects, funders, etc. Depending on the target users, access can vary from search and browse content to the consumption of statistics for monitoring and provision of feedback. Such graphs are populated over time as aggregations of multiple sources and therefore suffer from major entity-duplication problems. Although deduplication of graphs is a known and actual problem, existing solutions are dedicated to specific scenarios, operate on flat collections, local topology-drive challenges and cannot therefore be re-used in other contexts.

Design/methodology/approach

This work presents GDup, an integrated, scalable, general-purpose system that can be customized to address deduplication over arbitrary large information graphs. The paper presents its high-level architecture, its implementation as a service used within the OpenAIRE infrastructure system and reports numbers of real-case experiments.

Findings

GDup provides the functionalities required to deliver a fully-fledged entity deduplication workflow over a generic input graph. The system offers out-of-the-box Ground Truth management, acquisition of feedback from data curators and algorithms for identifying and merging duplicates, to obtain an output disambiguated graph.

Originality/value

To our knowledge GDup is the only system in the literature that offers an integrated and general-purpose solution for the deduplication graphs, while targeting big data scalability issues. GDup is today one of the key modules of the OpenAIRE infrastructure production system, which monitors Open Science trends on behalf of the European Commission, National funders and institutions.

Details

Data Technologies and Applications, vol. 54 no. 4
Type: Research Article
ISSN: 2514-9288

Keywords

Open Access
Article
Publication date: 22 November 2022

Kedong Yin, Yun Cao, Shiwei Zhou and Xinman Lv

The purposes of this research are to study the theory and method of multi-attribute index system design and establish a set of systematic, standardized, scientific index systems…

Abstract

Purpose

The purposes of this research are to study the theory and method of multi-attribute index system design and establish a set of systematic, standardized, scientific index systems for the design optimization and inspection process. The research may form the basis for a rational, comprehensive evaluation and provide the most effective way of improving the quality of management decision-making. It is of practical significance to improve the rationality and reliability of the index system and provide standardized, scientific reference standards and theoretical guidance for the design and construction of the index system.

Design/methodology/approach

Using modern methods such as complex networks and machine learning, a system for the quality diagnosis of index data and the classification and stratification of index systems is designed. This guarantees the quality of the index data, realizes the scientific classification and stratification of the index system, reduces the subjectivity and randomness of the design of the index system, enhances its objectivity and rationality and lays a solid foundation for the optimal design of the index system.

Findings

Based on the ideas of statistics, system theory, machine learning and data mining, the focus in the present research is on “data quality diagnosis” and “index classification and stratification” and clarifying the classification standards and data quality characteristics of index data; a data-quality diagnosis system of “data review – data cleaning – data conversion – data inspection” is established. Using a decision tree, explanatory structural model, cluster analysis, K-means clustering and other methods, classification and hierarchical method system of indicators is designed to reduce the redundancy of indicator data and improve the quality of the data used. Finally, the scientific and standardized classification and hierarchical design of the index system can be realized.

Originality/value

The innovative contributions and research value of the paper are reflected in three aspects. First, a method system for index data quality diagnosis is designed, and multi-source data fusion technology is adopted to ensure the quality of multi-source, heterogeneous and mixed-frequency data of the index system. The second is to design a systematic quality-inspection process for missing data based on the systematic thinking of the whole and the individual. Aiming at the accuracy, reliability, and feasibility of the patched data, a quality-inspection method of patched data based on inversion thought and a unified representation method of data fusion based on a tensor model are proposed. The third is to use the modern method of unsupervised learning to classify and stratify the index system, which reduces the subjectivity and randomness of the design of the index system and enhances its objectivity and rationality.

Details

Marine Economics and Management, vol. 5 no. 2
Type: Research Article
ISSN: 2516-158X

Keywords

Open Access
Article
Publication date: 3 June 2019

Lisa Maria Perkhofer, Peter Hofer, Conny Walchshofer, Thomas Plank and Hans-Christian Jetter

Big Data introduces high amounts and new forms of structured, unstructured and semi-structured data into the field of accounting and this requires alternative data management and…

11879

Abstract

Purpose

Big Data introduces high amounts and new forms of structured, unstructured and semi-structured data into the field of accounting and this requires alternative data management and reporting methods. Generating insights from these new data sources highlight the need for different and interactive forms of visualization in the field of visual analytics. Nonetheless, a considerable gap between the recommendations in research and the current usage in practice is evident. In order to understand and overcome this gap, a detailed analysis of the status quo as well as the identification of potential barriers for adoption is vital. The paper aims to discuss this issue.

Design/methodology/approach

A survey with 145 business accountants from Austrian companies from a wide array of business sectors and all hierarchy levels has been conducted. The survey is targeted toward the purpose of this study: identifying barriers, clustered as human-related and technological-related, as well as investigating current practice with respect to interactive visualization use for Big Data.

Findings

The lack of knowledge and experience regarding new visualization types and interaction techniques and the sole focus on Microsoft Excel as a visualization tool can be identified as the main barriers, while the use of multiple data sources and the gradual implementation of further software tools determine the first drivers of adoption.

Research limitations/implications

Due to the data collection with a standardized survey, there was no possibility of dealing with participants individually, which could lead to a misinterpretation of the given answers. Further, the sample population is Austrian, which might cause issues in terms of generalizing results to other geographical or cultural heritages.

Practical implications

The study shows that those knowledgeable and familiar with interactive Big Data visualizations indicate high perceived ease of use. It is, therefore, necessary to offer sufficient training as well as user-centered visualizations and technological support to further increase usage within the accounting profession.

Originality/value

A lot of research has been dedicated to the introduction of novel forms of interactive visualizations. However, little focus has been laid on the impact of these new tools for Big Data from a practitioner’s perspective and their needs.

Details

Journal of Applied Accounting Research, vol. 20 no. 4
Type: Research Article
ISSN: 0967-5426

Keywords

Open Access
Article
Publication date: 1 October 2019

Alex Zarifis, Christopher P. Holland and Alistair Milne

The increasing capabilities of artificial intelligence (AI) are changing the way organizations operate and interact with users both internally and externally. The insurance sector…

2016

Abstract

The increasing capabilities of artificial intelligence (AI) are changing the way organizations operate and interact with users both internally and externally. The insurance sector is currently using AI in several ways but its potential to disrupt insurance is not clear. This research evaluated the implementation of AI-led automation in 20 insurance companies. The findings indicate four business models (BM) emerging: In the first model the insurer takes a smaller part of the value chain allowing others with superior AI and data to take a larger part. In the second model the insurer keeps the same model and value chain but uses AI to improve effectiveness. In the third model the insurer adapts their model to fully utilize AI and seek new sources of data and customers. Lastly in the fourth model a technology focused company uses their existing AI prowess, superior data and extensive customer base, and adds insurance provision.

Details

Emerald Open Research, vol. 1 no. 1
Type: Research Article
ISSN: 2631-3952

Keywords

Open Access
Article
Publication date: 28 September 2017

Kalliopi Platanou, Kristiina Mäkelä, Anton Beletskiy and Anatoli Colicev

The purpose of this paper is to propose new directions for human resource management (HRM) research by drawing attention to online data as a complementary data source to…

5584

Abstract

Purpose

The purpose of this paper is to propose new directions for human resource management (HRM) research by drawing attention to online data as a complementary data source to traditional quantitative and qualitative data, and introducing network text analysis as a method for large quantities of textual material.

Design/methodology/approach

The paper first presents the added value and potential challenges of utilising online data in HRM research, and then proposes a four-step process for analysing online data with network text analysis.

Findings

Online data represent a naturally occuring source of real-time behavioural data that do not suffer from researcher intervention or hindsight bias. The authors argue that as such, this type of data provides a promising yet currently largely untapped empirical context for HRM research that is particularly suited for examining discourses and behavioural and social patterns over time.

Practical implications

While online data hold promise for many novel research questions, it is less appropriate for research questions that seek to establish causality between variables. When using online data, particular attention must be paid to ethical considerations, as well as the validity and representativeness of the sample.

Originality/value

The authors introduce online data and network text analysis as a new avenue for HRM research, with potential to address novel research questions at micro-, meso- and macro-levels of analysis.

Details

Journal of Organizational Effectiveness: People and Performance, vol. 5 no. 1
Type: Research Article
ISSN: 2051-6614

Keywords

1 – 10 of over 10000