Search results
1 – 10 of 487Kimmo Kettunen, Heikki Keskustalo, Sanna Kumpulainen, Tuula Pääkkönen and Juha Rautiainen
This study aims to identify user perception of different qualities of optical character recognition (OCR) in texts. The purpose of this paper is to study the effect of different…
Abstract
Purpose
This study aims to identify user perception of different qualities of optical character recognition (OCR) in texts. The purpose of this paper is to study the effect of different quality OCR on users' subjective perception through an interactive information retrieval task with a collection of one digitized historical Finnish newspaper.
Design/methodology/approach
This study is based on the simulated work task model used in interactive information retrieval. Thirty-two users made searches to an article collection of Finnish newspaper Uusi Suometar 1869–1918 which consists of ca. 1.45 million autosegmented articles. The article search database had two versions of each article with different quality OCR. Each user performed six pre-formulated and six self-formulated short queries and evaluated subjectively the top 10 results using a graded relevance scale of 0–3. Users were not informed about the OCR quality differences of the otherwise identical articles.
Findings
The main result of the study is that improved OCR quality affects subjective user perception of historical newspaper articles positively: higher relevance scores are given to better-quality texts.
Originality/value
To the best of the authors’ knowledge, this simulated interactive work task experiment is the first one showing empirically that users' subjective relevance assessments are affected by a change in the quality of an optically read text.
Details
Keywords
Aya Khaled Youssef Sayed Mohamed, Dagmar Auer, Daniel Hofer and Josef Küng
Data protection requirements heavily increased due to the rising awareness of data security, legal requirements and technological developments. Today, NoSQL databases are…
Abstract
Purpose
Data protection requirements heavily increased due to the rising awareness of data security, legal requirements and technological developments. Today, NoSQL databases are increasingly used in security-critical domains. Current survey works on databases and data security only consider authorization and access control in a very general way and do not regard most of today’s sophisticated requirements. Accordingly, the purpose of this paper is to discuss authorization and access control for relational and NoSQL database models in detail with respect to requirements and current state of the art.
Design/methodology/approach
This paper follows a systematic literature review approach to study authorization and access control for different database models. Starting with a research on survey works on authorization and access control in databases, the study continues with the identification and definition of advanced authorization and access control requirements, which are generally applicable to any database model. This paper then discusses and compares current database models based on these requirements.
Findings
As no survey works consider requirements for authorization and access control in different database models so far, the authors define their requirements. Furthermore, the authors discuss the current state of the art for the relational, key-value, column-oriented, document-based and graph database models in comparison to the defined requirements.
Originality/value
This paper focuses on authorization and access control for various database models, not concrete products. This paper identifies today’s sophisticated – yet general – requirements from the literature and compares them with research results and access control features of current products for the relational and NoSQL database models.
Details
Keywords
Simona Ibba, Filippo Eros Pani, John Gregory Stockton, Giulio Barabino, Michele Marchesi and Danilo Tigano
One of the main tasks of a researcher is to properly communicate the results he obtained. The choice of the journal in which to publish the work is therefore very important…
Abstract
Purpose
One of the main tasks of a researcher is to properly communicate the results he obtained. The choice of the journal in which to publish the work is therefore very important. However, not all journals have suitable characteristics for a correct dissemination of scientific knowledge. Some publishers turn out to be unreliable and, against a payment, they publish whatever researchers propose. The authors call “predatory journals” these untrustworthy journals. The purpose of this paper is to analyse the incidence of predatory journals in computer science literature and present a tool that was developed for this purpose.
Design/methodology/approach
The authors focused their attention on editors, universities and publishers that are involved in this kind of publishing process. The starting point of their research is the list of scholarly open-access publishers and open-access stand-alone journals created by Jeffrey Beall. Specifically, they analysed the presence of predatory journals in the search results obtained from Google Scholar in the engineering and computer science fields. They also studied the change over time of such incidence in the articles published between 2011 and 2015.
Findings
The analysis shows that the phenomenon of predatory journals somehow decreased in 2015, probably due to a greater awareness of the risks related to the reputation of the authors.
Originality/value
We focused on computer science field, using a specific sample of queries. We developed a software to automatically make queries to the search engine, and to detect predatory journals, using Beall’s list.
Details
Keywords
Maria Giovanna Confetto and Claudia Covucci
For companies that intend to respond to the modern conscious consumers' needs, a great competitive advantage is played on the ability to incorporate sustainability messages in…
Abstract
Purpose
For companies that intend to respond to the modern conscious consumers' needs, a great competitive advantage is played on the ability to incorporate sustainability messages in marketing communications. The aim of this paper is to address this important priority in the web context, building a semantic algorithm that allows content managers to evaluate the quality of sustainability web contents for search engines, considering the current semantic web development.
Design/methodology/approach
Following the Design Science (DS) methodological approach, the study develops the algorithm as an artefact capable of solving a practical problem and improving the operation of content managerial process.
Findings
The algorithm considers multiple factors of evaluation, grouped in three parameters: completeness, clarity and consistency. An applicability test of the algorithm was conducted on a sample of web pages of the Google blog on sustainability to highlight the correspondence between the established evaluation factors and those actually used by Google.
Practical implications
Studying content marketing for sustainability communication constitutes a new field of research that offers exciting opportunities. Writing sustainability contents in an effective way is a fundamental step to trigger stakeholder engagement mechanisms online. It could be a positive social engineering technique in the hands of marketers to make web users able to pursue sustainable development in their choices.
Originality/value
This is the first study that creates a theoretical connection between digital content marketing and sustainability communication focussing, especially, on the aspects of search engine optimization (SEO). The algorithm of “Sustainability-contents SEO” is the first operational software tool, with a regulatory nature, that is able to analyse the web contents, detecting the terms of the sustainability language and measuring the compliance to SEO requirements.
Details
Keywords
Tsuneo Yamada and Yosuke Morimoto
In order to sustain the development and utilization of learning content under the limited financial and human resources, the sharing and reuse of open educational resources (OERs…
Abstract
In order to sustain the development and utilization of learning content under the limited financial and human resources, the sharing and reuse of open educational resources (OERs) have been promoted in various school levels and disciplines. In many countries, central organization(s) collect both learning content and its metadata, and provide them to learners and teachers with distinctive value‐added services from their portals. Collaborations among the national portals have also emerged beyond borders.Global Learning Object Brokered Exchange (GLOBE) is an international consortium of the hub organizations, which manage a federated repository and/or a meta‐referatory in each country and region. GLOBE was established in September 2004 by five founding members, and now twelve organizations in the world participate in. By adopting IEEE LOM version 1.0 for the metadata standards, SQI for the query language, and OAI‐PMH for harvesting, GLOBE realized a global search and delivery infrastructure for lifelong learning (LLL). At present, while the repositories all over the world store huge number of learning content and metadata, the variance of the quality becomes much bigger. In addition, because the effectiveness of learning content is relative and dependent on contextual factors, both optimizations specialized in learner characteristics and localizations to each language/culture are indispensable. The users need some supports on the quality and pedagogical guidance when they find their right content. GLOBE, based on multilingualism, multiculturalism and pluralism, collaborates to build up standardized services on the quality assurance of the content utilizing the “Educational” items of LOM. In AAOU framework also, the organizations can promote “openness”, and share the quality learning content and educational information by using a standardized content search and delivery infrastructure.
Details
Keywords
Zhengfa Yang, Qian Liu, Baowen Sun and Xin Zhao
This paper aims to make it convenient for those who have only just begun their research into Community Question Answering (CQA) expert recommendation, and for those who are…
Abstract
Purpose
This paper aims to make it convenient for those who have only just begun their research into Community Question Answering (CQA) expert recommendation, and for those who are already concerned with this issue, to ease the extension of our understanding with future research.
Design/methodology/approach
In this paper, keywords such as “CQA”, “Social Question Answering”, “expert recommendation”, “question routing” and “expert finding” are used to search major digital libraries. The final sample includes a list of 83 relevant articles authored in academia as well as industry that have been published from January 1, 2008 to March 1, 2019.
Findings
This study proposes a comprehensive framework to categorize extant studies into three broad areas of CQA expert recommendation research: understanding profile modeling, recommendation approaches and recommendation system impacts.
Originality/value
This paper focuses on discussing and sorting out the key research issues from these three research genres. Finally, it was found that conflicting and contradictory research results and research gaps in the existing research, and then put forward the urgent research topics.
Details
Keywords
The subject of this paper is the idea of Brain–Computer Interface (BCI). The main goal is to assess the potential impact of BCI on the design, use and evaluation of information…
Abstract
Purpose
The subject of this paper is the idea of Brain–Computer Interface (BCI). The main goal is to assess the potential impact of BCI on the design, use and evaluation of information retrieval systems operating in libraries.
Design/methodology/approach
The method of literature review was used to establish the state of research. The search according to accepted queries was carried out in the Scopus database and complementary in Google Scholar. To determine the state of research on BCI on the basis of library and information science, a specialist LISTA abstract database was also searched. The most current papers published in the years 2015–2019 in the English language or having at least an abstract in this language were taken into account.
Findings
The analysis showed that BCI issues are extremely popular in subject literature from various fields, mainly computer science, but practically does not occur in the context of using this technology in information retrieval systems.
Research limitations/implications
Due to the fact that BCI solutions are not yet implemented in libraries and are rarely the subject of scientific considerations in the field of library and information science, this article is mainly based on literature from other disciplines. The goal was to consider how much BCI solutions can affect library information retrieval systems. The considerations presented in this article are theoretical in nature due to the lack of empirical materials on which to base. The author's assumption was to initiate a discussion about BCI on the basis of library and information science, not to propose final solutions.
Practical implications
The results can be widely used in practice as a framework for the implementation of BCI in libraries.
Social implications
The article can help to facilitate the debate on the role of implementing new technologies in libraries.
Originality/value
The problem of BCI is very rarely addressed in the subject literature in the field of library and information science.
Details
Keywords
Abstract
Purpose
Question-answering (QA) systems are being increasingly applied in learning contexts. However, the authors’ understanding of the relationship between such tools and traditional QA channels remains limited. Focusing on question-answering learning activities, the current research investigates the effect of QA systems on students' learning processes and outcomes, as well as the interplay between two QA channels, that is, QA systems and communication with instructors.
Design/methodology/approach
The authors designed and implemented a QA system for two university courses, and collected data from questionnaires and system logs that recorded the interaction between students and the system throughout a semester.
Findings
The results show that using a QA system alone does not improve students' learning processes or outcomes. However, the use of a QA system significantly improves the positive effect of instructor communication.
Originality/value
This study contributes to the literature on learning and education technology, and provides practical guidance on how to incorporate QA tools in learning.
Details
Keywords
Sofia Baroncini, Bruno Sartini, Marieke Van Erp, Francesca Tomasi and Aldo Gangemi
In the last few years, the size of Linked Open Data (LOD) describing artworks, in general or domain-specific Knowledge Graphs (KGs), is gradually increasing. This provides…
Abstract
Purpose
In the last few years, the size of Linked Open Data (LOD) describing artworks, in general or domain-specific Knowledge Graphs (KGs), is gradually increasing. This provides (art-)historians and Cultural Heritage professionals with a wealth of information to explore. Specifically, structured data about iconographical and iconological (icon) aspects, i.e. information about the subjects, concepts and meanings of artworks, are extremely valuable for the state-of-the-art of computational tools, e.g. content recognition through computer vision. Nevertheless, a data quality evaluation for art domains, fundamental for data reuse, is still missing. The purpose of this study is filling this gap with an overview of art-historical data quality in current KGs with a focus on the icon aspects.
Design/methodology/approach
This study’s analyses are based on established KG evaluation methodologies, adapted to the domain by addressing requirements from art historians’ theories. The authors first select several KGs according to Semantic Web principles. Then, the authors evaluate (1) their structures’ suitability to describe icon information through quantitative and qualitative assessment and (2) their content, qualitatively assessed in terms of correctness and completeness.
Findings
This study’s results reveal several issues on the current expression of icon information in KGs. The content evaluation shows that these domain-specific statements are generally correct but often not complete. The incompleteness is confirmed by the structure evaluation, which highlights the unsuitability of the KG schemas to describe icon information with the required granularity.
Originality/value
The main contribution of this work is an overview of the actual landscape of the icon information expressed in LOD. Therefore, it is valuable to cultural institutions by providing them a first domain-specific data quality evaluation. Since this study’s results suggest that the selected domain information is underrepresented in Semantic Web datasets, the authors highlight the need for the creation and fostering of such information to provide a more thorough art-historical dimension to LOD.
Details
Keywords
Taro Aso, Toshiyuki Amagasa and Hiroyuki Kitagawa
The purpose of this paper is to propose a scheme that allows users to interactively explore relations between entities in knowledge bases (KBs). KBs store a wide range of…
Abstract
Purpose
The purpose of this paper is to propose a scheme that allows users to interactively explore relations between entities in knowledge bases (KBs). KBs store a wide range of knowledge about real-world entities in a structured form as (subject, predicate, object). Although it is possible to query entities and relations among entities by specifying appropriate query expressions of SPARQL or keyword queries, the structure and the vocabulary are complicated, and it is hard for non-expert users to get the desired information. For this reason, many researchers have proposed faceted search interfaces for KBs. Nevertheless, existing ones are designed for finding entities and are insufficient for finding relations.
Design/methodology/approach
To this problem, the authors propose a novel “relation facet” to find relations between entities. To generate it, they applied clustering on predicates for grouping those predicates that are connected to common objects. Having generated clusters of predicates, the authors generated a facet according to the result. Specifically, they proposed to use a couple of clustering algorithms, namely, agglomerative hierarchical clustering (AHC) and CANDECOMP/PARAFAC (CP) tensor decomposition which is one of the tensor decomposition methods.
Findings
The authors experimentally show test the performance of clustering methods and found that AHC performs better than tensor decomposition. Besides, the authors conducted a user study and show that their proposed scheme performs better than existing ones in the task of searching relations.
Originality/value
The authors propose a relation-oriented faceted search method for KBs that allows users to explore relations between entities. As far as the authors know, this is the first method to focus on the exploration of relations between entities.
Details