Search results
1 – 10 of 986Yi-Hung Liu, Sheng-Fong Chen and Dan-Wei (Marian) Wen
Online medical repositories provide a platform for users to share information and dynamically access abundant electronic health data. It is important to determine whether case…
Abstract
Purpose
Online medical repositories provide a platform for users to share information and dynamically access abundant electronic health data. It is important to determine whether case report information can assist the general public in appropriately managing their diseases. Therefore, this paper aims to introduce a novel deep learning-based method that allows non-professionals to make inquiries using ordinary vocabulary, retrieving the most relevant case reports for accurate and effective health information.
Design/methodology/approach
The dataset of case reports was collected from both the patient-generated research network and the digital medical journal repository. To enhance the accuracy of obtaining relevant case reports, the authors propose a retrieval approach that combines BERT and BiLSTM methods. The authors identified representative health-related case reports and analyzed the retrieval performance, as well as user judgments.
Findings
This study aims to provide the necessary functionalities to deliver relevant health case reports based on input from ordinary terms. The proposed framework includes features for health management, user feedback acquisition and ranking by weights to obtain the most pertinent case reports.
Originality/value
This study contributes to health information systems by analyzing patients' experiences and treatments with the case report retrieval model. The results of this study can provide immense benefit to the general public who intend to find treatment decisions and experiences from relevant case reports.
Details
Keywords
Kianoosh Rashidi, Hajar Sotudeh and Alireza Nikseresht
This study aimed to investigate how the enrichment of medical documents' index terms by their comments improves the relevance and novelty of the top-ranked results retrieved by an…
Abstract
Purpose
This study aimed to investigate how the enrichment of medical documents' index terms by their comments improves the relevance and novelty of the top-ranked results retrieved by an NLP system.
Design/methodology/approach
A semi-experimental pre-test and post-test research was designed to compare NLP-based indexes before and after being expanded by the comment terms. The experiments were conducted on a test collection of 13,957 documents commented by F1000-Prime reviewers. They were indexed at title, abstract, body and full-text levels. In total, 100 seed documents were randomly selected and served as queries. The textual similarity of the documents and queries was calculated using Lucene-more-like-this function and evaluated by the semantic similarity of their MeSH. The results novelty was measured using maximal marginal relevance and evaluated by their MeSH novelties. Normalized discounted cumulative gain was used to compare the basic and expanded indexes' precisions at 10, 20 and 50 top ranks.
Findings
The relevance and novelty of the results ranked at the top precision points was improved after expanding the indexes by the comment terms. The finding implies that meta-texts are effective in representing their mother documents, by adding dynamic elements to their rather static contents. It also provides further evidence about the merits of the application of social intelligence and collective wisdom reflected in the actions and reactions of users in tackling the challenges faced by NLP-based systems.
Originality/value
This is the first study to confirm that social comments on scientific papers improve the performance of information systems in terms of relevance and novelty.
Peer review
The peer review history for this article is available at: https://publons.com/publon/10.1108/OIR-05-2022-0283.
Details
Keywords
The purpose of the paper is to analyze how the Neo-documentalist movement, initiated in 1996 by Michael Buckland, Boyd Rayward and Niels Lund, has evolved in its 27 years history…
Abstract
Purpose
The purpose of the paper is to analyze how the Neo-documentalist movement, initiated in 1996 by Michael Buckland, Boyd Rayward and Niels Lund, has evolved in its 27 years history, how the choice of documentation as name of the new program in Tromsø has made a difference in the LIS field and how different documentation scholars around the world has participated and approached the movement until now.
Design/methodology/approach
The paper has approached the “Neo-documentalist movement” in a historical perspective from 1996 to 2023 discussing what difference does the choice of a concept make, when the concept of documentation is chosen instead of information in the name of a program and for the general discussion of the object of an academic field like Library and Information Science.
Findings
The analysis shows that it did make a difference to choose the concept of documentation as name of the program in Tromsø and the Neo-documentalist movement contributed to a new focus and discussion of the informative objects, the documents and their creation, not only in Tromsø, but in different parts of the world across linguistic borders.
Originality/value
The paper is original by the fact that it is the first time that the neo documentalist movement has been reviewed on a global scale across linguistic barriers. It has value by a discussion of the ways in which a choice of concept matter in relation to defining a field and the research agenda.
Details
Keywords
Oussama Ayoub, Christophe Rodrigues and Nicolas Travers
This paper aims to manage the word gap in information retrieval (IR) especially for long documents belonging to specific domains. In fact, with the continuous growth of text data…
Abstract
Purpose
This paper aims to manage the word gap in information retrieval (IR) especially for long documents belonging to specific domains. In fact, with the continuous growth of text data that modern IR systems have to manage, existing solutions are needed to efficiently find the best set of documents for a given request. The words used to describe a query can differ from those used in related documents. Despite meaning closeness, nonoverlapping words are challenging for IR systems. This word gap becomes significant for long documents from specific domains.
Design/methodology/approach
To generate new words for a document, a deep learning (DL) masked language model is used to infer related words. Used DL models are pretrained on massive text data and carry common or specific domain knowledge to propose a better document representation.
Findings
The authors evaluate the approach of this study on specific IR domains with long documents to show the genericity of the proposed model and achieve encouraging results.
Originality/value
In this paper, to the best of the authors’ knowledge, an original unsupervised and modular IR system based on recent DL methods is introduced.
Details
Keywords
Wayne de Fremery and Michael Keeble Buckland
The purpose of this paper is to provide a new and useful formulation of relevance.
Abstract
Purpose
The purpose of this paper is to provide a new and useful formulation of relevance.
Design/methodology/approach
This paper is formulated as a conceptual argument. It makes the case for the utility of considering relevance to be function of use in creative processes.
Findings
There are several corollaries to formulating relevance as a function of use. These include the idea that objects by themselves cannot be relevant since use assumes interaction; the affordances of objects and how they are perceived can affect what becomes relevant but are not in themselves relevant; relevance is not an essential characteristic of objects; relevance is transient; potential relevance (what might be relevant in the future) can be distinguished from what is relevant in use and from what has been relevant in the past.
Originality/value
The paper shows that its new formulation of relevance brings improved conceptual and terminological clarity to the discourse about relevance in information science. It demonstrates that how relevance is articulated conceptually is important as its conceptualization can affect the ways that users are able to make use of information systems and, by extension, how information systems can facilitate or disable the co-production of creative outcomes. The paper also usefully expands investigative opportunities by suggesting relevance and creativity are interrelated.
Details
Keywords
Debasis Majhi and Bhaskar Mukherjee
The purpose of this study is to identify the research fronts by analysing highly cited core papers adjusted with the age of a paper in library and information science (LIS) where…
Abstract
Purpose
The purpose of this study is to identify the research fronts by analysing highly cited core papers adjusted with the age of a paper in library and information science (LIS) where natural language processing (NLP) is being applied significantly.
Design/methodology/approach
By excavating international databases, 3,087 core papers that received at least 5% of the total citations have been identified. By calculating the average mean years of these core papers, and total citations received, a CPT (citation/publication/time) value was calculated in all 20 fronts to understand how a front is relatively receiving greater attention among peers within a course of time. One theme article has been finally identified from each of these 20 fronts.
Findings
Bidirectional encoder representations from transformers with CPT value 1.608 followed by sentiment analysis with CPT 1.292 received highest attention in NLP research. Columbia University New York, in terms of University, Journal of the American Medical Informatics Association, in terms of journals, USA followed by People Republic of China, in terms of country and Xu, H., University of Texas, in terms of author are the top in these fronts. It is identified that the NLP applications boost the performance of digital libraries and automated library systems in the digital environment.
Practical implications
Any research fronts that are identified in the findings of this paper may be used as a base for researchers who intended to perform extensive research on NLP.
Originality/value
To the best of the authors’ knowledge, the methodology adopted in this paper is the first of its kind where meta-analysis approach has been used for understanding the research fronts in sub field like NLP for a broad domain like LIS.
Details
Keywords
Kimmo Kettunen, Heikki Keskustalo, Sanna Kumpulainen, Tuula Pääkkönen and Juha Rautiainen
This study aims to identify user perception of different qualities of optical character recognition (OCR) in texts. The purpose of this paper is to study the effect of different…
Abstract
Purpose
This study aims to identify user perception of different qualities of optical character recognition (OCR) in texts. The purpose of this paper is to study the effect of different quality OCR on users' subjective perception through an interactive information retrieval task with a collection of one digitized historical Finnish newspaper.
Design/methodology/approach
This study is based on the simulated work task model used in interactive information retrieval. Thirty-two users made searches to an article collection of Finnish newspaper Uusi Suometar 1869–1918 which consists of ca. 1.45 million autosegmented articles. The article search database had two versions of each article with different quality OCR. Each user performed six pre-formulated and six self-formulated short queries and evaluated subjectively the top 10 results using a graded relevance scale of 0–3. Users were not informed about the OCR quality differences of the otherwise identical articles.
Findings
The main result of the study is that improved OCR quality affects subjective user perception of historical newspaper articles positively: higher relevance scores are given to better-quality texts.
Originality/value
To the best of the authors’ knowledge, this simulated interactive work task experiment is the first one showing empirically that users' subjective relevance assessments are affected by a change in the quality of an optically read text.
Details
Keywords
Gitaek Lee, Seonghyeon Moon and Seokho Chi
Contractors must check the provisions that may cause disputes in the specifications to manage project risks when bidding for a construction project. However, since the…
Abstract
Purpose
Contractors must check the provisions that may cause disputes in the specifications to manage project risks when bidding for a construction project. However, since the specification is mainly written regarding many national standards, determining which standard each section of the specification is derived from and whether the content is appropriate for the local site is a labor-intensive task. To develop an automatic reference section identification model that helps complete the specification review process in short bidding steps, the authors proposed a framework that integrates rules and machine learning algorithms.
Design/methodology/approach
The study begins by collecting 7,795 sections from construction specifications and the national standards from different countries. Then, the collected sections were retrieved for similar section pairs with syntactic rules generated by the construction domain knowledge. Finally, to improve the reliability and expandability of the section paring, the authors built a deep structured semantic model that increases the cosine similarity between documents dealing with the same topic by learning human-labeled similarity information.
Findings
The integrated model developed in this study showed 0.812, 0.898, and 0.923 levels of performance in NDCG@1, NDCG@5, and NDCG@10, respectively, confirming that the model can adequately select document candidates that require comparative analysis of clauses for practitioners.
Originality/value
The results contribute to more efficient and objective identification of potential disputes within the specifications by automatically providing practitioners with the reference section most relevant to the analysis target section.
Details
Keywords
Raj Kumar Bhardwaj, Ritesh Kumar and Mohammad Nazim
This paper evaluates the precision of four metasearch engines (MSEs) – DuckDuckGo, Dogpile, Metacrawler and Startpage, to determine which metasearch engine exhibits the highest…
Abstract
Purpose
This paper evaluates the precision of four metasearch engines (MSEs) – DuckDuckGo, Dogpile, Metacrawler and Startpage, to determine which metasearch engine exhibits the highest level of precision and to identify the metasearch engine that is most likely to return the most relevant search results.
Design/methodology/approach
The research is divided into two parts: the first phase involves four queries categorized into two segments (4-Q-2-S), while the second phase includes six queries divided into three segments (6-Q-3-S). These queries vary in complexity, falling into three types: simple, phrase and complex. The precision, average precision and the presence of duplicates across all the evaluated metasearch engines are determined.
Findings
The study clearly demonstrated that Startpage returned the most relevant results and achieved the highest precision (0.98) among the four MSEs. Conversely, DuckDuckGo exhibited consistent performance across both phases of the study.
Research limitations/implications
The study only evaluated four metasearch engines, which may not be representative of all available metasearch engines. Additionally, a limited number of queries were used, which may not be sufficient to generalize the findings to all types of queries.
Practical implications
The findings of this study can be valuable for accreditation agencies in managing duplicates, improving their search capabilities and obtaining more relevant and precise results. These findings can also assist users in selecting the best metasearch engine based on precision rather than interface.
Originality/value
The study is the first of its kind which evaluates the four metasearch engines. No similar study has been conducted in the past to measure the performance of metasearch engines.
Details
Keywords
Samia Ebrahiem, Ahmed O. El-Kholei and Ghada Yassein
The article attempts to shed light on the social aspects of research that deal with Sustainable Development Goals (SDGs) and sustainable cities. The aim is to offer a global view…
Abstract
Purpose
The article attempts to shed light on the social aspects of research that deal with Sustainable Development Goals (SDGs) and sustainable cities. The aim is to offer a global view of these facets' evolution and to provide information on people-centered smart cities.
Design/methodology/approach
The research is qualitative. A systematic bibliometric approach is a framework for the research. The unit of analysis is publications on SDGs and Smart Cities (SCs) indexed in Scopus. The authors used VOSviewer text mining functionality to construct co-occurrence networks of socially related critical terms extracted from textual data. The co-occurrence of keywords presents a valuable method and process for attaining in-depth analysis and fast comprehension of trends and linkages in articles from a holistic approach.
Findings
Social media, social sustainability and social capital are the three multifaceted social keywords that co-occur in SDGs and SCs. The paper provides a brief compendium of resources and frameworks to build a socially sustainable smart city.
Research limitations/implications
The retrieval date was on 15 August 2019. The authors used the same search query for new papers released in 2019 and afterwards to update their findings. The authors collected 657 documents on SCs, compared to 2,975 documents about SDGs demonstrating that their findings are still trending in the same direction, emphasizing the importance of the research topic. SCs' social aspects are still chartered areas that require the attention to future research.
Originality/value
The authors’ decision to use two separate data sets for SCs and SDGs data files helps to provide a more comprehensive picture of the research landscape. It may identify areas where research is lacking or needs future research. The authors present an integrative agenda for a smart city to be socially sustainable. Innovative approaches to urban planning are required to empower the place and context and improve the users' satisfaction, where innovative solutions enable smart, sustainable and inclusive societies. Infrastructure governance is a critical keystone. It could guarantee that public investments contribute to sustainable urban development while enhancing city resilience, particularly in facing climate change and inclusive growth challenges.
Details