Search results
1 – 10 of 418Despite ongoing research into archival metadata standards, digital archives are unable to effectively represent records in their appropriate contexts. This study aims to propose a…
Abstract
Purpose
Despite ongoing research into archival metadata standards, digital archives are unable to effectively represent records in their appropriate contexts. This study aims to propose a knowledge graph that depicts the diverse relationships between heterogeneous digital archive entities.
Design/methodology/approach
This study introduces and describes a method for applying knowledge graphs to digital archives in a step-by-step manner. It examines archival metadata standards, such as Records in Context Ontology (RiC-O), for characterising digital records; explains the process of data refinement, enrichment and reconciliation with examples; and demonstrates the use of knowledge graphs constructed using semantic queries.
Findings
This study introduced the 97imf.kr archive as a knowledge graph, enabling meaningful exploration of relationships within the archive’s records. This approach facilitated comprehensive record descriptions about different record entities. Applying archival ontologies with general-purpose vocabularies to digital records was advised to enhance metadata coherence and semantic search.
Originality/value
Most digital archives serviced in Korea are limited in the proper use of archival metadata standards. The contribution of this study is to propose a practical application of knowledge graph technology for linking and exploring digital records. This study details the process of collecting raw data on archives, data preprocessing and data enrichment, and demonstrates how to build a knowledge graph connected to external data. In particular, the knowledge graph of RiC-O vocabulary, Wikidata and Schema.org vocabulary and the semantic query using it can be applied to supplement keyword search in conventional digital archives.
Details
Keywords
Sofia Baroncini, Bruno Sartini, Marieke Van Erp, Francesca Tomasi and Aldo Gangemi
In the last few years, the size of Linked Open Data (LOD) describing artworks, in general or domain-specific Knowledge Graphs (KGs), is gradually increasing. This provides…
Abstract
Purpose
In the last few years, the size of Linked Open Data (LOD) describing artworks, in general or domain-specific Knowledge Graphs (KGs), is gradually increasing. This provides (art-)historians and Cultural Heritage professionals with a wealth of information to explore. Specifically, structured data about iconographical and iconological (icon) aspects, i.e. information about the subjects, concepts and meanings of artworks, are extremely valuable for the state-of-the-art of computational tools, e.g. content recognition through computer vision. Nevertheless, a data quality evaluation for art domains, fundamental for data reuse, is still missing. The purpose of this study is filling this gap with an overview of art-historical data quality in current KGs with a focus on the icon aspects.
Design/methodology/approach
This study’s analyses are based on established KG evaluation methodologies, adapted to the domain by addressing requirements from art historians’ theories. The authors first select several KGs according to Semantic Web principles. Then, the authors evaluate (1) their structures’ suitability to describe icon information through quantitative and qualitative assessment and (2) their content, qualitatively assessed in terms of correctness and completeness.
Findings
This study’s results reveal several issues on the current expression of icon information in KGs. The content evaluation shows that these domain-specific statements are generally correct but often not complete. The incompleteness is confirmed by the structure evaluation, which highlights the unsuitability of the KG schemas to describe icon information with the required granularity.
Originality/value
The main contribution of this work is an overview of the actual landscape of the icon information expressed in LOD. Therefore, it is valuable to cultural institutions by providing them a first domain-specific data quality evaluation. Since this study’s results suggest that the selected domain information is underrepresented in Semantic Web datasets, the authors highlight the need for the creation and fostering of such information to provide a more thorough art-historical dimension to LOD.
Details
Keywords
Elaheh Hosseini, Kimiya Taghizadeh Milani and Mohammad Shaker Sabetnasab
This research aimed to visualize and analyze the co-word network and thematic clusters of the intellectual structure in the field of linked data during 1900–2021.
Abstract
Purpose
This research aimed to visualize and analyze the co-word network and thematic clusters of the intellectual structure in the field of linked data during 1900–2021.
Design/methodology/approach
This applied research employed a descriptive and analytical method, scientometric indicators, co-word techniques, and social network analysis. VOSviewer, SPSS, Python programming, and UCINet software were used for data analysis and network structure visualization.
Findings
The top ranks of the Web of Science (WOS) subject categorization belonged to various fields of computer science. Besides, the USA was the most prolific country. The keyword ontology had the highest frequency of co-occurrence. Ontology and semantic were the most frequent co-word pairs. In terms of the network structure, nine major topic clusters were identified based on co-occurrence, and 29 thematic clusters were identified based on hierarchical clustering. Comparisons between the two clustering techniques indicated that three clusters, namely semantic bioinformatics, knowledge representation, and semantic tools were in common. The most mature and mainstream thematic clusters were natural language processing techniques to boost modeling and visualization, context-aware knowledge discovery, probabilistic latent semantic analysis (PLSA), semantic tools, latent semantic indexing, web ontology language (OWL) syntax, and ontology-based deep learning.
Originality/value
This study adopted various techniques such as co-word analysis, social network analysis network structure visualization, and hierarchical clustering to represent a suitable, visual, methodical, and comprehensive perspective into linked data.
Details
Keywords
Jiaxin Ye, Huixiang Xiong, Jinpeng Guo and Xuan Meng
The purpose of this study is to investigate how book group recommendations can be used as a meaningful way to suggest suitable books to users, given the increasing number of…
Abstract
Purpose
The purpose of this study is to investigate how book group recommendations can be used as a meaningful way to suggest suitable books to users, given the increasing number of individuals engaging in sharing and discussing books on the web.
Design/methodology/approach
The authors propose reviews fine-grained classification (CFGC) and its related models such as CFGC1 for book group recommendation. These models can categorize reviews successively by function and role. Constructing the BERT-BiLSTM model to classify the reviews by function. The frequency characteristics of the reviews are mined by word frequency analysis, and the relationship between reviews and total book score is mined by correlation analysis. Then, the reviews are classified into three roles: celebrity, general and passerby. Finally, the authors can form user groups, mine group features and combine group features with book fine-grained ratings to make book group recommendations.
Findings
Overall, the best recommendations are made by Synopsis comments, with the accuracy, recall, F-value and Hellinger distance of 52.9%, 60.0%, 56.3% and 0.163, respectively. The F1 index of the recommendations based on the author and the writing comments is improved by 2.5% and 0.4%, respectively, compared to the Synopsis comments.
Originality/value
Previous studies on book recommendation often recommend relevant books for users by mining the similarity between books, so the set of book recommendations recommended to users, especially to groups, always focuses on the few types. The proposed method can effectively ensure the diversity of recommendations by mining users’ tendency to different review attributes of books and recommending books for the groups. In addition, this study also investigates which types of reviews should be used to make book recommendations when targeting groups with specific tendencies.
Details
Keywords
Ruan Wang, Jun Deng, Xinhui Guan and Yuming He
With the development of data mining technology, diverse and broader domain knowledge can be extracted automatically. However, the research on applying knowledge mapping and data…
Abstract
Purpose
With the development of data mining technology, diverse and broader domain knowledge can be extracted automatically. However, the research on applying knowledge mapping and data visualization techniques to genealogical data is limited. This paper aims to fill this research gap by providing a systematic framework and process guidance for practitioners seeking to uncover hidden knowledge from genealogy.
Design/methodology/approach
Based on a literature review of genealogy's current knowledge reasoning research, the authors constructed an integrated framework for knowledge inference and visualization application using a knowledge graph. Additionally, the authors applied this framework in a case study using “Manchu Clan Genealogy” as the data source.
Findings
The case study shows that the proposed framework can effectively decompose and reconstruct genealogy. It demonstrates the reasoning, discovery, and web visualization application process of implicit information in genealogy. It enhances the effective utilization of Manchu genealogy resources by highlighting the intricate relationships among people, places, and time entities.
Originality/value
This study proposed a framework for genealogy knowledge reasoning and visual analysis utilizing a knowledge graph, including five dimensions: the target layer, the resource layer, the data layer, the inference layer, and the application layer. It helps to gather the scattered genealogy information and establish a data network with semantic correlations while establishing reasoning rules to enable inference discovery and visualization of hidden relationships.
Details
Keywords
José Félix Yagüe, Ignacio Huitzil, Carlos Bobed and Fernando Bobillo
There is an increasing interest in the use of knowledge graphs to represent real-world knowledge and a common need to manage imprecise knowledge in many real-world applications…
Abstract
Purpose
There is an increasing interest in the use of knowledge graphs to represent real-world knowledge and a common need to manage imprecise knowledge in many real-world applications. This paper aims to study approaches to solve flexible queries over knowledge graphs.
Design/methodology/approach
By introducing fuzzy logic in the query answering process, the authors are able to obtain a novel algorithm to solve flexible queries over knowledge graphs. This approach is implemented in the FUzzy Knowledge Graphs system, a software tool with an intuitive user-graphical interface.
Findings
This approach makes it possible to reuse semantic web standards (RDF, SPARQL and OWL 2) and builds a fuzzy layer on top of them. The application to a use case shows that the system can aggregate information in different ways by selecting different fusion operators and adapting to different user needs.
Originality/value
This approach is more general than similar previous works in the literature and provides a specific way to represent the flexible restrictions (using fuzzy OWL 2 datatypes).
Details
Keywords
Ying Gao, Qiang Zhang, Xiaoran Wang, Yanmei Huang, Fanshuang Meng and Wan Tao
Currently, the Tang tomb mural cultural relic resources are presented in a multi-source and heterogeneous manner, with a lack of effective organization and sharing between…
Abstract
Purpose
Currently, the Tang tomb mural cultural relic resources are presented in a multi-source and heterogeneous manner, with a lack of effective organization and sharing between resources. Therefore, this study aims to propose a multidimensional knowledge discovery solution for Tang tomb mural cultural relic resources.
Design/methodology/approach
Taking the Tang tomb murals collected by the Shaanxi History Museum as an example, based on clarifying the relevant concepts of Tang tomb mural resources and considering both dynamic and static dimensions, a top-down approach was adopted to first construct an ontology model of Tang tomb mural type cultural relics resources. Then, the actual case data was imported into the Neo4J graph database according to the defined pattern hierarchy to complete the static organization of knowledge, and presented in a multimodal form in knowledge reasoning and retrieval. In addition, geographic information system (GIS) technology is used to dynamically display the spatiotemporal distribution of Tang tomb mural resources, and the distribution trend is analysed from a digital humanistic perspective.
Findings
The multi-dimensional knowledge discovery of Tang tomb mural cultural relics resources can help establish the correlation and spatiotemporal relationship between resources, providing support for semantic retrieval and navigation, knowledge discovery and visualization and so on.
Originality/value
This study takes the murals in the collection of the Shaanxi History Museum as an example, revealing potential knowledge associations in a static and intelligent way, achieving knowledge discovery and management of Tang tomb murals, and dynamically presents the spatial distribution of Tang tomb murals through GIS technology, meeting the knowledge presentation needs of different users and opening up new ideas for the study of Tang tomb murals.
Details
Keywords
In this study, the distinctive functional features of linked data (LD) catalogues were investigated to contrast with existing online public access catalogues (OPACs) and discovery…
Abstract
Purpose
In this study, the distinctive functional features of linked data (LD) catalogues were investigated to contrast with existing online public access catalogues (OPACs) and discovery systems using a checklist approach. The checklist was derived from a literature review and is composed of 10 items as follows: self-descriptive and dynamic content for resource description, linkage to external LD sources and online services, aggregation of knowledge contexts into knowledge graphs (KGs), URI-based link discovery, representation and query of LD relationships, URI-based serendipitous discovery, keyword recommendation, faceted limitation and browsing, visualization and openness of data.
Design/methodology/approach
Ten functional features derived from the literature were checked against existing LD catalogues offered by libraries, archives and museums (LAMs). The LD catalogues were regarded as qualified subjects if they offered functional features that were distinct from current OPACs and discovery systems through URI-based enrichment and aggregation from various LD sources. In addition to individual organizations, LD union catalogues were also included. However, LD hubs, such as ISNI, OCLC WorldCat Entities, VIAF and Wikidata, were excluded. In total, six LD catalogues from LAMs were selected as subjects for examination.
Findings
First, LD catalogues provide similar KG information through URI combination, and KGs also facilitate information serendipity, including social-document, intellectual, conceptual, spatial and temporal contexts and networks of corporate bodies, persons and families (CPFs). Second, LD catalogues have transformed the “seek first and browse later” paradigm into a “seek or browse” paradigm by refreshing the browsing function of traditional card catalogues with preview and new options to facilitate LD identification and discovery. Third, LD catalogues have refined keyword recommendation with the addition of the following fields: person’s title, CPF relationships, entity type and LD source. Lastly, a virtual union LD catalogue is offered.
Research limitations/implications
The proposed checklist revealed the unique/improved functional features of LD catalogues, allowing further investigation and comparison. More cases from the fields of medicine, engineering science and so on will be required to make revisions to fine-tune the proposed checklist approach.
Originality/value
To the best of the author’s knowledge, this is the first study to propose a checklist of functional features for LD catalogues and examine what the results and features of LD catalogues have achieved and are supported by from ontologies across LAMs. The findings suggest that LD provides a viable alternative to catalogues. The proposed checklist and results pave the way for the future development of LD catalogues and next-generation catalogues and also provide a basis for the future study of LD catalogues from other fields to refine the proposed checklist.
Details
Keywords
Huyen Nguyen, Haihua Chen, Jiangping Chen, Kate Kargozari and Junhua Ding
This study aims to evaluate a method of building a biomedical knowledge graph (KG).
Abstract
Purpose
This study aims to evaluate a method of building a biomedical knowledge graph (KG).
Design/methodology/approach
This research first constructs a COVID-19 KG on the COVID-19 Open Research Data Set, covering information over six categories (i.e. disease, drug, gene, species, therapy and symptom). The construction used open-source tools to extract entities, relations and triples. Then, the COVID-19 KG is evaluated on three data-quality dimensions: correctness, relatedness and comprehensiveness, using a semiautomatic approach. Finally, this study assesses the application of the KG by building a question answering (Q&A) system. Five queries regarding COVID-19 genomes, symptoms, transmissions and therapeutics were submitted to the system and the results were analyzed.
Findings
With current extraction tools, the quality of the KG is moderate and difficult to improve, unless more efforts are made to improve the tools for entity extraction, relation extraction and others. This study finds that comprehensiveness and relatedness positively correlate with the data size. Furthermore, the results indicate the performances of the Q&A systems built on the larger-scale KGs are better than the smaller ones for most queries, proving the importance of relatedness and comprehensiveness to ensure the usefulness of the KG.
Originality/value
The KG construction process, data-quality-based and application-based evaluations discussed in this paper provide valuable references for KG researchers and practitioners to build high-quality domain-specific knowledge discovery systems.
Details
Keywords
Jun Liu, Junyuan Dong, Mingming Hu and Xu Lu
Existing Simultaneous Localization and Mapping (SLAM) algorithms have been relatively well developed. However, when in complex dynamic environments, the movement of the dynamic…
Abstract
Purpose
Existing Simultaneous Localization and Mapping (SLAM) algorithms have been relatively well developed. However, when in complex dynamic environments, the movement of the dynamic points on the dynamic objects in the image in the mapping can have an impact on the observation of the system, and thus there will be biases and errors in the position estimation and the creation of map points. The aim of this paper is to achieve more accurate accuracy in SLAM algorithms compared to traditional methods through semantic approaches.
Design/methodology/approach
In this paper, the semantic segmentation of dynamic objects is realized based on U-Net semantic segmentation network, followed by motion consistency detection through motion detection method to determine whether the segmented objects are moving in the current scene or not, and combined with the motion compensation method to eliminate dynamic points and compensate for the current local image, so as to make the system robust.
Findings
Experiments comparing the effect of detecting dynamic points and removing outliers are conducted on a dynamic data set of Technische Universität München, and the results show that the absolute trajectory accuracy of this paper's method is significantly improved compared with ORB-SLAM3 and DS-SLAM.
Originality/value
In this paper, in the semantic segmentation network part, the segmentation mask is combined with the method of dynamic point detection, elimination and compensation, which reduces the influence of dynamic objects, thus effectively improving the accuracy of localization in dynamic environments.
Details