Search results
1 – 10 of 143Elaheh Hosseini, Kimiya Taghizadeh Milani and Mohammad Shaker Sabetnasab
This research aimed to visualize and analyze the co-word network and thematic clusters of the intellectual structure in the field of linked data during 1900–2021.
Abstract
Purpose
This research aimed to visualize and analyze the co-word network and thematic clusters of the intellectual structure in the field of linked data during 1900–2021.
Design/methodology/approach
This applied research employed a descriptive and analytical method, scientometric indicators, co-word techniques, and social network analysis. VOSviewer, SPSS, Python programming, and UCINet software were used for data analysis and network structure visualization.
Findings
The top ranks of the Web of Science (WOS) subject categorization belonged to various fields of computer science. Besides, the USA was the most prolific country. The keyword ontology had the highest frequency of co-occurrence. Ontology and semantic were the most frequent co-word pairs. In terms of the network structure, nine major topic clusters were identified based on co-occurrence, and 29 thematic clusters were identified based on hierarchical clustering. Comparisons between the two clustering techniques indicated that three clusters, namely semantic bioinformatics, knowledge representation, and semantic tools were in common. The most mature and mainstream thematic clusters were natural language processing techniques to boost modeling and visualization, context-aware knowledge discovery, probabilistic latent semantic analysis (PLSA), semantic tools, latent semantic indexing, web ontology language (OWL) syntax, and ontology-based deep learning.
Originality/value
This study adopted various techniques such as co-word analysis, social network analysis network structure visualization, and hierarchical clustering to represent a suitable, visual, methodical, and comprehensive perspective into linked data.
Details
Keywords
Shaodan Sun, Jun Deng and Xugong Qin
This paper aims to amplify the retrieval and utilization of historical newspapers through the application of semantic organization, all from the vantage point of a fine-grained…
Abstract
Purpose
This paper aims to amplify the retrieval and utilization of historical newspapers through the application of semantic organization, all from the vantage point of a fine-grained knowledge element perspective. This endeavor seeks to unlock the latent value embedded within newspaper contents while simultaneously furnishing invaluable guidance within methodological paradigms for research in the humanities domain.
Design/methodology/approach
According to the semantic organization process and knowledge element concept, this study proposes a holistic framework, including four pivotal stages: knowledge element description, extraction, association and application. Initially, a semantic description model dedicated to knowledge elements is devised. Subsequently, harnessing the advanced deep learning techniques, the study delves into the realm of entity recognition and relationship extraction. These techniques are instrumental in identifying entities within the historical newspaper contents and capturing the interdependencies that exist among them. Finally, an online platform based on Flask is developed to enable the recognition of entities and relationships within historical newspapers.
Findings
This article utilized the Shengjing Times·Changchun Compilation as the datasets for describing, extracting, associating and applying newspapers contents. Regarding knowledge element extraction, the BERT + BS consistently outperforms Bi-LSTM, CRF++ and even BERT in terms of Recall and F1 scores, making it a favorable choice for entity recognition in this context. Particularly noteworthy is the Bi-LSTM-Pro model, which stands out with the highest scores across all metrics, notably achieving an exceptional F1 score in knowledge element relationship recognition.
Originality/value
Historical newspapers transcend their status as mere artifacts, evolving into invaluable reservoirs safeguarding the societal and historical memory. Through semantic organization from a fine-grained knowledge element perspective, it can facilitate semantic retrieval, semantic association, information visualization and knowledge discovery services for historical newspapers. In practice, it can empower researchers to unearth profound insights within the historical and cultural context, broadening the landscape of digital humanities research and practical applications.
Details
Keywords
Ruan Wang, Jun Deng, Xinhui Guan and Yuming He
With the development of data mining technology, diverse and broader domain knowledge can be extracted automatically. However, the research on applying knowledge mapping and data…
Abstract
Purpose
With the development of data mining technology, diverse and broader domain knowledge can be extracted automatically. However, the research on applying knowledge mapping and data visualization techniques to genealogical data is limited. This paper aims to fill this research gap by providing a systematic framework and process guidance for practitioners seeking to uncover hidden knowledge from genealogy.
Design/methodology/approach
Based on a literature review of genealogy's current knowledge reasoning research, the authors constructed an integrated framework for knowledge inference and visualization application using a knowledge graph. Additionally, the authors applied this framework in a case study using “Manchu Clan Genealogy” as the data source.
Findings
The case study shows that the proposed framework can effectively decompose and reconstruct genealogy. It demonstrates the reasoning, discovery, and web visualization application process of implicit information in genealogy. It enhances the effective utilization of Manchu genealogy resources by highlighting the intricate relationships among people, places, and time entities.
Originality/value
This study proposed a framework for genealogy knowledge reasoning and visual analysis utilizing a knowledge graph, including five dimensions: the target layer, the resource layer, the data layer, the inference layer, and the application layer. It helps to gather the scattered genealogy information and establish a data network with semantic correlations while establishing reasoning rules to enable inference discovery and visualization of hidden relationships.
Details
Keywords
The article extends the distinction of semantic from syntactic labour to comprehend all forms of mental labour. It answers a critique from de Fremery and Buckland, which required…
Abstract
Purpose
The article extends the distinction of semantic from syntactic labour to comprehend all forms of mental labour. It answers a critique from de Fremery and Buckland, which required envisaging mental labour as a differentiated spectrum.
Design/methodology/approach
The paper adopts a discursive approach. It first reviews the significance and extensive diffusion of the distinction of semantic from syntactic labour. Second, it integrates semantic and syntactic labour along a vertical dimension within mental labour, indicating analogies in principle with, and differences in application from, the inherited distinction of intellectual from clerical labour. Third, it develops semantic labour to the very highest level, on a consistent principle of differentiation from syntactic labour. Finally, it reintegrates the understanding developed of semantic labour with syntactic labour, confirming that they can fully and informatively occupy mental labour.
Findings
The article further validates the distinction of semantic from syntactic labour. It enables to address Norbert Wiener's classic challenge of appropriately distributing activity between human and computer.
Research limitations/implications
The article transforms work in progress into knowledge for diffusion.
Practical implications
It has practical implications for determining what tasks to delegate to computational technology.
Social implications
The paper has social implications for the understanding of appropriate human and machine computational tasks and our own distinctive humanness.
Originality/value
The paper is highly original. Although based on preceding research, from the late 20th century, it is the first separately published full account of semantic and syntactic labour.
Details
Keywords
The purpose of this article is to contribute to the digital development and utilization of China’s intangible cultural heritage resources, research on the theft of intangible…
Abstract
Purpose
The purpose of this article is to contribute to the digital development and utilization of China’s intangible cultural heritage resources, research on the theft of intangible cultural heritage resources and knowledge integration based on linked data is proposed to promote the standardized description of intangible cultural heritage knowledge and realize the digital dissemination and development of intangible cultural heritage.
Design/methodology/approach
In this study, firstly, the knowledge organization theory and semantic Web technology are used to describe the intangible cultural heritage digital resource objects in metadata specifications. Secondly, the ontology theory and technical methods are used to build a conceptual model of the intangible cultural resources field and determine the concept sets and hierarchical relationships in this field. Finally, the semantic Web technology is used to establish semantic associations between intangible cultural heritage resource knowledge.
Findings
The study findings indicate that the knowledge organization of intangible cultural heritage resources constructed in this study provides a solution for the digital development of intangible cultural heritage in China. It also provides semantic retrieval with better knowledge granularity and helps to visualize the knowledge content of intangible cultural heritage.
Originality/value
This study summarizes and provides significant theoretical and practical value for the digital development of intangible cultural heritage and the resource description and knowledge fusion of intangible cultural heritage can help to discover the semantic relationship of intangible cultural heritage in multiple dimensions and levels.
Details
Keywords
Morteza Mohammadi Ostani, Jafar Ebadollah Amoughin and Mohadeseh Jalili Manaf
This study aims to adjust Thesis-type properties on Schema.org using metadata models and standards (MS) (Bibframe, electronic thesis and dissertations [ETD]-MS, Common European…
Abstract
Purpose
This study aims to adjust Thesis-type properties on Schema.org using metadata models and standards (MS) (Bibframe, electronic thesis and dissertations [ETD]-MS, Common European Research Information Format [CERIF] and Dublin Core [DC]) to enrich the Thesis-type properties for better description and processing on the Web.
Design/methodology/approach
This study is applied, descriptive analysis in nature and is based on content analysis in terms of method. The research population consisted of elements and attributes of the metadata model and standards (Bibframe, ETD-MS, CERIF and DC) and Thesis-type properties in the Schema.org. The data collection tool was a researcher-made checklist, and the data collection method was structured observation.
Findings
The results show that the 65 Thesis-type properties and the two levels of Thing and CreativeWork as its parents on Schema.org that corresponds to the elements and attributes of related models and standards. In addition, 12 properties are special to the Thesis type for better comprehensive description and processing, and 27 properties are added to the CreativeWork type.
Practical implications
Enrichment and expansion of Thesis-type properties on Schema.org is one of the practical applications of the present study, which have enabled more comprehensive description and processing and increased access points and visibility for ETDs in the environment Web and digital libraries.
Originality/value
This study has offered some new Thesis type properties and CreativeWork levels on Schema.org. To the best of the authors’ knowledge, this is the first time this issue is investigated.
Details
Keywords
Sudarsan Desul, Rabindra Kumar Mahapatra, Raj Kishore Patra, Mrutyunjay Sethy and Neha Pandey
The purpose of this study is to review the application of semantic technologies in cultural heritage (STCH) to achieve interoperability and enable advanced applications like 3D…
Abstract
Purpose
The purpose of this study is to review the application of semantic technologies in cultural heritage (STCH) to achieve interoperability and enable advanced applications like 3D modeling and augmented reality by enhancing the understanding and appreciation of CH. The study aims to identify the trends and patterns in using STCH and provide insights for scholars and policymakers on future research directions.
Design/methodology/approach
This research paper uses a bibliometric study to analyze the articles published in Scopus and Web of Science (WoS)-indexed journals from 1999 to 2022 on STCH. A total of 580 articles were analyzed using the Biblioshiny package in RStudio.
Findings
The study reveals a substantial increase in STCH publications since 2008, with Italy leading in contributions. Key research areas such as ontologies, semantic Web, linked data and digital humanities are extensively explored, highlighting their significance and characteristics within the STCH research domain.
Research limitations/implications
This study only analyzed articles published in Scopus and WoS-indexed journals in the English language. Further research could include articles published in other languages and non-indexed journals.
Originality/value
This study extensively analyses the research published on STCH over the past 23 years, identifying the leading authors, institutions, countries and top research topics. The findings provide guidelines for future research direction and contribute to the literature on promoting, preserving and managing the CH globally.
Details
Keywords
Na Xu, Yanxiang Liang, Chaoran Guo, Bo Meng, Xueqing Zhou, Yuting Hu and Bo Zhang
Safety management plays an important part in coal mine construction. Due to complex data, the implementation of the construction safety knowledge scattered in standards poses a…
Abstract
Purpose
Safety management plays an important part in coal mine construction. Due to complex data, the implementation of the construction safety knowledge scattered in standards poses a challenge. This paper aims to develop a knowledge extraction model to automatically and efficiently extract domain knowledge from unstructured texts.
Design/methodology/approach
Bidirectional encoder representations from transformers (BERT)-bidirectional long short-term memory (BiLSTM)-conditional random field (CRF) method based on a pre-training language model was applied to carry out knowledge entity recognition in the field of coal mine construction safety in this paper. Firstly, 80 safety standards for coal mine construction were collected, sorted out and marked as a descriptive corpus. Then, the BERT pre-training language model was used to obtain dynamic word vectors. Finally, the BiLSTM-CRF model concluded the entity’s optimal tag sequence.
Findings
Accordingly, 11,933 entities and 2,051 relationships in the standard specifications texts of this paper were identified and a language model suitable for coal mine construction safety management was proposed. The experiments showed that F1 values were all above 60% in nine types of entities such as security management. F1 value of this model was more than 60% for entity extraction. The model identified and extracted entities more accurately than conventional methods.
Originality/value
This work completed the domain knowledge query and built a Q&A platform via entities and relationships identified by the standard specifications suitable for coal mines. This paper proposed a systematic framework for texts in coal mine construction safety to improve efficiency and accuracy of domain-specific entity extraction. In addition, the pretraining language model was also introduced into the coal mine construction safety to realize dynamic entity recognition, which provides technical support and theoretical reference for the optimization of safety management platforms.
Details
Keywords
Recent archiving and curatorial practices took advantage of the advancement in digital technologies, creating immersive and interactive experiences to emphasize the plurality of…
Abstract
Purpose
Recent archiving and curatorial practices took advantage of the advancement in digital technologies, creating immersive and interactive experiences to emphasize the plurality of memory materials, encourage personalized sense-making and extract, manage and share the ever-growing surrounding knowledge. Audiovisual (AV) content, with its growing importance and popularity, is less explored on that end than texts and images. This paper examines the trend of datafication in AV archives and answers the critical question, “What to extract from AV materials and why?”.
Design/methodology/approach
This study roots in a comprehensive state-of-the-art review of digital methods and curatorial practices in AV archives. The thinking model for mapping AV archive data to purposes is based on pre-existing models for understanding multimedia content and metadata standards.
Findings
The thinking model connects AV content descriptors (data perspective) and purposes (curatorial perspective) and provides a theoretical map of how information extracted from AV archives should be fused and embedded for memory institutions. The model is constructed by looking into the three broad dimensions of audiovisual content – archival, affective and aesthetic, social and historical.
Originality/value
This paper contributes uniquely to the intersection of computational archives, audiovisual content and public sense-making experiences. It provides updates and insights to work towards datafied AV archives and cope with the increasing needs in the sense-making end using AV archives.
Details
Keywords
Integrating the Chat Generative Pre-Trained Transformer-type (ChatGPT-type) model with government services has great development prospects. Applying this model improves service…
Abstract
Purpose
Integrating the Chat Generative Pre-Trained Transformer-type (ChatGPT-type) model with government services has great development prospects. Applying this model improves service efficiency but has certain risks, thus having a dual impact on the public. For a responsible and democratic government, it is necessary to fully understand the factors influencing public acceptance and their causal relationships to truly encourage the public to accept and use government ChatGPT-type services.
Design/methodology/approach
This study used the Latent Dirichlet allocation (LDA) model to analyze comment texts and summarize 15 factors that affect public acceptance. Multiple-related matrices were established using the grey decision-making trial and evaluation laboratory (grey-DEMATEL) method to reveal causal relationships among factors. From the two opposite extraction rules of result priority and cause priority, the authors obtained an antagonistic topological model with comprehensive influence values using the total adversarial interpretive structure model (TAISM).
Findings
Fifteen factors were categorized in terms of cause and effect, and the antagonistic topological model with comprehensive influence values was also analyzed. The analysis showed that perceived risk, trust and meeting demand were the three most critical factors of public acceptance. Meanwhile, perceived risk and trust directly affected public acceptance and were affected by other factors. Supervision and accountability had the highest driving power and acted as the causal factor to influence other factors.
Originality/value
This study identified the factors affecting public acceptance of integrating the ChatGPT-type model with government services. It analyzed the relationship between the factors to provide a reference for decision-makers. This study introduced TAISM to form the LDA-grey-DEMATEL-TAISM method to provide an analytical paradigm for studying similar influencing factors.
Details