Search results
1 – 10 of over 7000Chuanming Yu, Zhengang Zhang, Lu An and Gang Li
In recent years, knowledge graph completion has gained increasing research focus and shown significant improvements. However, most existing models only use the structures of…
Abstract
Purpose
In recent years, knowledge graph completion has gained increasing research focus and shown significant improvements. However, most existing models only use the structures of knowledge graph triples when obtaining the entity and relationship representations. In contrast, the integration of the entity description and the knowledge graph network structure has been ignored. This paper aims to investigate how to leverage both the entity description and the network structure to enhance the knowledge graph completion with a high generalization ability among different datasets.
Design/methodology/approach
The authors propose an entity-description augmented knowledge graph completion model (EDA-KGC), which incorporates the entity description and network structure. It consists of three modules, i.e. representation initialization, deep interaction and reasoning. The representation initialization module utilizes entity descriptions to obtain the pre-trained representation of entities. The deep interaction module acquires the features of the deep interaction between entities and relationships. The reasoning component performs matrix manipulations with the deep interaction feature vector and entity representation matrix, thus obtaining the probability distribution of target entities. The authors conduct intensive experiments on the FB15K, WN18, FB15K-237 and WN18RR data sets to validate the effect of the proposed model.
Findings
The experiments demonstrate that the proposed model outperforms the traditional structure-based knowledge graph completion model and the entity-description-enhanced knowledge graph completion model. The experiments also suggest that the model has greater feasibility in different scenarios such as sparse data, dynamic entities and limited training epochs. The study shows that the integration of entity description and network structure can significantly increase the effect of the knowledge graph completion task.
Originality/value
The research has a significant reference for completing the missing information in the knowledge graph and improving the application effect of the knowledge graph in information retrieval, question answering and other fields.
Details
Keywords
Wenjing Wu, Caifeng Wen, Qi Yuan, Qiulan Chen and Yunzhong Cao
Learning from safety accidents and sharing safety knowledge has become an important part of accident prevention and improving construction safety management. Considering the…
Abstract
Purpose
Learning from safety accidents and sharing safety knowledge has become an important part of accident prevention and improving construction safety management. Considering the difficulty of reusing unstructured data in the construction industry, the knowledge in it is difficult to be used directly for safety analysis. The purpose of this paper is to explore the construction of construction safety knowledge representation model and safety accident graph through deep learning methods, extract construction safety knowledge entities through BERT-BiLSTM-CRF model and propose a data management model of data–knowledge–services.
Design/methodology/approach
The ontology model of knowledge representation of construction safety accidents is constructed by integrating entity relation and logic evolution. Then, the database of safety incidents in the architecture, engineering and construction (AEC) industry is established based on the collected construction safety incident reports and related dispute cases. The construction method of construction safety accident knowledge graph is studied, and the precision of BERT-BiLSTM-CRF algorithm in information extraction is verified through comparative experiments. Finally, a safety accident report is used as an example to construct the AEC domain construction safety accident knowledge graph (AEC-KG), which provides visual query knowledge service and verifies the operability of knowledge management.
Findings
The experimental results show that the combined BERT-BiLSTM-CRF algorithm has a precision of 84.52%, a recall of 92.35%, and an F1 value of 88.26% in named entity recognition from the AEC domain database. The construction safety knowledge representation model and safety incident knowledge graph realize knowledge visualization.
Originality/value
The proposed framework provides a new knowledge management approach to improve the safety management of practitioners and also enriches the application scenarios of knowledge graph. On the one hand, it innovatively proposes a data application method and knowledge management method of safety accident report that integrates entity relationship and matter evolution logic. On the other hand, the legal adjudication dimension is innovatively added to the knowledge graph in the construction safety field as the basis for the postincident disposal measures of safety accidents, which provides reference for safety managers' decision-making in all aspects.
Details
Keywords
Xiang Zheng, Mingjie Li, Ze Wan and Yan Zhang
This study aims to extract knowledge of ancient Chinese scientific and technological documents bibliographic summaries (STDBS) and provide the knowledge graph (KG) comprehensively…
Abstract
Purpose
This study aims to extract knowledge of ancient Chinese scientific and technological documents bibliographic summaries (STDBS) and provide the knowledge graph (KG) comprehensively and systematically. By presenting the relationship among content, discipline, and author, this study focuses on providing services for knowledge discovery of ancient Chinese scientific and technological documents.
Design/methodology/approach
This study compiles ancient Chinese STDBS and designs a knowledge mining and graph visualization framework. The authors define the summaries' entities, attributes, and relationships for knowledge representation, use deep learning techniques such as BERT-BiLSTM-CRF models and rules for knowledge extraction, unify the representation of entities for knowledge fusion, and use Neo4j and other visualization techniques for KG construction and application. This study presents the generation, distribution, and evolution of ancient Chinese agricultural scientific and technological knowledge in visualization graphs.
Findings
The knowledge mining and graph visualization framework is feasible and effective. The BERT-BiLSTM-CRF model has domain adaptability and accuracy. The knowledge generation of ancient Chinese agricultural scientific and technological documents has distinctive time features. The knowledge distribution is uneven and concentrated, mainly concentrated on C1-Planting and cultivation, C2-Silkworm, and C3-Mulberry and water conservancy. The knowledge evolution is apparent, and differentiation and integration coexist.
Originality/value
This study is the first to visually present the knowledge connotation and association of ancient Chinese STDBS. It solves the problems of the lack of in-depth knowledge mining and connotation visualization of ancient Chinese STDBS.
Details
Keywords
Margie Foster, Hossein Arvand, Hugh T. Graham and Denise Bedford
This chapter makes a case for extending institutional preservation strategies to the entire landscape of knowledge capital. First, the authors define the three primary types of…
Abstract
Chapter Summary
This chapter makes a case for extending institutional preservation strategies to the entire landscape of knowledge capital. First, the authors define the three primary types of capital – physical, financial, and knowledge. Knowledge capital is further broken down into three categories – human, structural, and relational. The individual types of knowledge capital are defined, along with their variant economic properties and behaviors. The challenges these variations present for preservation are discussed. The authors also highlight these assets’ significant opportunities for curating new knowledge. Each type of knowledge capital is described, along with the preservation challenges and the curation opportunities.
This study aims to develop a synthetic knowledge repository consisted of interrelated Web Ontology Language.
Abstract
Purpose
This study aims to develop a synthetic knowledge repository consisted of interrelated Web Ontology Language.
Design/methodology/approach
The ontology composes the main framework to categorize data of product life cycle with eco-design mode (PLC-EDM) and automatically infer specialists’ knowledge for data confirmation, eventually assisting the utilizations and generation of strategies toward decision-making
Findings
(i) utilization of a novel model with ontology mode for information reuse cross the different eco-design applications; (ii) generation of a sound platform toward life cycle evaluation; and (iii) implementation of the PLC-EDM model along the product generation process.
Research limitations/implications
It cannot substitute an evaluation tool of life cycle. Certainly, this model does not predict the “target and range” and/or the depiction of the “utility module” that are basic activities in life cycle assessments as characterized through the international organization for standardization regulations.
Practical implications
As portion of this framework, a prototype Web application is presented which is applied to produce, reuse and verify knowledge of product life cycle.
Social implications
By counting upon the ontology, the information conducted by the utilization is certainly semantically represented to promote the data sharing among various participants and tools. Besides, the data can be verified against possible faults by inferring over the ontology. Hence, a feasible way to a popular topic in the domain of eco-design applications extension in the industry.
Originality/value
The goals are: to lean on rigid modeling principles; and to promote the interoperability and diffusion of the ontology toward particular utilization demands.
Details
Keywords
Edoardo Ramalli and Barbara Pernici
Experiments are the backbone of the development process of data-driven predictive models for scientific applications. The quality of the experiments directly impacts the model…
Abstract
Purpose
Experiments are the backbone of the development process of data-driven predictive models for scientific applications. The quality of the experiments directly impacts the model performance. Uncertainty inherently affects experiment measurements and is often missing in the available data sets due to its estimation cost. For similar reasons, experiments are very few compared to other data sources. Discarding experiments based on the missing uncertainty values would preclude the development of predictive models. Data profiling techniques are fundamental to assess data quality, but some data quality dimensions are challenging to evaluate without knowing the uncertainty. In this context, this paper aims to predict the missing uncertainty of the experiments.
Design/methodology/approach
This work presents a methodology to forecast the experiments’ missing uncertainty, given a data set and its ontological description. The approach is based on knowledge graph embeddings and leverages the task of link prediction over a knowledge graph representation of the experiments database. The validity of the methodology is first tested in multiple conditions using synthetic data and then applied to a large data set of experiments in the chemical kinetic domain as a case study.
Findings
The analysis results of different test case scenarios suggest that knowledge graph embedding can be used to predict the missing uncertainty of the experiments when there is a hidden relationship between the experiment metadata and the uncertainty values. The link prediction task is also resilient to random noise in the relationship. The knowledge graph embedding outperforms the baseline results if the uncertainty depends upon multiple metadata.
Originality/value
The employment of knowledge graph embedding to predict the missing experimental uncertainty is a novel alternative to the current and more costly techniques in the literature. Such contribution permits a better data quality profiling of scientific repositories and improves the development process of data-driven models based on scientific experiments.
Details
Keywords
Elaheh Hosseini, Kimiya Taghizadeh Milani and Mohammad Shaker Sabetnasab
This research aimed to visualize and analyze the co-word network and thematic clusters of the intellectual structure in the field of linked data during 1900–2021.
Abstract
Purpose
This research aimed to visualize and analyze the co-word network and thematic clusters of the intellectual structure in the field of linked data during 1900–2021.
Design/methodology/approach
This applied research employed a descriptive and analytical method, scientometric indicators, co-word techniques, and social network analysis. VOSviewer, SPSS, Python programming, and UCINet software were used for data analysis and network structure visualization.
Findings
The top ranks of the Web of Science (WOS) subject categorization belonged to various fields of computer science. Besides, the USA was the most prolific country. The keyword ontology had the highest frequency of co-occurrence. Ontology and semantic were the most frequent co-word pairs. In terms of the network structure, nine major topic clusters were identified based on co-occurrence, and 29 thematic clusters were identified based on hierarchical clustering. Comparisons between the two clustering techniques indicated that three clusters, namely semantic bioinformatics, knowledge representation, and semantic tools were in common. The most mature and mainstream thematic clusters were natural language processing techniques to boost modeling and visualization, context-aware knowledge discovery, probabilistic latent semantic analysis (PLSA), semantic tools, latent semantic indexing, web ontology language (OWL) syntax, and ontology-based deep learning.
Originality/value
This study adopted various techniques such as co-word analysis, social network analysis network structure visualization, and hierarchical clustering to represent a suitable, visual, methodical, and comprehensive perspective into linked data.
Details
Keywords
Anish Khobragade, Shashikant Ghumbre and Vinod Pachghare
MITRE and the National Security Agency cooperatively developed and maintained a D3FEND knowledge graph (KG). It provides concepts as an entity from the cybersecurity…
Abstract
Purpose
MITRE and the National Security Agency cooperatively developed and maintained a D3FEND knowledge graph (KG). It provides concepts as an entity from the cybersecurity countermeasure domain, such as dynamic, emulated and file analysis. Those entities are linked by applying relationships such as analyze, may_contains and encrypt. A fundamental challenge for collaborative designers is to encode knowledge and efficiently interrelate the cyber-domain facts generated daily. However, the designers manually update the graph contents with new or missing facts to enrich the knowledge. This paper aims to propose an automated approach to predict the missing facts using the link prediction task, leveraging embedding as representation learning.
Design/methodology/approach
D3FEND is available in the resource description framework (RDF) format. In the preprocessing step, the facts in RDF format converted to subject–predicate–object triplet format contain 5,967 entities and 98 relationship types. Progressive distance-based, bilinear and convolutional embedding models are applied to learn the embeddings of entities and relations. This study presents a link prediction task to infer missing facts using learned embeddings.
Findings
Experimental results show that the translational model performs well on high-rank results, whereas the bilinear model is superior in capturing the latent semantics of complex relationship types. However, the convolutional model outperforms 44% of the true facts and achieves a 3% improvement in results compared to other models.
Research limitations/implications
Despite the success of embedding models to enrich D3FEND using link prediction under the supervised learning setup, it has some limitations, such as not capturing diversity and hierarchies of relations. The average node degree of D3FEND KG is 16.85, with 12% of entities having a node degree less than 2, especially there are many entities or relations with few or no observed links. This results in sparsity and data imbalance, which affect the model performance even after increasing the embedding vector size. Moreover, KG embedding models consider existing entities and relations and may not incorporate external or contextual information such as textual descriptions, temporal dynamics or domain knowledge, which can enhance the link prediction performance.
Practical implications
Link prediction in the D3FEND KG can benefit cybersecurity countermeasure strategies in several ways, such as it can help to identify gaps or weaknesses in the existing defensive methods and suggest possible ways to improve or augment them; it can help to compare and contrast different defensive methods and understand their trade-offs and synergies; it can help to discover novel or emerging defensive methods by inferring new relations from existing data or external sources; and it can help to generate recommendations or guidance for selecting or deploying appropriate defensive methods based on the characteristics and objectives of the system or network.
Originality/value
The representation learning approach helps to reduce incompleteness using a link prediction that infers possible missing facts by using the existing entities and relations of D3FEND.
Details
Keywords
Abigail Richard, Fred Ahrens and Benjamin George
This study aims to introduce a new prescriptive model to aid both managers and researchers in partner selection for innovation-orientated collaboration. This framework…
Abstract
Purpose
This study aims to introduce a new prescriptive model to aid both managers and researchers in partner selection for innovation-orientated collaboration. This framework demonstrates how prospective partner firms’ complementing bodies of knowledge and goal alignment interact to affect the success of a collaboration.
Design/methodology/approach
The authors use geometric modeling to represent the interrelationships among knowledge similarity/dissimilarity, goal congruence, knowledge complementarity (KC) and innovation in alliance formation. Using this model as a framework, the authors derive relationships among predictors of innovation success and determine how they affect the nature of partnerships under varying conditions of KC.
Findings
This research shows how innovation success is strongly determined by partner selection. Specifically, the authors examine the influence of KC and partner goals on three aspects of a potential research and development (R&D) alliance – the potential level of innovation outcome for the alliance, the boundaries of knowledge sharing and limitations arising from knowledge and goal incongruence and the nature of cooperation.
Originality/value
Although there is broad empirical support that innovation success is influenced by the similarity of R&D partners’ knowledge, further research is still needed to model the relationship more precisely between partner KC and goal alignment. The authors address this gap by developing a model that is both prescriptive and predictive of how innovation success can be achieved in the context of disparate but complementing knowledge and goal sets. The authors conclude with practical implications for practice and future research directions.
Details
Keywords
Pierre Jouan and Pierre Hallot
The purpose of this paper is to address the challenging issue of developing a quantitative approach for the representation of cultural significance data in heritage information…
Abstract
Purpose
The purpose of this paper is to address the challenging issue of developing a quantitative approach for the representation of cultural significance data in heritage information systems (HIS). The authors propose to provide experts in the field with a dedicated framework to structure and integrate targeted data about historical objects' significance in such environments.
Design/methodology/approach
This research seeks the identification of key indicators which allow to better inform decision-makers about cultural significance. Identified concepts are formalized in a data structure through conceptual data modeling, taking advantage on unified modeling language (HIS). The design science research (DSR) method is implemented to facilitate the development of the data model.
Findings
This paper proposes a practical solution for the formalization of data related to the significance of objects in HIS. The authors end up with a data model which enables multiple knowledge representations through data analysis and information retrieval.
Originality/value
The framework proposed in this article supports a more sustainable vision of heritage preservation as the framework enhances the involvement of all stakeholders in the conservation and management of historical sites. The data model supports explicit communications of the significance of historical objects and strengthens the synergy between the stakeholders involved in different phases of the conservation process.
Details