Search results

1 – 10 of 32
Article
Publication date: 3 October 2023

Haklae Kim

Despite ongoing research into archival metadata standards, digital archives are unable to effectively represent records in their appropriate contexts. This study aims to propose a…

Abstract

Purpose

Despite ongoing research into archival metadata standards, digital archives are unable to effectively represent records in their appropriate contexts. This study aims to propose a knowledge graph that depicts the diverse relationships between heterogeneous digital archive entities.

Design/methodology/approach

This study introduces and describes a method for applying knowledge graphs to digital archives in a step-by-step manner. It examines archival metadata standards, such as Records in Context Ontology (RiC-O), for characterising digital records; explains the process of data refinement, enrichment and reconciliation with examples; and demonstrates the use of knowledge graphs constructed using semantic queries.

Findings

This study introduced the 97imf.kr archive as a knowledge graph, enabling meaningful exploration of relationships within the archive’s records. This approach facilitated comprehensive record descriptions about different record entities. Applying archival ontologies with general-purpose vocabularies to digital records was advised to enhance metadata coherence and semantic search.

Originality/value

Most digital archives serviced in Korea are limited in the proper use of archival metadata standards. The contribution of this study is to propose a practical application of knowledge graph technology for linking and exploring digital records. This study details the process of collecting raw data on archives, data preprocessing and data enrichment, and demonstrates how to build a knowledge graph connected to external data. In particular, the knowledge graph of RiC-O vocabulary, Wikidata and Schema.org vocabulary and the semantic query using it can be applied to supplement keyword search in conventional digital archives.

Details

The Electronic Library , vol. 42 no. 1
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 9 November 2023

Gustavo Candela, Nele Gabriëls, Sally Chambers, Milena Dobreva, Sarah Ames, Meghan Ferriter, Neil Fitzgerald, Victor Harbo, Katrine Hofmann, Olga Holownia, Alba Irollo, Mahendra Mahey, Eileen Manchester, Thuy-An Pham, Abigail Potter and Ellen Van Keer

The purpose of this study is to offer a checklist that can be used for both creating and evaluating digital collections, which are also sometimes referred to as data sets as part…

Abstract

Purpose

The purpose of this study is to offer a checklist that can be used for both creating and evaluating digital collections, which are also sometimes referred to as data sets as part of the collections as data movement, suitable for computational use.

Design/methodology/approach

The checklist was built by synthesising and analysing the results of relevant research literature, articles and studies and the issues and needs obtained in an observational study. The checklist was tested and applied both as a tool for assessing a selection of digital collections made available by galleries, libraries, archives and museums (GLAM) institutions as proof of concept and as a supporting tool for creating collections as data.

Findings

Over the past few years, there has been a growing interest in making available digital collections published by GLAM organisations for computational use. Based on previous work, the authors defined a methodology to build a checklist for the publication of Collections as data. The authors’ evaluation showed several examples of applications that can be useful to encourage other institutions to publish their digital collections for computational use.

Originality/value

While some work on making available digital collections suitable for computational use exists, giving particular attention to data quality, planning and experimentation, to the best of the authors’ knowledge, none of the work to date provides an easy-to-follow and robust checklist to publish collection data sets in GLAM institutions. This checklist intends to encourage small- and medium-sized institutions to adopt the collection as data principles in daily workflows following best practices and guidelines.

Details

Global Knowledge, Memory and Communication, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2514-9342

Keywords

Article
Publication date: 10 March 2023

Marta Ortiz-de-Urbina-Criado, Alberto Abella and Diego García-Luna

This paper aims to highlight the importance of open data and the role that knowledge management and open innovation can play in its identification and use. Open data has great…

Abstract

Purpose

This paper aims to highlight the importance of open data and the role that knowledge management and open innovation can play in its identification and use. Open data has great potential to create social and economic value, but its main problem is that it is often not easily reusable. The aim of this paper is to propose a unique identifier for open data-sets that would facilitate search and access to them and help to reduce heterogeneity in the publication of data in open data portals.

Design/methodology/approach

Considering a model of the impact process of open data reuse and based on the digital object identifier system, this paper develops a proposal of a unique identifier for open data-sets called Open Data-set Identifier (OpenDatId).

Findings

This paper presents some examples of the application and advantages of OpenDatId. For example, users can easily consult the available content catalogues, search the data in an automated way and examine the content for reuse. It is also possible to find out where this data comes from, solving the problems caused by the increasingly frequent federation of data in open data portals and enabling the creation of additional services based on open data.

Originality/value

From an integrated perspective of knowledge management and open innovation, this paper presents a new unique identifier for open data-sets (OpenDatId) and a new concept for data-set, the FAIR Open Data-sets.

Details

Journal of Knowledge Management, vol. 27 no. 10
Type: Research Article
ISSN: 1367-3270

Keywords

Article
Publication date: 25 January 2024

Besiki Stvilia and Dong Joon Lee

This study addresses the need for a theory-guided, rich, descriptive account of research data repositories' (RDRs) understanding of data quality and the structures of their data…

Abstract

Purpose

This study addresses the need for a theory-guided, rich, descriptive account of research data repositories' (RDRs) understanding of data quality and the structures of their data quality assurance (DQA) activities. Its findings can help develop operational DQA models and best practice guides and identify opportunities for innovation in the DQA activities.

Design/methodology/approach

The study analyzed 122 data repositories' applications for the Core Trustworthy Data Repositories, interview transcripts of 32 curators and repository managers and data curation-related webpages of their repository websites. The combined dataset represented 146 unique RDRs. The study was guided by a theoretical framework comprising activity theory and an information quality evaluation framework.

Findings

The study provided a theory-based examination of the DQA practices of RDRs summarized as a conceptual model. The authors identified three DQA activities: evaluation, intervention and communication and their structures, including activity motivations, roles played and mediating tools and rules and standards. When defining data quality, study participants went beyond the traditional definition of data quality and referenced seven facets of ethical and effective information systems in addition to data quality. Furthermore, the participants and RDRs referenced 13 dimensions in their DQA models. The study revealed that DQA activities were prioritized by data value, level of quality, available expertise, cost and funding incentives.

Practical implications

The study's findings can inform the design and construction of digital research data curation infrastructure components on university campuses that aim to provide access not just to big data but trustworthy data. Communities of practice focused on repositories and archives could consider adding FAIR operationalizations, extensions and metrics focused on data quality. The availability of such metrics and associated measurements can help reusers determine whether they can trust and reuse a particular dataset. The findings of this study can help to develop such data quality assessment metrics and intervention strategies in a sound and systematic way.

Originality/value

To the best of the authors' knowledge, this paper is the first data quality theory guided examination of DQA practices in RDRs.

Details

Journal of Documentation, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0022-0418

Keywords

Article
Publication date: 9 April 2024

Ishrat Ayub Sofi, Ajra Bhat and Rahat Gulzar

The study aims to shed light on the current state of “Dataset repositories” indexed in Directory of Open Access Repositories (OpenDOAR).

Abstract

Purpose

The study aims to shed light on the current state of “Dataset repositories” indexed in Directory of Open Access Repositories (OpenDOAR).

Design/methodology/approach

From each repository/record information, the Open-Access Policies, Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH), year of creation and the number of data sets archived in the repositories were manually searched, documented and analyzed.

Findings

Developed countries like the United Kingdom and the USA are primarily involved in the development of institutional open-access repositories comprising significant components of OpenDOAR. The most extensively used software is DSpace. Most data set archives are OAI-PMH compliant but do not follow open-access rules. The study also highlights the sites’ embrace of Web 2.0 capabilities and discovers really simple syndication feeds and Atom integration. The use of social media has made its presence known. Furthermore, the study concludes that the number of data sets kept in repositories is insufficient, although the expansion of such repositories has been consistent over the years.

Practical implications

The work has the potential to benefit both researchers in general and policymakers in particular. Scholars interested in research data, data sharing and data reuse can learn about the present state of repositories that preserve data sets in OpenDOAR. At the same time, policymakers can develop recommendations and policies to assist in the construction and maintenance of repositories for data sets.

Originality/value

According to the literature, there have been numerous studies on open-access repositories and OpenDOAR internationally, but no research has focused on repositories preserving content-type data sets. As a result, the study attempts to uncover various characteristics of OpenDOAR Data set repositories.

Details

Digital Library Perspectives, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2059-5816

Keywords

Open Access
Article
Publication date: 30 April 2024

Marguerite Alice Nel, Pfano Makhera, Mabjala Mercia Moreana and Marinda Maritz

Although universities have extensive research and initiatives in place that align with the United Nations’ Sustainable Development Goals (SDGs), there is still a significant gap…

Abstract

Purpose

Although universities have extensive research and initiatives in place that align with the United Nations’ Sustainable Development Goals (SDGs), there is still a significant gap in documenting and assessing these efforts. This paper aims to discuss how academic libraries can apply their information management skills and open-access platforms, to facilitate the discoverability and retrieval of evidence on SDGs.

Design/methodology/approach

Introduced by a brief literature review on the role of libraries in contributing to the SDGs in general, the authors draw on their personal experiences as metadata specialists, participating in a project aimed at linking their university’s research output to the SDGs. A case study, from the University of Pretoria’s Veterinary Science Library, is used as an example to demonstrate the benefits of resourceful metadata in organising, communicating and raising awareness about the SDGs in the field of veterinary science.

Findings

Through practical examples and recommended workflows, this paper illustrates that metadata specialists are perfectly positioned to apply their information management skills and library platforms to facilitate the discoverability and retrieval of evidence on SDGs.

Originality/value

Although there are increasing reports on the contributions of libraries to support the successful implementation of the SDGs, limited information exists on the role of metadata specialists, as well as those with a practical focus.

Details

Digital Library Perspectives, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2059-5816

Keywords

Article
Publication date: 20 November 2023

Laksmi Laksmi, Muhammad Fadly Suhendra, Shamila Mohamed Shuhidan and Umanto Umanto

This study aims to identify the readiness of institutional repositories in Indonesia to implement digital humanities (DH) data curation. Data curation is a method of managing…

Abstract

Purpose

This study aims to identify the readiness of institutional repositories in Indonesia to implement digital humanities (DH) data curation. Data curation is a method of managing research data that maintains the data’s accuracy and makes it available for reuse. It requires controlled data management.

Design/methodology/approach

The study uses a qualitative approach. Data collection was carried out through a focus group discussion in September–October 2022, interviews and document analysis. The informants came from four institutions in Indonesia.

Findings

The findings reveal that the national research repository has implemented data curation, albeit not optimally. Within the case study, one of the university repositories diligently curates its humanities data and has established networks extending to various ASEAN countries. Both the national archive repository and the other university repository have implemented rudimentary data curation practices but have not prioritized them. In conclusion, the readiness of the national research repository and the university repository stand at the high-capacity stage, while the national archive repository and the other university repository are at the established and early stages of data curation, respectively.

Research limitations/implications

This study examined only four repositories due to time constraints. Nonetheless, the four institutions were able to provide a comprehensive picture of their readiness for DH data curation management.

Practical implications

This study provides insight into strategies for developing DH data curation activities in institutional repositories. It also highlights the need for professional development for curators so they can devise and implement stronger ownership policies and data privacy to support a data-driven research agenda.

Originality/value

This study describes the preparations that must be considered by institutional repositories in the development of DH data curation activities.

Open Access
Article
Publication date: 1 March 2022

Elisabetta Colucci, Francesca Matrone, Francesca Noardo, Vanessa Assumma, Giulia Datola, Federica Appiotti, Marta Bottero, Filiberto Chiabrando, Patrizia Lombardi, Massimo Migliorini, Enrico Rinaldi, Antonia Spanò and Andrea Lingua

The study, within the Increasing Resilience of Cultural Heritage (ResCult) project, aims to support civil protection to prevent, lessen and mitigate disasters impacts on cultural…

2062

Abstract

Purpose

The study, within the Increasing Resilience of Cultural Heritage (ResCult) project, aims to support civil protection to prevent, lessen and mitigate disasters impacts on cultural heritage using a unique standardised-3D geographical information system (GIS), including both heritage and risk and hazard information.

Design/methodology/approach

A top-down approach, starting from existing standards (an INSPIRE extension integrated with other parts from the standardised and shared structure), was completed with a bottom-up integration according to current requirements for disaster prevention procedures and risk analyses. The results were validated and tested in case studies (differentiated concerning the hazard and type of protected heritage) and refined during user forums.

Findings

Besides the ensuing reusable database structure, the filling with case studies data underlined the tough challenges and allowed proposing a sample of workflows and possible guidelines. The interfaces are provided to use the obtained knowledge base.

Originality/value

The increasing number of natural disasters could severely damage the cultural heritage, causing permanent damage to movable and immovable assets and tangible and intangible heritage. The study provides an original tool properly relating the (spatial) information regarding cultural heritage and the risk factors in a unique archive as a standard-based European tool to cope with these frequent losses, preventing risk.

Details

Journal of Cultural Heritage Management and Sustainable Development, vol. 14 no. 2
Type: Research Article
ISSN: 2044-1266

Keywords

Article
Publication date: 13 October 2023

Judit Gárdos, Julia Egyed-Gergely, Anna Horváth, Balázs Pataki, Roza Vajda and András Micsik

The present study is about generating metadata to enhance thematic transparency and facilitate research on interview collections at the Research Documentation Centre, Centre for…

Abstract

Purpose

The present study is about generating metadata to enhance thematic transparency and facilitate research on interview collections at the Research Documentation Centre, Centre for Social Sciences (TK KDK) in Budapest. It explores the use of artificial intelligence (AI) in producing, managing and processing social science data and its potential to generate useful metadata to describe the contents of such archives on a large scale.

Design/methodology/approach

The authors combined manual and automated/semi-automated methods of metadata development and curation. The authors developed a suitable domain-oriented taxonomy to classify a large text corpus of semi-structured interviews. To this end, the authors adapted the European Language Social Science Thesaurus (ELSST) to produce a concise, hierarchical structure of topics relevant in social sciences. The authors identified and tested the most promising natural language processing (NLP) tools supporting the Hungarian language. The results of manual and machine coding will be presented in a user interface.

Findings

The study describes how an international social scientific taxonomy can be adapted to a specific local setting and tailored to be used by automated NLP tools. The authors show the potential and limitations of existing and new NLP methods for thematic assignment. The current possibilities of multi-label classification in social scientific metadata assignment are discussed, i.e. the problem of automated selection of relevant labels from a large pool.

Originality/value

Interview materials have not yet been used for building manually annotated training datasets for automated indexing of scientifically relevant topics in a data repository. Comparing various automated-indexing methods, this study shows a possible implementation of a researcher tool supporting custom visualizations and the faceted search of interview collections.

Open Access
Article
Publication date: 17 November 2023

Peiman Tavakoli, Ibrahim Yitmen, Habib Sadri and Afshin Taheri

The purpose of this study is to focus on structured data provision and asset information model maintenance and develop a data provenance model on a blockchain-based digital twin…

Abstract

Purpose

The purpose of this study is to focus on structured data provision and asset information model maintenance and develop a data provenance model on a blockchain-based digital twin smart and sustainable built environment (DT) for predictive asset management (PAM) in building facilities.

Design/methodology/approach

Qualitative research data were collected through a comprehensive scoping review of secondary sources. Additionally, primary data were gathered through interviews with industry specialists. The analysis of the data served as the basis for developing blockchain-based DT data provenance models and scenarios. A case study involving a conference room in an office building in Stockholm was conducted to assess the proposed data provenance model. The implementation utilized the Remix Ethereum platform and Sepolia testnet.

Findings

Based on the analysis of results, a data provenance model on blockchain-based DT which ensures the reliability and trustworthiness of data used in PAM processes was developed. This was achieved by providing a transparent and immutable record of data origin, ownership and lineage.

Practical implications

The proposed model enables decentralized applications (DApps) to publish real-time data obtained from dynamic operations and maintenance processes, enhancing the reliability and effectiveness of data for PAM.

Originality/value

The research presents a data provenance model on a blockchain-based DT, specifically tailored to PAM in building facilities. The proposed model enhances decision-making processes related to PAM by ensuring data reliability and trustworthiness and providing valuable insights for specialists and stakeholders interested in the application of blockchain technology in asset management and data provenance.

Details

Smart and Sustainable Built Environment, vol. 13 no. 1
Type: Research Article
ISSN: 2046-6099

Keywords

1 – 10 of 32