Search results

1 – 10 of over 58000

Abstract

Details

Using Subject Headings for Online Retrieval: Theory, Practice and Potential
Type: Book
ISBN: 978-0-12221-570-4

Abstract

Details

Using Subject Headings for Online Retrieval: Theory, Practice and Potential
Type: Book
ISBN: 978-0-12221-570-4

Article
Publication date: 5 June 2009

Bojana Dimić and Dušan Surla

The purpose of this paper is to model and implement an extensible markup language (XML)‐based editor for library cataloguing. The editor model should support data input in the…

1871

Abstract

Purpose

The purpose of this paper is to model and implement an extensible markup language (XML)‐based editor for library cataloguing. The editor model should support data input in the form of free text with interactive control of structure and content validity of records specified in the UNIMARC and MARC 21 formats. The editor is implemented in the Java programming language in the form of a software package.

Design/methodology/approach

The unified modelling language (UML 2.0) is used for the specification of both the information requirements and the model architecture. The object oriented methodology is used for design and implementation of the software packages, as well as the corresponding CASE tools.

Findings

The result is an editor for UNIMARC and MARC 21 cataloguing. The editor is based on the XML technologies by which the two basic characteristics are achieved as follows: a possibility of integrating the editor into different library software systems and, moving to another format requires only the changes of the module for bibliographic record data control.

Research limitations/implications

A basic limitation of the system is related to the subsystem that controls validation of the bibliographic records and its expansion for work with other bibliographic formats. In the proposed solution, a part of the control of data input is included into the implementation itself and it is related to the UNIMARC format. That is, a part of data by which the control is done, such as repeatability of the record elements and the codebooks, is contained in the XML document of the format that is input information in the editor. However, the control that is related to validation of the format of content in record elements cannot be performed for any other format without modification in the implementation. Therefore, the research could be continued by considering the separation of data used for content control as input information for the application. In that way, this segment would also become implementation independent. One of the solutions should be extending the XML document of the format by this data. Some other solution should mean creating a totally separate system for the content validation. Moreover, the proposed editor supports processing of a bibliographic record only in the UNIMARC and MARC 21 formats. Processing of records in other formats requires considerable changes in the model.

Practical implications

The model of a new editor is developed on the basis of the experience and needs of electronic management in city and special libraries. Based on the given model a new editor is implemented and integrated into the BISIS software system used by the mentioned libraries. Testing and verification are performed on the bibliographic records of the public city libraries.

Originality/value

The contribution of this work is in the system architecture that is based on the XML documents and is independent of the bibliographic format. The XML document that contains data about the bibliographic format represents the editor input information. After a bibliographic record is created in this editor, the record is stored into an XML document that represents the editor output information. This XML document can be stored into various software systems for data storage and retrieval.

Details

The Electronic Library, vol. 27 no. 3
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 11 January 2019

Myongho Yi

The use of “open data” can help the public find value in various areas of interests. Many governments have created and published a huge amount of open data; however, people have a…

1043

Abstract

Purpose

The use of “open data” can help the public find value in various areas of interests. Many governments have created and published a huge amount of open data; however, people have a hard time using open data because of data quality issues. The UK, the USA and Korea have created and published open data; however, the rate of open data implementation and level of open data impact is very low because of data quality issues like incompatible data formats and incomplete data. This study aims to compare the statuses of data quality from open government sites in the UK, the USA and Korea and also present guidelines for publishing data format and enhancing data completeness.

Design/methodology/approach

This study uses statistical analysis of different data formats and examination of data completeness to explore key issues of data quality in open government data.

Findings

Findings show that the USA and the UK have published more than 50 per cent of open data in level one. Korea has published 52.8 per cent of data in level three. Level one data are not machine-readable; therefore, users have a hard time using them. The level one data are found in portable document format and hyper text markup language (HTML) and are locked up in documents; therefore, machines cannot extract out the data. Findings show that incomplete data are existing in all three governments’ open data.

Originality/value

Governments should investigate data incompleteness of all open data and correct incomplete data of the most used data. Governments can find the most used data easily by monitoring data sets that have been downloaded most frequently over a certain period.

Details

The Electronic Library, vol. 37 no. 1
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 15 February 2011

Dragan Ivanović, Dušan Surla and Zora Konjović

The purpose of this research is to observe all data from the Common European Research Information Format (CERIF) data model that can be described using bibliographic standards and…

Abstract

Purpose

The purpose of this research is to observe all data from the Common European Research Information Format (CERIF) data model that can be described using bibliographic standards and move those data to a data model of bibliographic standard.

Design/methodology/approach

Analysis of the CERIF data model and the MARC 21 format has shown that some elements of the CERIF data model could be mapped to the MARC 21 bibliographic record. A CERIF compatible data model based on the MARC 21 format is proposed. The data model was created using PowerDesigner CASE tool. The proposed data model is represented using a physical data model in the conceptual notation that is adopted in the literature for representing the CERIF data model.

Findings

A CERIF compatible data model based on the MARC 21 format is proposed. The proposed model contains all the data from the CERIF2008 data model. The central part of the proposed model is MARC 21 data model that is used as a replacement for 27 entities of the CERIF data model, including all their attributes as well as part of the attributes in entities related to organisational unit. The mappings between attributes of entities of the CERIF data model and the data model of the MARC 21 format are described.

Research limitations/implications

The CERIF compatible data model based on the MARC 21 format does not support all restrictions on data types, which are defined by the CERIF data model. This means that such restrictions have to be controlled by software.

Practical implications

The central part of the proposed CERIF compatible data model is a data model of MARC 21 format. It means that most of the data are modelled according to bibliographic standard, which is very widespread worldwide. This implies that the proposed CERIF model can be easily implemented within the existing library infrastructure. In addition, the proposed model can be used for other purposes, such as the evaluation of scientific research results, generating bibliographies of researchers, and institutions, the citations etc. A research management system based on the proposed model is implemented. Also, this system is verified and tested on data about published results of researchers employed at University of Novi Sad, Serbia.

Originality/value

A new data model compatible with the CERIF data model is proposed. The basic idea is to map part of the CERIF data model related to published results of scientific research to some well‐known bibliographic standard. It was shown that this part of the data model could be mapped to the MARC 21 data model. It can be mapped to data models of any other MARC standards in a similar way.

Details

The Electronic Library, vol. 29 no. 1
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 3 August 2012

Lidija Ivanović, Dragan Ivanović and Dušan Surla

The aim of this research is to define a data model of theses and dissertations that enables data exchange with CERIF‐compatible CRIS systems and data exchange according to OAI‐PMH…

Abstract

Purpose

The aim of this research is to define a data model of theses and dissertations that enables data exchange with CERIF‐compatible CRIS systems and data exchange according to OAI‐PMH protocol in different metadata formats (Dublin Core, EDT‐MS, etc.).

Design/methodology/approach

Various systems that contain metadata about theses and dissertations are analyzed. There are different standards and protocols that enable the interoperability of those systems: CERIF standard, AOI‐PMH protocol, etc. A physical data model that enables interoperability with almost all of those systems is created using the PowerDesigner CASE tool.

Findings

A set of metadata about theses and dissertations that contain all the metadata required by CERIF data model, Dublin Core format, EDT‐MS format and all the metadata prescribed by the University of Novi Sad is defined. Defined metadata can be stored in the CERIF‐compatible data model based on the MARC21 format.

Practical implications

CRIS‐UNS is a CRIS which has been developed at the University of Novi Sad since 2008. The system is based on the proposed data model, which enables the system's interoperability with other CERIF‐compatible CRIS systems. Also, the system based on the proposed model can become a member of NDLTD.

Social implications

A system based on the proposed model increases the availability of theses and dissertations, and thus encourages the development of the knowledge‐based society.

Originality/value

A data model of theses and dissertations that enables interoperability with CERIF‐compatible CRIS systems is proposed. A software system based on the proposed model could become a member of NDLTD and exchange metadata with institutional repositories. The proposed model increases the availability of theses and dissertations.

Details

Online Information Review, vol. 36 no. 4
Type: Research Article
ISSN: 1468-4527

Keywords

Article
Publication date: 1 June 1996

A. Dolenc and I. Mäkelä

Overviews the current data exchange formats in use today in rapid prototyping (RP), and explores their role in connection with data preparation. Tries to explain the interplay…

611

Abstract

Overviews the current data exchange formats in use today in rapid prototyping (RP), and explores their role in connection with data preparation. Tries to explain the interplay between data exchange formats, application programs, and final functionality which these programs should deliver to a user of RP technology. RP is closely related to the effective use of CAD and related systems, and can demonstrate how geometric models can be transferred from one system to another, and how they are converted from one representation to another. Argues that the functionality that can be implemented and offered to a user in CAD‐related systems is closely related to how geometric models are represented. Since data exchange formats are the vehicles used to represent geometric models for data transfer purposes, they are important building blocks for the effective use of RP technologies. Overviews data exchange formats used in RP applications and discusses their role in developing useful software tools for RP applications.

Details

Rapid Prototyping Journal, vol. 2 no. 2
Type: Research Article
ISSN: 1355-2546

Keywords

Article
Publication date: 1 February 1994

Peter Simmons

Librarians in countries without well‐established national bibliographic systems increasingly find themselves faced with the problem of establishing local formats for…

Abstract

Librarians in countries without well‐established national bibliographic systems increasingly find themselves faced with the problem of establishing local formats for machine‐readable cataloguing and for referral data. Often they lack the background and the resources — especially trained staff — either to adopt an existing MARC format or to develop their own. Such international formats as UNIMARC and CCF (Common Communication Format), despite widespread international use, present problems of their own: MARC formats are not practical for agencies that do not follow standard cataloguing rules, and CCF offers little guidance to agencies wishing to adapt it for local use. A number of techniques useful in adapting and implementing international and national standard formats are presented, with some guidelines for preserving compatibility with standards.

Details

Program, vol. 28 no. 2
Type: Research Article
ISSN: 0033-0337

Article
Publication date: 13 April 2010

Bojana Dimić, Branko Milosavljević and Dušan Surla

The purpose of this paper is to create a model for an XML document that will carry information about bibliographic formats. The model will be given in the form of an XML schema…

1245

Abstract

Purpose

The purpose of this paper is to create a model for an XML document that will carry information about bibliographic formats. The model will be given in the form of an XML schema describing two bibliographic formats, UNIMARC and MARC 21.

Design/methodology/approach

The description of bibliographic formats using the XML schema language may be discussed in two ways. The first one relates to creating an XML schema in a way that all elements of the bibliographic format are described separately. The second way, used in this paper, is creating an XML schema as a set of elements that presents concepts of bibliographic formats. A schema created in the second way is appropriate for use in implementation of cataloguing software.

Findings

The result is an XML schema that describes MARC 21 and UNIMARC formats. The instance of that schema is an XML document describing a bibliographic format that will be used in software systems for cataloguing. An XML document that is an instance of the proposed XML schema is applied in the development of the editor for cataloguing in the BISIS library information system. This XML document represents input information for that editor. In this way, the implementation of the editor becomes independent of the bibliographic format.

Practical implications

The created XML schema cannot serve as an electronic manual because there is some information about the format that is not included in it. In order to overcome this shortcoming an additional XML schema that will contain remaining format data may be provided.

Originality/value

The originality lies in the idea of creating one XML schema for two bibliographic formats. The schema contains elements that are models for data used in cataloguing tools. On the basis of that XML schema, the object model of bibliographic formats is implemented as well as software component for manipulating format data. This component can be used in development of library software systems.

Details

The Electronic Library, vol. 28 no. 2
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 31 December 2006

Hooman Homayounfar and Fangju Wang

XML is becoming one of the most important structures for data exchange on the web. Despite having many advantages, XML structure imposes several major obstacles to large document…

Abstract

XML is becoming one of the most important structures for data exchange on the web. Despite having many advantages, XML structure imposes several major obstacles to large document processing. Inconsistency between the linear nature of the current algorithms (e.g. for caching and prefetch) used in operating systems and databases, and the non‐linear structure of XML data makes XML processing more costly. In addition to verbosity (e.g. tag redundancy), interpreting (i.e. parsing) depthfirst (DF) structure of XML documents is a significant overhead to processing applications (e.g. query engines). Recent research on XML query processing has learned that sibling clustering can improve performance significantly. However, the existing clustering methods are not able to avoid parsing overhead as they are limited by larger document sizes. In this research, We have developed a better data organization for native XML databases, named sibling‐first (SF) format that improves query performance significantly. SF uses an embedded index for fast accessing to child nodes. It also compresses documents by eliminating extra information from the original DF format. The converted SF documents can be processed for XPath query purposes without being parsed. We have implemented the SF storage in virtual memory as well as a format on disk. Experimental results with real data have showed that significantly higher performance can be achieved when XPath queries are conducted on very large SF documents.

Details

International Journal of Web Information Systems, vol. 2 no. 3/4
Type: Research Article
ISSN: 1744-0084

Keywords

1 – 10 of over 58000