Search results

1 – 10 of over 21000
Book part
Publication date: 18 January 2023

Shane W. Reid, Aaron F. McKenny and Jeremy C. Short

A growing body of research outlines how to best facilitate and ensure methodological rigor when using dictionary-based computerized text analyses (DBCTA) in organizational…

Abstract

A growing body of research outlines how to best facilitate and ensure methodological rigor when using dictionary-based computerized text analyses (DBCTA) in organizational research. However, these best practices are currently scattered across several methodological and empirical manuscripts, making it difficult for scholars new to the technique to implement DBCTA in their own research. To better equip researchers looking to leverage this technique, this methodological report consolidates current best practices for applying DBCTA into a single, practical guide. In doing so, we provide direction regarding how to make key design decisions and identify valuable resources to help researchers from the beginning of the research process through final publication. Consequently, we advance DBCTA methods research by providing a one-stop reference for novices and experts alike concerning current best practices and available resources.

Article
Publication date: 1 January 1980

D. ANDREW ROBERTS and RICHARD B. LIGHT

A survey of the current state of documentation practice in museums is presented. This concentrates on the broad themes of the practice, making comparisons with analogous library…

Abstract

A survey of the current state of documentation practice in museums is presented. This concentrates on the broad themes of the practice, making comparisons with analogous library procedures, where appropriate. A brief introduction to museums and their organizational framework within the United Kingdom is given. With this as background, the methods of documentation used by museums are reviewed, and a survey presented of current developments on an international and national scale.

Details

Journal of Documentation, vol. 36 no. 1
Type: Research Article
ISSN: 0022-0418

Article
Publication date: 1 February 1993

BRIAN VICKERY and ALINA VICKERY

There is a huge amount of information and data stored in publicly available online databases that consist of large text files accessed by Boolean search techniques. It is widely…

Abstract

There is a huge amount of information and data stored in publicly available online databases that consist of large text files accessed by Boolean search techniques. It is widely held that less use is made of these databases than could or should be the case, and that one reason for this is that potential users find it difficult to identify which databases to search, to use the various command languages of the hosts and to construct the Boolean search statements required. This reasoning has stimulated a considerable amount of exploration and development work on the construction of search interfaces, to aid the inexperienced user to gain effective access to these databases. The aim of our paper is to review aspects of the design of such interfaces: to indicate the requirements that must be met if maximum aid is to be offered to the inexperienced searcher; to spell out the knowledge that must be incorporated in an interface if such aid is to be given; to describe some of the solutions that have been implemented in experimental and operational interfaces; and to discuss some of the problems encountered. The paper closes with an extensive bibliography of references relevant to online search aids, going well beyond the items explicitly mentioned in the text. An index to software appears after the bibliography at the end of the paper.

Details

Journal of Documentation, vol. 49 no. 2
Type: Research Article
ISSN: 0022-0418

Article
Publication date: 1 May 2006

Carmen Galvez and Félix de Moya‐Anegón

To evaluate the accuracy of conflation methods based on finite‐state transducers (FSTs).

Abstract

Purpose

To evaluate the accuracy of conflation methods based on finite‐state transducers (FSTs).

Design/methodology/approach

Incorrectly lemmatized and stemmed forms may lead to the retrieval of inappropriate documents. Experimental studies to date have focused on retrieval performance, but very few on conflation performance. The process of normalization we used involved a linguistic toolbox that allowed us to construct, through graphic interfaces, electronic dictionaries represented internally by FSTs. The lexical resources developed were applied to a Spanish test corpus for merging term variants in canonical lemmatized forms. Conflation performance was evaluated in terms of an adaptation of recall and precision measures, based on accuracy and coverage, not actual retrieval. The results were compared with those obtained using a Spanish version of the Porter algorithm.

Findings

The conclusion is that the main strength of lemmatization is its accuracy, whereas its main limitation is the underanalysis of variant forms.

Originality/value

The report outlines the potential of transducers in their application to normalization processes.

Details

Journal of Documentation, vol. 62 no. 3
Type: Research Article
ISSN: 0022-0418

Keywords

Article
Publication date: 1 August 2005

Carmen Galvez, Félix de Moya‐Anegón and Víctor H. Solana

To propose a categorization of the different conflation procedures at the two basic approaches, non‐linguistic and linguistic techniques, and to justify the application of…

1323

Abstract

Purpose

To propose a categorization of the different conflation procedures at the two basic approaches, non‐linguistic and linguistic techniques, and to justify the application of normalization methods within the framework of linguistic techniques.

Design/methodology/approach

Presents a range of term conflation methods, that can be used in information retrieval. The uniterm and multiterm variants can be considered equivalent units for the purposes of automatic indexing. Stemming algorithms, segmentation rules, association measures and clustering techniques are well evaluated non‐linguistic methods, and experiments with these techniques show a wide variety of results. Alternatively, the lemmatisation and the use of syntactic pattern‐matching, through equivalence relations represented in finite‐state transducers (FST), are emerging methods for the recognition and standardization of terms.

Findings

The survey attempts to point out the positive and negative effects of the linguistic approach and its potential as a term conflation method.

Originality/value

Outlines the importance of FSTs for the normalization of term variants.

Details

Journal of Documentation, vol. 61 no. 4
Type: Research Article
ISSN: 0022-0418

Keywords

Article
Publication date: 1 March 1998

Alexander M. Robertson and Peter Willett

This paper provides an introduction to the use of n‐grams in textual information systems, where an n‐gram is a string of n, usually adjacent, characters extracted from a section…

Abstract

This paper provides an introduction to the use of n‐grams in textual information systems, where an n‐gram is a string of n, usually adjacent, characters extracted from a section of continuous text. Applications that can be implemented efficiently and effectively using sets of n‐grams include spelling error detection and correction, query expansion, information retrieval with serial, inverted and signature files, dictionary look‐up, text compression, and language identification.

Details

Journal of Documentation, vol. 54 no. 1
Type: Research Article
ISSN: 0022-0418

Keywords

Article
Publication date: 1 July 2006

Peter Willett

In 1980, Porter presented a simple algorithm for stemming English language words. This paper summarises the main features of the algorithm, and highlights its role not just in…

3587

Abstract

Purpose

In 1980, Porter presented a simple algorithm for stemming English language words. This paper summarises the main features of the algorithm, and highlights its role not just in modern information retrieval research, but also in a range of related subject domains.

Design/methodology/approach

Review of literature and research involving use of the Porter algorithm.

Findings

The algorithm has been widely adopted and extended so that it has become the standard approach to word conflation for information retrieval in a wide range of languages.

Orinality/value

The 1980 paper in Program by Porter describing his algorithm has been highly cited. This paper provides a context for the original paper as well as an overview of its subsequent use.

Details

Program, vol. 40 no. 3
Type: Research Article
ISSN: 0033-0337

Keywords

Article
Publication date: 1 January 1984

Günther Haedrich and Dipl.‐Kfm. Edgar Kreilkamp

Einführung in das Portfolio‐Management Allgemein ist in den letzten 5–10 Jahren zunehmendes Interesse der Wirtschaft und Verwaltung für Probleme der strategischen Planung…

Abstract

Einführung in das Portfolio‐Management Allgemein ist in den letzten 5–10 Jahren zunehmendes Interesse der Wirtschaft und Verwaltung für Probleme der strategischen Planung festzustellen, nicht zuletzt aufgrund der Diskussion über das Portfolio‐Management. Dieses neue Planungsmodell rückt durch die Definition sogenannter Strategischer Geschäftsfelder (“business units”, vgl. Abschnitt 2 dieses Aufsatzes) als strategische Planung aus gesamtunternehmerischer Sicht in den Mittelpunkt der Betrachtung und löst sich damit los von der engen Funktionalorientierung, bei der jeder strategische Bereich als eigene und relativ unabhängige Planungseinheit auftritt. Abell/Hammond sprechen in diesem Zusammenhang von einem 3‐zyklischen Planungsprozess (Abell/Hammond 1979, S. 14) (vgl. Abb. 1); die Längeund Breite der einzelnen Planungsstufen geben gleichzeitig ihre zeitliche Dauer und ihre relative Bedeutung wieder.

Details

The Tourist Review, vol. 39 no. 1
Type: Research Article
ISSN: 0251-3102

Article
Publication date: 1 January 1988

Martin Porter and Valerie Galpin

This paper reports on the successful introduction of a sophisticated online catalogue system at the library of the Scott Polar Research Institute, Cambridge, using the Muscat…

Abstract

This paper reports on the successful introduction of a sophisticated online catalogue system at the library of the Scott Polar Research Institute, Cambridge, using the Muscat program package. The system provides to both end‐users and library staff a choice between boolean searching on keywords and access using relevance feedback based on free text in English, mixed with UDC classification numbers. The system is implemented on an IBM 3084 computer. Significant benefits from the application of relevance feedback are reported with 10,000 records on file.

Details

Program, vol. 22 no. 1
Type: Research Article
ISSN: 0033-0337

Article
Publication date: 1 January 1990

Alan F. Smeaton

Database management systems (DBMS) and information retrieval (IR) systems can both be used as online information systems but they differ in the type of data and the types of…

Abstract

Database management systems (DBMS) and information retrieval (IR) systems can both be used as online information systems but they differ in the type of data and the types of retrieval they provide for users. Many previous attempts have been made to couple DBMS and IR systems together, either by integrating the two into a unified framework, or by using a DBMS as an implementation tool for information retrieval functionality. This paper reports on some of these previous attempts and describes a system, retriev, which uses a DBMS to implement an IR system for teaching and research purposes. The implementation of retriev is described in detail and the effects that the current trends in database research will have on the relationship between DBMS and IR systems, are discussed.

Details

Program, vol. 24 no. 1
Type: Research Article
ISSN: 0033-0337

1 – 10 of over 21000