Search results

11 – 20 of over 27000
Article
Publication date: 1 June 1992

E.G. Sieverts, M. Hofstede, G. Lobbestael, B. Oude Groeniger, F. Provost and P. Šikovà

In this article, the fifth in a series on microcomputer software for information storage and retrieval, test results of seven programs are presented and various properties and…

Abstract

In this article, the fifth in a series on microcomputer software for information storage and retrieval, test results of seven programs are presented and various properties and qualities of these programs are discussed. In this instalment of the series we discuss programs for information storage and retrieval which are primarily characterised by the properties of personal information managers (PIMs), hypertext programs, or best match and ranking retrieval systems. The programs reviewed in this issue are the personal information managers 3by5/RediReference, askSam, Dayflo Tracker, and Ize; Personal Librarian uses best match and ranking; the hypertext programs are Folio Views and the HyperKRS/HyperCard combination (askSam, Ize and Personal Librarian boast hypertext features as well). HyperKRS/HyperCard is only available for the Apple Macintosh. All other programs run under MS‐DOS; versions of Personal Librarian also run under Windows and some other systems. For each of the seven programs about 100 facts and test results are tabulated. The programs are also discussed individually.

Details

The Electronic Library, vol. 10 no. 6
Type: Research Article
ISSN: 0264-0473

Book part
Publication date: 10 February 2012

Kin Fun Li, Yali Wang and Wei Yu

Purpose — To develop methodologies to evaluate search engines according to an individual's preference in an easy and reliable manner, and to formulate user-oriented metrics to…

Abstract

Purpose — To develop methodologies to evaluate search engines according to an individual's preference in an easy and reliable manner, and to formulate user-oriented metrics to compare freshness and duplication in search results.

Design/methodology/approach — A personalised evaluation model for comparing search engines is designed as a hierarchy of weighted parameters. These commonly found search engine features and performance measures are given quantitative and qualitative ratings by an individual user. Furthermore, three performance measurement metrics are formulated and presented as histograms for visual inspection. A methodology is introduced to quantitatively compare and recognise the different histogram patterns within the context of search engine performance.

Findings — Precision and recall are the fundamental measures used in many search engine evaluations due to their simplicity, fairness and reliability. Most recent evaluation models are user oriented and focus on relevance issues. Identifiable statistical patterns are found in performance measures of search engines.

Research limitations/implications — The specific parameters used in the evaluation model could be further refined. A larger scale user study would confirm the validity and usefulness of the model. The three performance measures presented give a reasonably informative overview of the characteristics of a search engine. However, additional performance parameters and their resulting statistical patterns would make the methodology more valuable to the users.

Practical implications — The easy-to-use personalised search engine evaluation model can be tailored to an individual's preference and needs simply by changing the weights and modifying the features considered. A user is able to get an idea of the characteristics of a search engine quickly using the quantitative measure of histogram patterns that represent the search performance metrics introduced.

Originality/value — The presented work is considered original as one of the first search engine evaluation models that can be personalised. This enables a Web searcher to choose an appropriate search engine for his/her needs and hence finding the right information in the shortest time with the least effort.

Article
Publication date: 29 December 2022

Kianoosh Rashidi, Hajar Sotudeh and Alireza Nikseresht

This study aimed to investigate how the enrichment of medical documents' index terms by their comments improves the relevance and novelty of the top-ranked results retrieved by an…

Abstract

Purpose

This study aimed to investigate how the enrichment of medical documents' index terms by their comments improves the relevance and novelty of the top-ranked results retrieved by an NLP system.

Design/methodology/approach

A semi-experimental pre-test and post-test research was designed to compare NLP-based indexes before and after being expanded by the comment terms. The experiments were conducted on a test collection of 13,957 documents commented by F1000-Prime reviewers. They were indexed at title, abstract, body and full-text levels. In total, 100 seed documents were randomly selected and served as queries. The textual similarity of the documents and queries was calculated using Lucene-more-like-this function and evaluated by the semantic similarity of their MeSH. The results novelty was measured using maximal marginal relevance and evaluated by their MeSH novelties. Normalized discounted cumulative gain was used to compare the basic and expanded indexes' precisions at 10, 20 and 50 top ranks.

Findings

The relevance and novelty of the results ranked at the top precision points was improved after expanding the indexes by the comment terms. The finding implies that meta-texts are effective in representing their mother documents, by adding dynamic elements to their rather static contents. It also provides further evidence about the merits of the application of social intelligence and collective wisdom reflected in the actions and reactions of users in tackling the challenges faced by NLP-based systems.

Originality/value

This is the first study to confirm that social comments on scientific papers improve the performance of information systems in terms of relevance and novelty.

Peer review

The peer review history for this article is available at: https://publons.com/publon/10.1108/OIR-05-2022-0283.

Details

Online Information Review, vol. 47 no. 6
Type: Research Article
ISSN: 1468-4527

Keywords

Article
Publication date: 2 January 2024

Raunaque Mujeeb Quaiser and Praveen Ranjan Srivastava

This research aims to identify the key factors affecting Outbound Open Innovation between Startups and Big organizations using the multiple criteria decision-making analysis…

Abstract

Purpose

This research aims to identify the key factors affecting Outbound Open Innovation between Startups and Big organizations using the multiple criteria decision-making analysis (MCDM) approach. The MCDM technique ranks the four key factors identified from the literature study that can help to improve collaboration opportunities with Startups.

Design/methodology/approach

Identification of key factors affecting Outbound Open Innovation between Startups and big organizations based on extant literature. A questionnaire is prepared based on these four identified key factors to gather views of the startup's employees, from the designer level to the startup's founder. MCDM techniques are used to evaluate the questionnaire. The ensemble technique is used to rank the key factors coming from three different MCDM methods.

Findings

The findings from the MCDM approach and Ensemble techniques give insight to the big organizations to facilitate outbound Open Innovation effectively. It also provides insight into the requirements of the startups and the kind of support they seek from the big organizations. The ranking can help the big organization close the gaps and make an informed decision to increase the effectiveness of the collaborations and boost innovation.

Originality/value

This is a unique research work where the MCDM approach is used to identify the ranking of key factors affecting outbound open innovation between startups and big organizations. The MCDM technique is followed by the ensemble method to rationalize the findings. Technology Relevance ranks highest, followed by Innovation Ecosystem, Organization commitment and Knowledge Sharing.

Details

Management Decision, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0025-1747

Keywords

Article
Publication date: 19 April 2011

Shihchieh Chou, Chinyi Cheng and Szujui Huang

The purpose of this paper is to establish a new approach for solving the expansion term problem.

Abstract

Purpose

The purpose of this paper is to establish a new approach for solving the expansion term problem.

Design/methodology/approach

This study develops an expansion term weighting function derived from the valuable concepts used by previous approaches. These concepts include probability measurement, adjustment according to situations, and summation of weights. Formal tests have been conducted to compare the proposed weighting function with the baseline ranking model and other weighting functions.

Findings

The results reveal stable performance by the proposed expansion term weighting function. It proves more effective than the baseline ranking model and outperforms other weighting functions.

Research limitations/implications

The paper finds that testing additional data sets and potential applications to real working situations is required before the generalisability and superiority of the proposed expansion term weighting function can be asserted.

Originality/value

Stable performance and an acceptable level of effectiveness for the proposed expansion term weighting function indicate the potential for further study and development of this approach. This would add to the current methods studied by the information retrieval community for culling information from documents.

Details

Online Information Review, vol. 35 no. 2
Type: Research Article
ISSN: 1468-4527

Keywords

Article
Publication date: 2 August 2013

Lourdes Moreno and Paloma Martinez

The purpose of this paper is to show that the pursuit of a high search engine relevance ranking for a webpage is not necessarily incompatible with the pursuit of web accessibility.

3768

Abstract

Purpose

The purpose of this paper is to show that the pursuit of a high search engine relevance ranking for a webpage is not necessarily incompatible with the pursuit of web accessibility.

Design/methodology/approach

The research described arose from an investigation into the observed phenomenon that pages from accessible websites regularly appear near the top of search engine (such as Google) results, without any deliberate effort having been made through the application of search engine optimization (SEO) techniques to achieve this. The reasons for this phenomenon appear to be found in the numerous similarities and overlapping characteristics between SEO factors and web accessibility guidelines. Context is provided through a review of sources including accessibility standards and relevant SEO studies and the relationship between SEO and web accessibility is described. The particular overlapping factors between the two are identified and the precise nature of the overlaps is explained in greater detail.

Findings

The available literature provides firm evidence that the overlapping factors not only serve to ensure the accessibility of a website for all users, but are also useful for the optimization of the website's search engine ranking. The research demonstrates that any SEO project undertaken should include, as a prerequisite, the proper design of accessible web content, inasmuch as search engines will interpret the web accessibility achieved as an indicator of quality and will be able to better access and index the resulting web content.

Originality/value

The present study indicates how developing websites with high visibility in search engine results also makes their content more accessible.

Details

Online Information Review, vol. 37 no. 4
Type: Research Article
ISSN: 1468-4527

Keywords

Article
Publication date: 1 January 1996

PETER INGWERSEN

The objective of the paper is to amalgamate theories of text retrieval from various research traditions into a cognitive theory for information retrieval interaction. Set in a…

2458

Abstract

The objective of the paper is to amalgamate theories of text retrieval from various research traditions into a cognitive theory for information retrieval interaction. Set in a cognitive framework, the paper outlines the concept of polyrepresentation applied to both the user's cognitive space and the information space of IR systems. The concept seeks to represent the current user's information need, problem state, and domain work task or interest in a structure of causality. Further, it implies that we should apply different methods of representation and a variety of IR techniques of different cognitive and functional origin simultaneously to each semantic full‐text entity in the information space. The cognitive differences imply that by applying cognitive overlaps of information objects, originating from different interpretations of such objects through time and by type, the degree of uncertainty inherent in IR is decreased. Polyrepresentation and the use of cognitive overlaps are associated with, but not identical to, data fusion in IR. By explicitly incorporating all the cognitive structures participating in the interactive communication processes during IR, the cognitive theory provides a comprehensive view of these processes. It encompasses the ad hoc theories of text retrieval and IR techniques hitherto developed in mainstream retrieval research. It has elements in common with van Rijsbergen and Lalmas' logical uncertainty theory and may be regarded as compatible with that conception of IR. Epistemologically speaking, the theory views IR interaction as processes of cognition, potentially occurring in all the information processing components of IR, that may be applied, in particular, to the user in a situational context. The theory draws upon basic empirical results from information seeking investigations in the operational online environment, and from mainstream IR research on partial matching techniques and relevance feedback. By viewing users, source systems, intermediary mechanisms and information in a global context, the cognitive perspective attempts a comprehensive understanding of essential IR phenomena and concepts, such as the nature of information needs, cognitive inconsistency and retrieval overlaps, logical uncertainty, the concept of ‘document’, relevance measures and experimental settings. An inescapable consequence of this approach is to rely more on sociological and psychological investigative methods when evaluating systems and to view relevance in IR as situational, relative, partial, differentiated and non‐linear. The lack of consistency among authors, indexers, evaluators or users is of an identical cognitive nature. It is unavoidable, and indeed favourable to IR. In particular, for full‐text retrieval, alternative semantic entities, including Salton et al.'s ‘passage retrieval’, are proposed to replace the traditional document record as the basic retrieval entity. These empirically observed phenomena of inconsistency and of semantic entities and values associated with data interpretation support strongly a cognitive approach to IR and the logical use of polyrepresentation, cognitive overlaps, and both data fusion and data diffusion.

Details

Journal of Documentation, vol. 52 no. 1
Type: Research Article
ISSN: 0022-0418

Article
Publication date: 14 May 2018

Sholeh Arastoopoor

The degree to which a text is considered readable depends on the capability of the reader. This assumption puts different information retrieval systems at the risk of retrieving…

Abstract

Purpose

The degree to which a text is considered readable depends on the capability of the reader. This assumption puts different information retrieval systems at the risk of retrieving unreadable or hard-to-be-read yet relevant documents for their users. This paper aims to examine the potential use of concept-based readability measures along with classic measures for re-ranking search results in information retrieval systems, specifically in the Persian language.

Design/methodology/approach

Flesch–Dayani as a classic readability measure along with document scope (DS) and document cohesion (DC) as domain-specific measures have been applied for scoring the retrieved documents from Google (181 documents) and the RICeST database (215 documents) in the field of computer science and information technology (IT). The re-ranked result has been compared with the ranking of potential users regarding their readability.

Findings

The results show that there is a difference among subcategories of the computer science and IT field according to their readability and understandability. This study also shows that it is possible to develop a hybrid score based on DS and DC measures and, among all four applied scores in re-ranking the documents, the re-ranked list of documents based on the DSDC score shows correlation with re-ranking of the participants in both groups.

Practical implications

The findings of this study would foster a new option in re-ranking search results based on their difficulty for experts and non-experts in different fields.

Originality/value

The findings and the two-mode re-ranking model proposed in this paper along with its primary focus on domain-specific readability in the Persian language would help Web search engines and online databases in further refining the search results in pursuit of retrieving useful texts for users with differing expertise.

Article
Publication date: 1 February 1997

Xiaoying Dong and Louise T. Su

The World Wide Web's search engines are the main tools for indexing and retrieval of Internet resources today. Comparison and evaluation of their performance is of great…

3077

Abstract

The World Wide Web's search engines are the main tools for indexing and retrieval of Internet resources today. Comparison and evaluation of their performance is of great importance for system developers and information professionals, as well as end‐users, for the improvement and development of better tools. The paper describes categories and special features of Web‐based databases and compares them with traditional databases. It then presents a review of the literature on the testing and evaluation of Web‐based search engines. Different methodologies and measures used in previous studies are described and their findings are summarised. The paper presents some evaluative comments on previous studies and suggests areas for future investigation, particularly evaluation of Web‐based search engines from the end‐user's perspective.

Details

Online and CD-Rom Review, vol. 21 no. 2
Type: Research Article
ISSN: 1353-2642

Article
Publication date: 28 October 2020

Adamu Garba, Shah Khalid, Irfan Ullah, Shah Khusro and Diyawu Mumin

There have been many challenges in crawling deep web by search engines due to their proprietary nature or dynamic content. Distributed Information Retrieval (DIR) tries to solve…

Abstract

Purpose

There have been many challenges in crawling deep web by search engines due to their proprietary nature or dynamic content. Distributed Information Retrieval (DIR) tries to solve these problems by providing a unified searchable interface to these databases. Since a DIR must search across many databases, selecting a specific database to search against the user query is challenging. The challenge can be solved if the past queries of the users are considered in selecting collections to search in combination with word embedding techniques. Combining these would aid the best performing collection selection method to speed up retrieval performance of DIR solutions.

Design/methodology/approach

The authors propose a collection selection model based on word embedding using Word2Vec approach that learns the similarity between the current and past queries. They used the cosine and transformed cosine similarity models in computing the similarities among queries. The experiment is conducted using three standard TREC testbeds created for federated search.

Findings

The results show significant improvements over the baseline models.

Originality/value

Although the lexical matching models for collection selection using similarity based on past queries exist, to the best our knowledge, the proposed work is the first of its kind that uses word embedding for collection selection by learning from past queries.

Details

Data Technologies and Applications, vol. 54 no. 5
Type: Research Article
ISSN: 2514-9288

Keywords

11 – 20 of over 27000