Search results

1 – 10 of over 16000
Article
Publication date: 1 January 2005

Diederik Aerts and Liane Gabora

To develop a theory of concepts that solves the combination problem, i.e. to deliver a description of the combination of concepts. We also investigate the so‐called “pet fish…

Abstract

Purpose

To develop a theory of concepts that solves the combination problem, i.e. to deliver a description of the combination of concepts. We also investigate the so‐called “pet fish problem” in concept research.

Design/methodology/approach

The set of contexts and properties of a concept are embedded in the complex Hilbert space of quantum mechanics. States are unit vectors or density operators and context and properties are orthogonal projections.

Findings

The way calculations are done in Hilbert space makes it possible to model how context influences the state of a concept. Moreover, a solution to the combination problem is proposed. Using the tensor product, a natural product in Hilbert space mathematics, a procedure for describing combined concepts is elaborated. This procedure also provides a solution to the pet‐fish problem, and it allows the modeling of an arbitrary number of combined concepts. By way of example, a model for a simple sentence containing a subject, a predicate and an object, is presented.

Originality/value

The combination problem is considered to be one of the crucial unsolved problems in concept research. Also the pet‐fish problem has not been solved by earlier attempts of modeling.

Details

Kybernetes, vol. 34 no. 1/2
Type: Research Article
ISSN: 0368-492X

Keywords

Article
Publication date: 1 January 1979

GERARD SALTON

The development of a given discipline in science and technology often depends on the availability of theories capable of describing the processes which control the field and of…

Abstract

The development of a given discipline in science and technology often depends on the availability of theories capable of describing the processes which control the field and of modelling the interactions between these processes. The absence of an accepted theory of information retrieval has been blamed for the relative disorder and the lack of technical advances in the area. The main mathematical approaches to information retrieval are examined in this study, including both algebraic and probabilistic models, and the difficulties which impede the formalization of information retrieval processes are described. A number of developments are covered where new theoretical understandings have directly led to the improvement of retrieval techniques and operations.

Details

Journal of Documentation, vol. 35 no. 1
Type: Research Article
ISSN: 0022-0418

Article
Publication date: 21 May 2018

Dongmei Han, Wen Wang, Suyuan Luo, Weiguo Fan and Songxin Wang

This paper aims to apply vector space model (VSM)-PCR model to compute the similarity of Fault zone ontology semantics, which verified the feasibility and effectiveness of the…

Abstract

Purpose

This paper aims to apply vector space model (VSM)-PCR model to compute the similarity of Fault zone ontology semantics, which verified the feasibility and effectiveness of the application of VSM-PCR method in uncertainty mapping of ontologies.

Design/methodology/approach

The authors first define the concept of uncertainty ontology and then propose the method of ontology mapping. The proposed method fully considers the properties of ontology in measuring the similarity of concept. It expands the single VSM of concept meaning or instance set to the “meaning, properties, instance” three-dimensional VSM and uses membership degree or correlation to express the level of uncertainty.

Findings

It provides a relatively better accuracy which verified the feasibility and effectiveness of VSM-PCR method in treating the uncertainty mapping of ontology.

Research limitations/implications

The future work will focus on exploring the similarity measure and combinational methods in every dimension.

Originality/value

This paper presents an uncertain mapping method of ontology concept based on three-dimensional combination weighted VSM, namely, VSM-PCR. It expands the single VSM of concept meaning or instance set to the “meaning, properties, instance” three-dimensional VSM. The model uses membership degree or correlation which is used to express the degree of uncertainty; as a result, a three-dimensional VSM is obtained. The authors finally provide an example to verify the feasibility and effectiveness of VSM-PCR method in treating the uncertainty mapping of ontology.

Details

Information Discovery and Delivery, vol. 46 no. 2
Type: Research Article
ISSN: 2398-6247

Keywords

Article
Publication date: 8 July 2022

Chuanming Yu, Zhengang Zhang, Lu An and Gang Li

In recent years, knowledge graph completion has gained increasing research focus and shown significant improvements. However, most existing models only use the structures of…

Abstract

Purpose

In recent years, knowledge graph completion has gained increasing research focus and shown significant improvements. However, most existing models only use the structures of knowledge graph triples when obtaining the entity and relationship representations. In contrast, the integration of the entity description and the knowledge graph network structure has been ignored. This paper aims to investigate how to leverage both the entity description and the network structure to enhance the knowledge graph completion with a high generalization ability among different datasets.

Design/methodology/approach

The authors propose an entity-description augmented knowledge graph completion model (EDA-KGC), which incorporates the entity description and network structure. It consists of three modules, i.e. representation initialization, deep interaction and reasoning. The representation initialization module utilizes entity descriptions to obtain the pre-trained representation of entities. The deep interaction module acquires the features of the deep interaction between entities and relationships. The reasoning component performs matrix manipulations with the deep interaction feature vector and entity representation matrix, thus obtaining the probability distribution of target entities. The authors conduct intensive experiments on the FB15K, WN18, FB15K-237 and WN18RR data sets to validate the effect of the proposed model.

Findings

The experiments demonstrate that the proposed model outperforms the traditional structure-based knowledge graph completion model and the entity-description-enhanced knowledge graph completion model. The experiments also suggest that the model has greater feasibility in different scenarios such as sparse data, dynamic entities and limited training epochs. The study shows that the integration of entity description and network structure can significantly increase the effect of the knowledge graph completion task.

Originality/value

The research has a significant reference for completing the missing information in the knowledge graph and improving the application effect of the knowledge graph in information retrieval, question answering and other fields.

Details

Aslib Journal of Information Management, vol. 75 no. 3
Type: Research Article
ISSN: 2050-3806

Keywords

Article
Publication date: 17 April 2020

Barkha Bansal and Sangeet Srivastava

Aspect based sentiment classification is valuable for providing deeper insight into online consumer reviews (OCR). However, the majority of the previous studies explicitly…

Abstract

Purpose

Aspect based sentiment classification is valuable for providing deeper insight into online consumer reviews (OCR). However, the majority of the previous studies explicitly determine the orientation of aspect related sentiment bearing word and overlook the aspect-context. Therefore, this paper aims to propose an aspect-context aware sentiment classification of OCR for deeper and more accurate insights.

Design/methodology/approach

In the proposed methodology, first, aspect descriptions and sentiment bearing words are extracted. Then, the skip-gram model is used to extract the first set of features to capture contextual information. For the second category of features, cosine similarity is used between a pre-defined seed word list and aspects, to capture aspect context sensitive sentiments. The third set of features includes weighted word vectors using term frequency-inverse document frequency. After concatenating features, ensemble classifier is used using three base classifiers.

Findings

Experimental results on two real-world data sets with variable lengths, acquired from Amazon.com and TripAdvisor.com, show that the advised ensemble approach significantly outperforms sentiment classification accuracy of state-of-the-art and baseline methods.

Originality/value

This method is capable of capturing the correct sentiment of ambiguous words and other special words by extracting aspect-context using word vector similarity instead of expensive lexical resources, and hence, shows superior performance in terms of accuracy as compared to other methods.

Details

Information Discovery and Delivery, vol. 48 no. 3
Type: Research Article
ISSN: 2398-6247

Keywords

Article
Publication date: 27 November 2020

Mingwei Tang, Jiangping Chen, Haihua Chen, Zhenyuan Xu, Yueyao Wang, Mengting Xie and Jiangwei Lin

The purpose of this paper is to provide an integrated semantic information retrieval (IR) solution based on an ontology-improved vector space model for situations where a digital…

Abstract

Purpose

The purpose of this paper is to provide an integrated semantic information retrieval (IR) solution based on an ontology-improved vector space model for situations where a digital collection is established or curated. It aims to create a retrieval approach which could return the results by meanings rather than by keywords.

Design/methodology/approach

In this paper, the authors propose a semantic term frequency algorithm to create a semantic vector space model (SeVSM) based on ontology. To support the calculation, a multi-branches tree model is created to represent the ontology and a set of algorithms is developed to operate it. Then, a semantic ontology-based IR system based on the SeVSM model is designed and developed to verify the effectiveness of the proposed model.

Findings

The experimental study using 30 queries from 15 different domains confirms the effectiveness of the SeVSM and the usability of the proposed system. The results demonstrate that the proposed model and system can be a significant exploration to enhance IR in specific domains, such as a digital library and e-commerce.

Originality/value

This research not only creates a semantic retrieval model, but also provides the application approach via designing and developing a semantic retrieval system based on the model. Comparing with most of the current related research, the proposed research studies the whole process of realizing a semantic retrieval.

Details

The Electronic Library , vol. 38 no. 5/6
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 26 March 2021

Azra Nazir, Roohie Naaz Mir and Shaima Qureshi

Natural languages have a fundamental quality of suppleness that makes it possible to present a single idea in plenty of different ways. This feature is often exploited in the…

Abstract

Purpose

Natural languages have a fundamental quality of suppleness that makes it possible to present a single idea in plenty of different ways. This feature is often exploited in the academic world, leading to the theft of work referred to as plagiarism. Many approaches have been put forward to detect such cases based on various text features and grammatical structures of languages. However, there is a huge scope of improvement for detecting intelligent plagiarism.

Design/methodology/approach

To realize this, the paper introduces a hybrid model to detect intelligent plagiarism by breaking the entire process into three stages: (1) clustering, (2) vector formulation in each cluster based on semantic roles, normalization and similarity index calculation and (3) Summary generation using encoder-decoder. An effective weighing scheme has been introduced to select terms used to build vectors based on K-means, which is calculated on the synonym set for the said term. If the value calculated in the last stage lies above a predefined threshold, only then the next semantic argument is analyzed. When the similarity score for two documents is beyond the threshold, a short summary for plagiarized documents is created.

Findings

Experimental results show that this method is able to detect connotation and concealment used in idea plagiarism besides detecting literal plagiarism.

Originality/value

The proposed model can help academics stay updated by providing summaries of relevant articles. It would eliminate the practice of plagiarism infesting the academic community at an unprecedented pace. The model will also accelerate the process of reviewing academic documents, aiding in the speedy publishing of research articles.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 14 no. 3
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 30 July 2020

V. Srilakshmi, K. Anuradha and C. Shoba Bindu

This paper aims to model a technique that categorizes the texts from huge documents. The progression in internet technologies has raised the count of document accessibility, and…

Abstract

Purpose

This paper aims to model a technique that categorizes the texts from huge documents. The progression in internet technologies has raised the count of document accessibility, and thus the documents available online become countless. The text documents comprise of research article, journal papers, newspaper, technical reports and blogs. These large documents are useful and valuable for processing real-time applications. Also, these massive documents are used in several retrieval methods. Text classification plays a vital role in information retrieval technologies and is considered as an active field for processing massive applications. The aim of text classification is to categorize the large-sized documents into different categories on the basis of its contents. There exist numerous methods for performing text-related tasks such as profiling users, sentiment analysis and identification of spams, which is considered as a supervised learning issue and is addressed with text classifier.

Design/methodology/approach

At first, the input documents are pre-processed using the stop word removal and stemming technique such that the input is made effective and capable for feature extraction. In the feature extraction process, the features are extracted using the vector space model (VSM) and then, the feature selection is done for selecting the highly relevant features to perform text categorization. Once the features are selected, the text categorization is progressed using the deep belief network (DBN). The training of the DBN is performed using the proposed grasshopper crow optimization algorithm (GCOA) that is the integration of the grasshopper optimization algorithm (GOA) and Crow search algorithm (CSA). Moreover, the hybrid weight bounding model is devised using the proposed GCOA and range degree. Thus, the proposed GCOA + DBN is used for classifying the text documents.

Findings

The performance of the proposed technique is evaluated using accuracy, precision and recall is compared with existing techniques such as naive bayes, k-nearest neighbors, support vector machine and deep convolutional neural network (DCNN) and Stochastic Gradient-CAViaR + DCNN. Here, the proposed GCOA + DBN has improved performance with the values of 0.959, 0.959 and 0.96 for precision, recall and accuracy, respectively.

Originality/value

This paper proposes a technique that categorizes the texts from massive sized documents. From the findings, it can be shown that the proposed GCOA-based DBN effectively classifies the text documents.

Details

International Journal of Web Information Systems, vol. 16 no. 3
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 1 February 2000

Ralf Östermark

The performance of Aoki’s state space algorithm and the Cartesian ARIMA search algorithm (CARlMA) of Östermark and Höglund is compared. The analysis is carried out on a set of…

Abstract

The performance of Aoki’s state space algorithm and the Cartesian ARIMA search algorithm (CARlMA) of Östermark and Höglund is compared. The analysis is carried out on a set of stock prices on the Helsinki (Finland) and Stockholm (Sweden) Stock Exchanges. Demonstrates that the Finnish and Swedish stock markets differ in predictability of stock prices. With Finnish stock data, Aoki’s state space algorithm outperforms the subset of MAPE minimizing forecasts. In contrast, with Swedish stock data, ARIMA‐models of a fairly simple structure outperform Aoki’s algorithm. The stock markets are seen to differ in complexity of time series models as well as in predictability of individual asset prices.

Details

Kybernetes, vol. 29 no. 1
Type: Research Article
ISSN: 0368-492X

Keywords

Article
Publication date: 1 January 1992

Ross Wilkinson and Philip Hingston

The task of a document retrieval system is to match a query, perhaps in natural language, against a large number of natural language documents. Neural networks are known to be…

Abstract

The task of a document retrieval system is to match a query, perhaps in natural language, against a large number of natural language documents. Neural networks are known to be good pattern matchers. This article describes investigations in implementing a document retrieval system based on a neural network model. It shows that many of the standard strategies of information retrieval are applicable in a neural network model.

Details

Library Hi Tech, vol. 10 no. 1/2
Type: Research Article
ISSN: 0737-8831

1 – 10 of over 16000