Search results

1 – 10 of over 43000
Article
Publication date: 18 October 2019

Jairo Francisco de Souza, Sean Wolfgand Matsui Siqueira and Bernardo Nunes

Although ontology matchers are annually proposed to address different aspects of the semantic heterogeneity problem, finding the most suitable alignment approach is still an…

Abstract

Purpose

Although ontology matchers are annually proposed to address different aspects of the semantic heterogeneity problem, finding the most suitable alignment approach is still an issue. This study aims to propose a computational solution for ontology meta-matching (OMM) and a framework designed for developers to make use of alignment techniques in their applications.

Design/methodology/approach

The framework includes some similarity functions that can be chosen by developers and then, automatically, set weights for each function to obtain better alignments. To evaluate the framework, several simulations were performed with a data set from the Ontology Alignment Evaluation Initiative. Simple similarity functions were used, rather than aligners known in the literature, to demonstrate that the results would be more influenced by the proposed meta-alignment approach than the functions used.

Findings

The results showed that the framework is able to adapt to different test cases. The approach achieved better results when compared with existing ontology meta-matchers.

Originality/value

Although approaches for OMM have been proposed, it is not easy to use them during software development. On the other hand, this work presents a framework that can be used by developers to align ontologies. New ontology matchers can be added and the framework is extensible to new methods. Moreover, this work presents a novel OMM approach modeled as a linear equation system which can be easily computed.

Details

International Journal of Web Information Systems, vol. 16 no. 2
Type: Research Article
ISSN: 1744-0084

Keywords

Open Access
Article
Publication date: 26 November 2020

Bernadette Bouchon-Meunier and Giulianella Coletti

The paper is dedicated to the analysis of fuzzy similarity measures in uncertainty analysis in general, and in economic decision-making in particular. The purpose of this paper is…

1310

Abstract

Purpose

The paper is dedicated to the analysis of fuzzy similarity measures in uncertainty analysis in general, and in economic decision-making in particular. The purpose of this paper is to explain how a similarity measure can be chosen to quantify a qualitative description of similarities provided by experts of a given domain, in the case where the objects to compare are described through imprecise or linguistic attribute values represented by fuzzy sets. The case of qualitative dissimilarities is also addressed and the particular case of their representation by distances is presented.

Design/methodology/approach

The approach is based on measurement theory, following Tversky’s well-known paradigm.

Findings

A list of axioms which may or may not be satisfied by a qualitative comparative similarity between fuzzy objects is proposed, as extensions of axioms satisfied by similarities between crisp objects. They enable to express necessary and sufficient conditions for a numerical similarity measure to represent a comparative similarity between fuzzy objects. The representation of comparative dissimilarities is also addressed by means of specific functions depending on the distance between attribute values.

Originality/value

Examples of functions satisfying certain axioms to represent comparative similarities are given. They are based on the choice of operators to compute intersection, union and difference of fuzzy sets. A simple application of this methodology to economy is given, to show how a measure of similarity can be chosen to represent intuitive similarities expressed by an economist by means of a quantitative measure easily calculable. More detailed and formal results are given in Coletti and Bouchon-Meunier (2020) for similarities and Coletti et al. (2020) for dissimilarities.

Details

Asian Journal of Economics and Banking, vol. 4 no. 3
Type: Research Article
ISSN: 2615-9821

Keywords

Article
Publication date: 7 August 2017

Ke Zhang, Qiupin Zhong and Yuan Zuo

The purpose of this paper is to overcome the shortcomings of existing multivariate grey incidence models that cannot analyze the similarity of behavior matrixes.

Abstract

Purpose

The purpose of this paper is to overcome the shortcomings of existing multivariate grey incidence models that cannot analyze the similarity of behavior matrixes.

Design/methodology/approach

First, the feasibility of using gradient to measure the similarity of continuous functions is analyzed theoretically and intuitively. Then, a grey incidence degree is constructed for multivariable continuous functions. The model employs the gradient to measure the local similarity, as incidence coefficient function, of two functions, and combines local similarity into global similarity, as grey incidence degree by double integral. Third, the gradient incidence degree model for behavior matrix is proposed by discretizing the continuous models. Furthermore, the properties and satisfaction of grey incidence atom of the proposed model are research, respectively. Finally, a financial case is studied to examine the validity of the model.

Findings

The proposed model satisfies properties of invariance under mean value transformation, multiple transformation and linear transformation, which proves it is a model constructed from similarity perspective. Meanwhile, the case study shows that proposed model performs effectively.

Practical implications

The method proposed in the paper could be used in financial multivariable time series clustering, personalized recommendation in e-commerce, etc., when the behavior matrixes need to be analyzed from trend similarity perspective.

Originality/value

It will promote the accuracy of multivariate grey incidence model.

Details

Grey Systems: Theory and Application, vol. 7 no. 2
Type: Research Article
ISSN: 2043-9377

Keywords

Article
Publication date: 2 November 2017

Etienne St-Jean, Miruna Radu-Lefebvre and Cynthia Mathieu

One of the main goals of entrepreneurial mentoring programs is to strengthen the mentees’ self-efficacy. However, the conditions in which entrepreneurial self-efficacy (ESE) is…

2025

Abstract

Purpose

One of the main goals of entrepreneurial mentoring programs is to strengthen the mentees’ self-efficacy. However, the conditions in which entrepreneurial self-efficacy (ESE) is developed through mentoring are not yet fully explored. The purpose of this paper is to test the combined effects of mentee’s learning goal orientation (LGO) and perceived similarity with the mentor and demonstrates the role of these two variables in mentoring relationships.

Design/methodology/approach

The current study is based on a sample of 360 novice Canadian entrepreneurs who completed an online questionnaire. The authors used a cross-sectional analysis as research design.

Findings

Findings indicate that the development of ESE is optimal when mentees present low levels of LGO and perceive high similarities between their mentor and themselves. Mentees with high LGO decreased their level of ESE with more in-depth mentoring received.

Research limitations/implications

This study investigated a formal mentoring program with volunteer (unpaid) mentors. Generalization to informal mentoring relationships needs to be tested.

Practical implications

The study shows that, in order to effectively develop self-efficacy in a mentoring situation, LGO should be taken into account. Mentors can be trained to modify mentees’ LGO to increase their impact on this mindset and mentees’ ESE.

Originality/value

This is the first empirical study that demonstrates the effects of mentoring on ESE and reveals a triple moderating effect of LGO and perceived similarity in mentoring relationships.

Details

International Journal of Entrepreneurial Behavior & Research, vol. 24 no. 1
Type: Research Article
ISSN: 1355-2554

Keywords

Article
Publication date: 1 February 2016

Manoj Manuja and Deepak Garg

Syntax-based text classification (TC) mechanisms have been overtly replaced by semantic-based systems in recent years. Semantic-based TC systems are particularly useful in those…

Abstract

Purpose

Syntax-based text classification (TC) mechanisms have been overtly replaced by semantic-based systems in recent years. Semantic-based TC systems are particularly useful in those scenarios where similarity among documents is computed considering semantic relationships among their terms. Kernel functions have received major attention because of the unprecedented popularity of SVMs in the field of TC. Most of the kernel functions exploit syntactic structures of the text, but quite a few also use a priori semantic information for knowledge extraction. The purpose of this paper is to investigate semantic kernel functions in the context of TC.

Design/methodology/approach

This work presents performance and accuracy analysis of seven semantic kernel functions (Semantic Smoothing Kernel, Latent Semantic Kernel, Semantic WordNet-based Kernel, Semantic Smoothing Kernel having Implicit Superconcept Expansions, Compactness-based Disambiguation Kernel Function, Omiotis-based S-VSM semantic kernel function and Top-k S-VSM semantic kernel) being implemented with SVM as kernel method. All seven semantic kernels are implemented in SVM-Light tool.

Findings

Performance and accuracy parameters of seven semantic kernel functions have been evaluated and compared. The experimental results show that Top-k S-VSM semantic kernel has the highest performance and accuracy among all the evaluated kernel functions which make it a preferred building block for kernel methods for TC and retrieval.

Research limitations/implications

A combination of semantic kernel function with syntactic kernel function needs to be investigated as there is a scope of further improvement in terms of accuracy and performance in all the seven semantic kernel functions.

Practical implications

This research provides an insight into TC using a priori semantic knowledge. Three commonly used data sets are being exploited. It will be quite interesting to explore these kernel functions on live web data which may test their actual utility in real business scenarios.

Originality/value

Comparison of performance and accuracy parameters is the novel point of this research paper. To the best of the authors’ knowledge, this type of comparison has not been done previously.

Details

Program, vol. 50 no. 1
Type: Research Article
ISSN: 0033-0337

Keywords

Open Access
Article
Publication date: 4 August 2020

Kanak Meena, Devendra K. Tayal, Oscar Castillo and Amita Jain

The scalability of similarity joins is threatened by the unexpected data characteristic of data skewness. This is a pervasive problem in scientific data. Due to skewness, the…

737

Abstract

The scalability of similarity joins is threatened by the unexpected data characteristic of data skewness. This is a pervasive problem in scientific data. Due to skewness, the uneven distribution of attributes occurs, and it can cause a severe load imbalance problem. When database join operations are applied to these datasets, skewness occurs exponentially. All the algorithms developed to date for the implementation of database joins are highly skew sensitive. This paper presents a new approach for handling data-skewness in a character- based string similarity join using the MapReduce framework. In the literature, no such work exists to handle data skewness in character-based string similarity join, although work for set based string similarity joins exists. Proposed work has been divided into three stages, and every stage is further divided into mapper and reducer phases, which are dedicated to a specific task. The first stage is dedicated to finding the length of strings from a dataset. For valid candidate pair generation, MR-Pass Join framework has been suggested in the second stage. MRFA concepts are incorporated for string similarity join, which is named as “MRFA-SSJ” (MapReduce Frequency Adaptive – String Similarity Join) in the third stage which is further divided into four MapReduce phases. Hence, MRFA-SSJ has been proposed to handle skewness in the string similarity join. The experiments have been implemented on three different datasets namely: DBLP, Query log and a real dataset of IP addresses & Cookies by deploying Hadoop framework. The proposed algorithm has been compared with three known algorithms and it has been noticed that all these algorithms fail when data is highly skewed, whereas our proposed method handles highly skewed data without any problem. A set-up of the 15-node cluster has been used in this experiment, and we are following the Zipf distribution law for the analysis of skewness factor. Also, a comparison among existing and proposed techniques has been shown. Existing techniques survived till Zipf factor 0.5 whereas the proposed algorithm survives up to Zipf factor 1. Hence the proposed algorithm is skew insensitive and ensures scalability with a reasonable query processing time for string similarity database join. It also ensures the even distribution of attributes.

Details

Applied Computing and Informatics, vol. 18 no. 1/2
Type: Research Article
ISSN: 2634-1964

Keywords

Article
Publication date: 1 January 2005

R. Obiała, B.H.V. Topping, G.M. Seed and D.E.R. Clark

This paper describes how non‐orthogonal geometric models may be transformed into orthogonal polyhedral models. The main purpose of the transformation is to obtain a geometric…

Abstract

Purpose

This paper describes how non‐orthogonal geometric models may be transformed into orthogonal polyhedral models. The main purpose of the transformation is to obtain a geometric model that is easy to describe and further modify without loss of topological information from the original model.

Design/methodology/approach

The transformation method presented in this paper is based on fuzzy logic (FL). The idea of using FL for this type of transformation was first described by Takahashi and Shimizu. This paper describes both philosophy and techniques behind the transformation method as well as its application to some example 2D and 3D models. The problem in this paper is to define a transformation technique that will change a non‐orthogonal model into a similar orthogonal model. The orthogonal model is unknown at the start of the transformation and will only be specified once the transformation is complete. The model has to satisfy certain conditions, i.e. it should be orthogonal.

Findings

The group of non‐orthogonal models that contain triangular faces such as tetrahedra or pyramids cannot be successfully recognized using this method. This algorithm fails to transform these types of problem because to do so requires modification of the structure of the model. It appears that only when the edges are divided into pieces and the sharp angles are smoothed then the method can be successfully applied. Even though the method cannot be applied to all geometric models many successful examples for 2D and 3D transformation are presented. Orthogonal models with the same topology, which make them easier to describe, are obtained.

Originality/value

This transformation makes it possible to apply simple algorithms to orthogonal models enabling the solution of complex problems usually requiring non‐orthogonal models and more complex algorithms.

Details

Engineering Computations, vol. 22 no. 1
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 27 November 2018

Rajat Kumar Mudgal, Rajdeep Niyogi, Alfredo Milani and Valentina Franzoni

The purpose of this paper is to propose and experiment a framework for analysing the tweets to find the basis of popularity of a person and extract the reasons supporting the…

Abstract

Purpose

The purpose of this paper is to propose and experiment a framework for analysing the tweets to find the basis of popularity of a person and extract the reasons supporting the popularity. Although the problem of analysing tweets to detect popular events and trends has recently attracted extensive research efforts, not much emphasis has been given to find out the reasons behind the popularity of a person based on tweets.

Design/methodology/approach

In this paper, the authors introduce a framework to find out the reasons behind the popularity of a person based on the analysis of events and the evaluation of a Web-based semantic set similarity measure applied to tweets. The methodology uses the semantic similarity measure to group similar tweets in events. Although the tweets cannot contain identical hashtags, they can refer to a unique topic with equivalent or related terminology. A special data structure maintains event information, related keywords and statistics to extract the reasons supporting popularity.

Findings

An implementation of the algorithms has been experimented on a data set of 218,490 tweets from five different countries for popularity detection and reasons extraction. The experimental results are quite encouraging and consistent in determining the reasons behind popularity. The use of Web-based semantic similarity measure is based on statistics extracted from search engines, it allows to dynamically adapt the similarity values to the variation on the correlation of words depending on current social trends.

Originality/value

To the best of the authors’ knowledge, the proposed method for finding the reason of popularity in short messages is original. The semantic set similarity presented in the paper is an original asymmetric variant of a similarity scheme developed in the context of semantic image recognition.

Details

International Journal of Web Information Systems, vol. 14 no. 4
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 18 May 2020

Xiang Chen, Yaohui Pan and Bin Luo

One challenge for tourism recommendation systems (TRSs) is the long-tail phenomenon of ratings or popularity among tourist products. This paper aims to improve the diversity and…

Abstract

Purpose

One challenge for tourism recommendation systems (TRSs) is the long-tail phenomenon of ratings or popularity among tourist products. This paper aims to improve the diversity and efficiency of TRSs utilizing the power-law distribution of long-tail data.

Design/methodology/approach

Using Sina Weibo check-in data for example, this paper demonstrates that the long-tail phenomenon exists in user travel behaviors and fits the long-tail travel data with power-law distribution. To solve data sparsity in the long-tail part and increase recommendation diversity of TRSs, the paper proposes a collaborative filtering (CF) recommendation algorithm combining with power-law distribution. Furthermore, by combining power-law distribution with locality sensitive hashing (LSH), the paper optimizes user similarity calculation to improve the calculation efficiency of TRSs.

Findings

The comparison experiments show that the proposed algorithm greatly improves the recommendation diversity and calculation efficiency while maintaining high precision and recall of recommendation, providing basis for further dynamic recommendation.

Originality/value

TRSs provide a better solution to the problem of information overload in the tourism field. However, based on the historical travel data over the whole population, most current TRSs tend to recommend hot and similar spots to users, lacking in diversity and failing to provide personalized recommendations. Meanwhile, the large high-dimensional sparse data in online social networks (OSNs) brings huge computational cost when calculating user similarity with traditional CF algorithms. In this paper, by integrating the power-law distribution of travel data and tourism recommendation technology, the authors’ work solves the problem existing in traditional TRSs that recommendation results are overly narrow and lack in serendipity, and provides users with a wider range of choices and hence improves user experience in TRSs. Meanwhile, utilizing locality sensitive hash functions, the authors’ work hashes users from high-dimensional vectors to one-dimensional integers and maps similar users into the same buckets, which realizes fast nearest neighbors search in high-dimensional space and solves the extreme sparsity problem of high dimensional travel data. Furthermore, applying the hashing results to user similarity calculation, the paper greatly reduces computational complexity and improves calculation efficiency of TRSs, which reduces the system load and enables TRSs to provide effective and timely recommendations for users.

Details

Industrial Management & Data Systems, vol. 121 no. 6
Type: Research Article
ISSN: 0263-5577

Keywords

Article
Publication date: 19 June 2017

Khai Tan Huynh, Tho Thanh Quan and Thang Hoai Bui

Service-oriented architecture is an emerging software architecture, in which web service (WS) plays a crucial role. In this architecture, the task of WS composition and…

Abstract

Purpose

Service-oriented architecture is an emerging software architecture, in which web service (WS) plays a crucial role. In this architecture, the task of WS composition and verification is required when handling complex requirement of services from users. When the number of WS becomes very huge in practice, the complexity of the composition and verification is also correspondingly high. In this paper, the authors aim to propose a logic-based clustering approach to solve this problem by separating the original repository of WS into clusters. Moreover, they also propose a so-called quality-controlled clustering approach to ensure the quality of generated clusters in a reasonable execution time.

Design/methodology/approach

The approach represents WSs as logical formulas on which the authors conduct the clustering task. They also combine two most popular clustering approaches of hierarchical agglomerative clustering (HAC) and k-means to ensure the quality of generated clusters.

Findings

This logic-based clustering approach really helps to increase the performance of the WS composition and verification significantly. Furthermore, the logic-based approach helps us to maintain the soundness and completeness of the composition solution. Eventually, the quality-controlled strategy can ensure the quality of generated clusters in low complexity time.

Research limitations/implications

The work discussed in this paper is just implemented as a research tool known as WSCOVER. More work is needed to make it a practical and usable system for real life applications.

Originality/value

In this paper, the authors propose a logic-based paradigm to represent and cluster WSs. Moreover, they also propose an approach of quality-controlled clustering which combines and takes advantages of two most popular clustering approaches of HAC and k-means.

1 – 10 of over 43000