Search results

1 – 10 of over 2000
Open Access
Article
Publication date: 26 November 2020

Bernadette Bouchon-Meunier and Giulianella Coletti

The paper is dedicated to the analysis of fuzzy similarity measures in uncertainty analysis in general, and in economic decision-making in particular. The purpose of this paper is…

1310

Abstract

Purpose

The paper is dedicated to the analysis of fuzzy similarity measures in uncertainty analysis in general, and in economic decision-making in particular. The purpose of this paper is to explain how a similarity measure can be chosen to quantify a qualitative description of similarities provided by experts of a given domain, in the case where the objects to compare are described through imprecise or linguistic attribute values represented by fuzzy sets. The case of qualitative dissimilarities is also addressed and the particular case of their representation by distances is presented.

Design/methodology/approach

The approach is based on measurement theory, following Tversky’s well-known paradigm.

Findings

A list of axioms which may or may not be satisfied by a qualitative comparative similarity between fuzzy objects is proposed, as extensions of axioms satisfied by similarities between crisp objects. They enable to express necessary and sufficient conditions for a numerical similarity measure to represent a comparative similarity between fuzzy objects. The representation of comparative dissimilarities is also addressed by means of specific functions depending on the distance between attribute values.

Originality/value

Examples of functions satisfying certain axioms to represent comparative similarities are given. They are based on the choice of operators to compute intersection, union and difference of fuzzy sets. A simple application of this methodology to economy is given, to show how a measure of similarity can be chosen to represent intuitive similarities expressed by an economist by means of a quantitative measure easily calculable. More detailed and formal results are given in Coletti and Bouchon-Meunier (2020) for similarities and Coletti et al. (2020) for dissimilarities.

Details

Asian Journal of Economics and Banking, vol. 4 no. 3
Type: Research Article
ISSN: 2615-9821

Keywords

Open Access
Article
Publication date: 15 February 2022

Martin Nečaský, Petr Škoda, David Bernhauer, Jakub Klímek and Tomáš Skopal

Semantic retrieval and discovery of datasets published as open data remains a challenging task. The datasets inherently originate in the globally distributed web jungle, lacking…

1210

Abstract

Purpose

Semantic retrieval and discovery of datasets published as open data remains a challenging task. The datasets inherently originate in the globally distributed web jungle, lacking the luxury of centralized database administration, database schemes, shared attributes, vocabulary, structure and semantics. The existing dataset catalogs provide basic search functionality relying on keyword search in brief, incomplete or misleading textual metadata attached to the datasets. The search results are thus often insufficient. However, there exist many ways of improving the dataset discovery by employing content-based retrieval, machine learning tools, third-party (external) knowledge bases, countless feature extraction methods and description models and so forth.

Design/methodology/approach

In this paper, the authors propose a modular framework for rapid experimentation with methods for similarity-based dataset discovery. The framework consists of an extensible catalog of components prepared to form custom pipelines for dataset representation and discovery.

Findings

The study proposes several proof-of-concept pipelines including experimental evaluation, which showcase the usage of the framework.

Originality/value

To the best of authors’ knowledge, there is no similar formal framework for experimentation with various similarity methods in the context of dataset discovery. The framework has the ambition to establish a platform for reproducible and comparable research in the area of dataset discovery. The prototype implementation of the framework is available on GitHub.

Details

Data Technologies and Applications, vol. 56 no. 4
Type: Research Article
ISSN: 2514-9288

Keywords

Open Access
Article
Publication date: 21 June 2021

Bufei Xing, Haonan Yin, Zhijun Yan and Jiachen Wang

The purpose of this paper is to propose a new approach to retrieve similar questions in online health communities to improve the efficiency of health information retrieval and…

Abstract

Purpose

The purpose of this paper is to propose a new approach to retrieve similar questions in online health communities to improve the efficiency of health information retrieval and sharing.

Design/methodology/approach

This paper proposes a hybrid approach to combining domain knowledge similarity and topic similarity to retrieve similar questions in online health communities. The domain knowledge similarity can evaluate the domain distance between different questions. And the topic similarity measures questions’ relationship base on the extracted latent topics.

Findings

The experiment results show that the proposed method outperforms the baseline methods.

Originality/value

This method conquers the problem of word mismatch and considers the named entities included in questions, which most of existing studies did not.

Details

International Journal of Crowd Science, vol. 5 no. 2
Type: Research Article
ISSN: 2398-7294

Keywords

Open Access
Article
Publication date: 3 July 2023

Jandre J. van Rensburg, Catarina M. Santos and Simon B. de Jong

An underlying assumption in the shared mental model (SMM) literature is that SMMs improve whilst team members work together for longer. However, whether dyad members indeed have…

Abstract

Purpose

An underlying assumption in the shared mental model (SMM) literature is that SMMs improve whilst team members work together for longer. However, whether dyad members indeed have higher perceived SMMs with higher shared tenure has not been explored. This study aims to, therefore, firstly, investigate this idea, and we do so by focusing on perceived SMMs at the dyadic level. Secondly, because in today’s fast-paced world perceived SMMs often need to be built quickly for dyads to perform, we assess if goal interdependence can reduce the dyadic tenure required for higher perceived SMM similarity. Thirdly, we analyse if these processes are related to dyadic performance.

Design/methodology/approach

We collected a dual-source sample of 88 leader–member dyads across various industries. We conducted PROCESS analyses to test their first-stage moderated mediation model.

Findings

Results showed that dyadic tenure was positively related to perceived SMM similarity, and that goal interdependence moderated this relationship. Additionally, perceived SMM similarity mediated the relationship between dyadic tenure and dyadic performance. Lastly, the overall moderated mediation model was supported.

Originality/value

We contribute to the perceived SMM literature by: investigating perceived SMMs in dyads, testing a key idea regarding the influence of dyadic tenure on perceived SMMs and investigating how goal interdependence may prompt perceived SMM similarity earlier in dyadic tenure and, ultimately, improve dyadic performance.

Details

Team Performance Management: An International Journal, vol. 29 no. 3/4
Type: Research Article
ISSN: 1352-7592

Keywords

Open Access
Article
Publication date: 11 October 2023

Bachriah Fatwa Dhini, Abba Suganda Girsang, Unggul Utan Sufandi and Heny Kurniawati

The authors constructed an automatic essay scoring (AES) model in a discussion forum where the result was compared with scores given by human evaluators. This research proposes…

Abstract

Purpose

The authors constructed an automatic essay scoring (AES) model in a discussion forum where the result was compared with scores given by human evaluators. This research proposes essay scoring, which is conducted through two parameters, semantic and keyword similarities, using a SentenceTransformers pre-trained model that can construct the highest vector embedding. Combining these models is used to optimize the model with increasing accuracy.

Design/methodology/approach

The development of the model in the study is divided into seven stages: (1) data collection, (2) pre-processing data, (3) selected pre-trained SentenceTransformers model, (4) semantic similarity (sentence pair), (5) keyword similarity, (6) calculate final score and (7) evaluating model.

Findings

The multilingual paraphrase-multilingual-MiniLM-L12-v2 and distilbert-base-multilingual-cased-v1 models got the highest scores from comparisons of 11 pre-trained multilingual models of SentenceTransformers with Indonesian data (Dhini and Girsang, 2023). Both multilingual models were adopted in this study. A combination of two parameters is obtained by comparing the response of the keyword extraction responses with the rubric keywords. Based on the experimental results, proposing a combination can increase the evaluation results by 0.2.

Originality/value

This study uses discussion forum data from the general biology course in online learning at the open university for the 2020.2 and 2021.2 semesters. Forum discussion ratings are still manual. In this survey, the authors created a model that automatically calculates the value of discussion forums, which are essays based on the lecturer's answers moreover rubrics.

Details

Asian Association of Open Universities Journal, vol. 18 no. 3
Type: Research Article
ISSN: 1858-3431

Keywords

Open Access
Article
Publication date: 2 December 2016

Juan Aparicio

The purpose of this paper is to provide an outline of the major contributions in the literature on the determination of the least distance in data envelopment analysis (DEA). The…

2216

Abstract

Purpose

The purpose of this paper is to provide an outline of the major contributions in the literature on the determination of the least distance in data envelopment analysis (DEA). The focus herein is primarily on methodological developments. Specifically, attention is mainly paid to modeling aspects, computational features, the satisfaction of properties and duality. Finally, some promising avenues of future research on this topic are stated.

Design/methodology/approach

DEA is a methodology based on mathematical programming for the assessment of relative efficiency of a set of decision-making units (DMUs) that use several inputs to produce several outputs. DEA is classified in the literature as a non-parametric method because it does not assume a particular functional form for the underlying production function and presents, in this sense, some outstanding properties: the efficiency of firms may be evaluated independently on the market prices of the inputs used and outputs produced; it may be easily used with multiple inputs and outputs; a single score of efficiency for each assessed organization is obtained; this technique ranks organizations based on relative efficiency; and finally, it yields benchmarking information. DEA models provide both benchmarking information and efficiency scores for each of the evaluated units when it is applied to a dataset of observations and variables (inputs and outputs). Without a doubt, this benchmarking information gives DEA a distinct advantage over other efficiency methodologies, such as stochastic frontier analysis (SFA). Technical inefficiency is typically measured in DEA as the distance between the observed unit and a “benchmarking” target on the estimated piece-wise linear efficient frontier. The choice of this target is critical for assessing the potential performance of each DMU in the sample, as well as for providing information on how to increase its performance. However, traditional DEA models yield targets that are determined by the “furthest” efficient projection to the evaluated DMU. The projected point on the efficient frontier obtained as such may not be a representative projection for the judged unit, and consequently, some authors in the literature have suggested determining closest targets instead. The general argument behind this idea is that closer targets suggest directions of enhancement for the inputs and outputs of the inefficient units that may lead them to the efficiency with less effort. Indeed, authors like Aparicio et al. (2007) have shown, in an application on airlines, that it is possible to find substantial differences between the targets provided by applying the criterion used by the traditional DEA models, and those obtained when the criterion of closeness is utilized for determining projection points on the efficient frontier. The determination of closest targets is connected to the calculation of the least distance from the evaluated unit to the efficient frontier of the reference technology. In fact, the former is usually computed through solving mathematical programming models associated with minimizing some type of distance (e.g. Euclidean). In this particular respect, the main contribution in the literature is the paper by Briec (1998) on Hölder distance functions, where formally technical inefficiency to the “weakly” efficient frontier is defined through mathematical distances.

Findings

All the interesting features of the determination of closest targets from a benchmarking point of view have generated, in recent times, the increasing interest of researchers in the calculation of the least distance to evaluate technical inefficiency (Aparicio et al., 2014a). So, in this paper, we present a general classification of published contributions, mainly from a methodological perspective, and additionally, we indicate avenues for further research on this topic. The approaches that we cite in this paper differ in the way that the idea of similarity is made operative. Similarity is, in this sense, implemented as the closeness between the values of the inputs and/or outputs of the assessed units and those of the obtained projections on the frontier of the reference production possibility set. Similarity may be measured through multiple distances and efficiency measures. In turn, the aim is to globally minimize DEA model slacks to determine the closest efficient targets. However, as we will show later in the text, minimizing a mathematical distance in DEA is not an easy task, as it is equivalent to minimizing the distance to the complement of a polyhedral set, which is not a convex set. This complexity will justify the existence of different alternatives for solving these types of models.

Originality/value

As we are aware, this is the first survey in this topic.

Details

Journal of Centrum Cathedra, vol. 9 no. 2
Type: Research Article
ISSN: 1851-6599

Keywords

Open Access
Article
Publication date: 1 July 2019

Marco van Veller

This paper aims to the identification of journal articles that probably report on interdisciplinary research at Wageningen University & Research (WUR).

1982

Abstract

Purpose

This paper aims to the identification of journal articles that probably report on interdisciplinary research at Wageningen University & Research (WUR).

Design/methodology/approach

For identification of interdisciplinary research, an analysis is performed on journals from which articles have been cited in articles (co-)authored by WUR staff. The journals with cited articles are inventoried from the reference lists of the WUR articles. For each WUR article, a mean dissimilarity is calculated between the journal in which it has been published and the journals inventoried from the reference lists. Dissimilarities are derived from a large matrix with similarity values between journals, calculated from co-occurrence of these journals in the WUR articles’ reference lists.

Findings

For 21,191 WUR articles published between 2006 and 2015 in 2,535 journals mean dissimilarities have been calculated. The analysis shows that WUR articles with high mean dissimilarities often are published in multidisciplinary journals. Also, WUR articles with high mean dissimilarities are found in non-multidisciplinary (research field-specific) journals. For these articles (with high mean dissimilarities), this paper shows that citations are often made to more various research fields than for articles with lower mean dissimilarities.

Originality/value

Identification of articles reporting on interdisciplinary research may be important to WUR policy for strategic purposes or for the evaluation of researchers or groups. Also, this analysis enables to identify journals with high mean dissimilarities (due to WUR articles citing more various research fields). Identification of these journals with a more interdisciplinary scope can be important for collection management by the library.

Details

Collection and Curation, vol. 38 no. 3
Type: Research Article
ISSN: 2514-9326

Keywords

Open Access
Article
Publication date: 4 September 2017

Yuqin Wang, Bing Liang, Wen Ji, Shiwei Wang and Yiqiang Chen

In the past few years, millions of people started to acquire knowledge from the Massive Open Online Courses (MOOCs). MOOCs contain massive video courses produced by instructors…

2427

Abstract

Purpose

In the past few years, millions of people started to acquire knowledge from the Massive Open Online Courses (MOOCs). MOOCs contain massive video courses produced by instructors, and learners all over the world can get access to these courses via the internet. However, faced with massive courses, learners often waste much time finding courses they like. This paper aims to explore the problem that how to make accurate personalized recommendations for MOOC users.

Design/methodology/approach

This paper proposes a multi-attribute weight algorithm based on collaborative filtering (CF) to select a recommendation set of courses for target MOOC users.

Findings

The recall of the proposed algorithm in this paper is higher than both the traditional CF and a CF-based algorithm – uncertain neighbors’ collaborative filtering recommendation algorithm. The higher the recall is, the more accurate the recommendation result is.

Originality/value

This paper reflects the target users’ preferences for the first time by calculating separately the weight of the attributes and the weight of attribute values of the courses.

Details

International Journal of Crowd Science, vol. 1 no. 3
Type: Research Article
ISSN: 2398-7294

Keywords

Open Access
Article
Publication date: 10 November 2023

Sue-Ting Chang and Jia-Jhou Wu

The study aims to propose an instrument for measuring product-centeredness (i.e. the extent to which comment content is related to a product) using word embedding techniques as…

2411

Abstract

Purpose

The study aims to propose an instrument for measuring product-centeredness (i.e. the extent to which comment content is related to a product) using word embedding techniques as well as explore its determinants.

Design/methodology/approach

The study collected branded posts from 205 Instagram influencers and empirically examined how four factors (i.e. authenticity, vividness, coolness and influencer–product congruence) influence the content of the comments on branded posts.

Findings

Post authenticity and congruence are shown to have positive effects on product-centeredness. The interaction between coolness and authenticity is also significant. The number of comments or likes on branded posts is not correlated with product-centeredness.

Originality/value

In social media influencer marketing, volume-based metrics such as the numbers of likes and comments have been researched and applied extensively. However, content-based metrics are urgently needed, as fans may ignore brands and focus on influencers. The proposed instrument for assessing comment content enables marketers to construct content-based metrics. Additionally, the authors' findings enhance the understanding of social media users' engagement behaviors.

Details

Industrial Management & Data Systems, vol. 124 no. 1
Type: Research Article
ISSN: 0263-5577

Keywords

Open Access
Article
Publication date: 28 July 2020

Prabhat Pokharel, Roshan Pokhrel and Basanta Joshi

Analysis of log message is very important for the identification of a suspicious system and network activity. This analysis requires the correct extraction of variable entities…

1076

Abstract

Analysis of log message is very important for the identification of a suspicious system and network activity. This analysis requires the correct extraction of variable entities. The variable entities are extracted by comparing the logs messages against the log patterns. Each of these log patterns can be represented in the form of a log signature. In this paper, we present a hybrid approach for log signature extraction. The approach consists of two modules. The first module identifies log patterns by generating log clusters. The second module uses Named Entity Recognition (NER) to extract signatures by using the extracted log clusters. Experiments were performed on event logs from Windows Operating System, Exchange and Unix and validation of the result was done by comparing the signatures and the variable entities against the standard log documentation. The outcome of the experiments was that extracted signatures were ready to be used with a high degree of accuracy.

Details

Applied Computing and Informatics, vol. 19 no. 1/2
Type: Research Article
ISSN: 2634-1964

Keywords

1 – 10 of over 2000