Search results

21 – 30 of over 5000
Article
Publication date: 10 January 2020

Khawla Asmi, Dounia Lotfi and Mohamed El Marraki

The state-of-the-art methods designed for overlapping community detection are limited by their high execution time as in CPM or the need to provide some parameters like the number…

Abstract

Purpose

The state-of-the-art methods designed for overlapping community detection are limited by their high execution time as in CPM or the need to provide some parameters like the number of communities in Bigclam and Nise_sph, which is a nontrivial information. Hence, there is a need to develop the accuracy that represents the primordial goal, where the actual state-of-the-art methods do not succeed to achieve high correspondence with the ground truth for many instances of networks. The paper aims to discuss this issue.

Design/methodology/approach

The authors offer a new method that explore the union of all maximum spanning trees (UMST) and models the strength of links between nodes. Also, each node in the UMST is linked with its most similar neighbor. From this model, the authors extract local community for each node, and then they combine the produced communities according to their number of shared nodes.

Findings

The experiments on eight real-world data sets and four sets of artificial networks show that the proposed method achieves obvious improvements over four state-of-the-art (BigClam, OSLOM, Demon, SE, DMST and ST) methods in terms of the F-score and ONMI for the networks with ground truth (Amazon, Youtube, LiveJournal and Orkut). Also, for the other networks, it provides communities with a good overlapping modularity.

Originality/value

In this paper, the authors investigate the UMST for the overlapping community detection.

Book part
Publication date: 29 January 2024

Shafeeq Ahmed Ali, Mohammad H. Allaymoun, Ahmad Yahia Mustafa Al Astal and Rehab Saleh

This chapter focuses on a case study of Kareem Exchange Company and its use of big data analysis to detect and prevent fraud and suspicious financial transactions. The chapter…

Abstract

This chapter focuses on a case study of Kareem Exchange Company and its use of big data analysis to detect and prevent fraud and suspicious financial transactions. The chapter describes the various phases of the big data analysis cycle, including discovery, data preparation, model planning, model building, operationalization, and communicating results, and how the Kareem Exchange Company team implemented each phase. This chapter emphasizes the importance of identifying the business problem, understanding the resources and stakeholders involved, and developing an initial hypothesis to guide the analysis. The case study results demonstrate the potential of big data analysis to improve fraud detection capabilities in financial institutions, leading to informed decision making and action.

Details

Digital Technology and Changing Roles in Managerial and Financial Accounting: Theoretical Knowledge and Practical Application
Type: Book
ISBN: 978-1-80455-973-4

Keywords

Article
Publication date: 17 July 2023

Marcelo Cajias and Joseph-Alexander Zeitler

The paper employs a unique online user-generated housing search dataset and introduces a novel measure for housing demand, namely “contacts per listing” as explained by hedonic…

Abstract

Purpose

The paper employs a unique online user-generated housing search dataset and introduces a novel measure for housing demand, namely “contacts per listing” as explained by hedonic, geographic and socioeconomic variables.

Design/methodology/approach

The authors explore housing demand by employing an extensive Internet search dataset from a German housing market platform. The authors apply state-of-the-art artificial intelligence, the eXtreme Gradient Boosting, to quantify factors that lead an apartment to be in demand.

Findings

The authors compare the results to alternative parametric models and find evidence of the superiority of the nonparametric model. The authors use eXplainable artificial intelligence (XAI) techniques to show economic meanings and inferences of the results. The results suggest that hedonic, socioeconomic and spatial aspects influence search intensity. The authors further find differences in temporal dynamics and geographical variations.

Originality/value

To the best of the authors’ knowledge, it is the first study of its kind. The statistical model of housing search draws on insights from decision theory, AI and qualitative studies on housing search. The econometric approach employed is new as it considers standard regression models and an eXtreme Gradient Boosting (XGB or XGBoost) approach followed by a model-agnostic interpretation of the underlying effects.

Details

Journal of European Real Estate Research, vol. 16 no. 2
Type: Research Article
ISSN: 1753-9269

Keywords

Book part
Publication date: 24 September 2018

Jonna Bornemark

What happens when we limit our understanding of reason to a calculating competence? In this chapter, I will approach the contemporary introduction of New Public Management (NPM…

Abstract

What happens when we limit our understanding of reason to a calculating competence? In this chapter, I will approach the contemporary introduction of New Public Management (NPM) in the Swedish public sector from the point of view of the fifteenth century philosopher Nicholas of Cusa and his critical analysis of reason and not-knowing. Cusa emphasises not-knowing as something which we cannot and should not avoid. As such it is central to every creation of knowledge. Reason, as the process to gaining knowledge also includes the capacity to relate to not-knowing. In modernity, the understanding of not-knowing has decreased and accordingly changed our understanding of reason. Reason became a calculating capacity, what Cusa calls ratio, rather than a reflecting capacity, what Cusa calls intellectus. The introduction of NPM in the Swedish public sector can, from this point of view, be seen as a kind of ratio-organisation, and I will point out three characteristics of this ratiofication: First, it includes a ‘concept imperialism’, as concepts from outside of the public service-activities displaces concepts that come from within. In this displacement, easily measurable concepts and concepts that frame a measurement-culture displace concepts that belong to the intellect. Second, we can see an ‘empaperment’ when every act has to be documented in order to be counted as complete, and where the empapered world of ratio becomes more central than the lived world with its constant presence of not-knowing. Third, this also results in a ‘remote controlling’ of activities when the acts of the staff are governed from the outside, and the competence to listen to the not-knowing of each situation is not valued.

Article
Publication date: 28 August 2009

Vassiliki A. Koutsonikola, Sophia G. Petridou, Athena I. Vakali and Georgios I. Papadimitriou

Web users' clustering is an important mining task since it contributes in identifying usage patterns, a beneficial task for a wide range of applications that rely on the web. The…

Abstract

Purpose

Web users' clustering is an important mining task since it contributes in identifying usage patterns, a beneficial task for a wide range of applications that rely on the web. The purpose of this paper is to examine the usage of Kullback‐Leibler (KL) divergence, an information theoretic distance, as an alternative option for measuring distances in web users clustering.

Design/methodology/approach

KL‐divergence is compared with other well‐known distance measures and clustering results are evaluated using a criterion function, validity indices, and graphical representations. Furthermore, the impact of noise (i.e. occasional or mistaken page visits) is evaluated, since it is imperative to assess whether a clustering process exhibits tolerance in noisy environments such as the web.

Findings

The proposed KL clustering approach is of similar performance when compared with other distance measures under both synthetic and real data workloads. Moreover, imposing extra noise on real data, the approach shows minimum deterioration among most of the other conventional distance measures.

Practical implications

The experimental results show that a probabilistic measure such as KL‐divergence has proven to be quite efficient in noisy environments and thus constitute a good alternative, the web users clustering problem.

Originality/value

This work is inspired by the usage of divergence in clustering of biological data and it is introduced by the authors in the area of web clustering. According to the experimental results presented in this paper, KL‐divergence can be considered as a good alternative for measuring distances in noisy environments such as the web.

Details

International Journal of Web Information Systems, vol. 5 no. 3
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 14 June 2019

Nora Madi, Rawan Al-Matham and Hend Al-Khalifa

The purpose of this paper is to provide an overall review of grammar checking and relation extraction (RE) literature, their techniques and the open challenges associated with…

Abstract

Purpose

The purpose of this paper is to provide an overall review of grammar checking and relation extraction (RE) literature, their techniques and the open challenges associated with them; and, finally, suggest future directions.

Design/methodology/approach

The review on grammar checking and RE was carried out using the following protocol: we prepared research questions, planed for searching strategy, addressed paper selection criteria to distinguish relevant works, extracted data from these works, and finally, analyzed and synthesized the data.

Findings

The output of error detection models could be used for creating a profile of a certain writer. Such profiles can be used for author identification, native language identification or even the level of education, to name a few. The automatic extraction of relations could be used to build or complete electronic lexical thesauri and knowledge bases.

Originality/value

Grammar checking is the process of detecting and sometimes correcting erroneous words in the text, while RE is the process of detecting and categorizing predefined relationships between entities or words that were identified in the text. The authors found that the most obvious challenge is the lack of data sets, especially for low-resource languages. Also, the lack of unified evaluation methods hinders the ability to compare results.

Details

Data Technologies and Applications, vol. 53 no. 3
Type: Research Article
ISSN: 2514-9288

Keywords

Article
Publication date: 1 February 2016

Chongchong Zhao, Chao Dong and Xiaoming Zhang

The integration and retrieval of the vast data have attracted sufficient attention, thus the W3C workgroup releases R2RML to standardize the transformation from relational data to…

2321

Abstract

Purpose

The integration and retrieval of the vast data have attracted sufficient attention, thus the W3C workgroup releases R2RML to standardize the transformation from relational data to semantic-aware data. However, it only provides a data transform mechanism to resource description framework (RDF). The generation of mapping alignments still needs manual work or other algorithms. Therefore, the purpose of this paper is to propose a domain-oriented automatic mapping method and an application of the R2RML standard.

Design/methodology/approach

In this paper, materials science is focussed to show an example of domain-oriented mapping. source field concept and M3B2 (Metal Materials Mapping Background Base) knowledge bases are established to support the auto-recommending algorithm. As for the generation of RDF files, the idea is to generate the triples and the links, respectively. The links of the triples follow the object-subject relationship, and the links of the object properties can be achieved by the range individuals and the trail path.

Findings

Consequently based on the previous work, the authors proposed Engine for Metal Materials Mapping Background Base (EM3B2), a semantic integration engine for materials science. EM3B2 not only offers friendly graphical interfaces, but also provides auto-recommending mapping based on materials knowledge to enable users to avoid vast manually work. The experimental result indicates that EM3B2 supplies accurate mapping. Moreover, the running time of E3MB2 is also competitive as classical methods.

Originality/value

This paper proposed EM3B2 semantic integration engine, which contributes to the relational database-to-RDF mapping by the application of W3C R2RML standard and the domain-oriented mapping.

Details

Program, vol. 50 no. 1
Type: Research Article
ISSN: 0033-0337

Keywords

Book part
Publication date: 15 January 2010

Matthieu de Lapparent

This article addresses simultaneously two important features in random utility maximisation (RUM) choice modelling: choice set generation and unobserved taste heterogeneity. It is…

Abstract

This article addresses simultaneously two important features in random utility maximisation (RUM) choice modelling: choice set generation and unobserved taste heterogeneity. It is proposed to develop and to compare definitions and properties of econometric specifications that are based on mixed logit (MXL) and latent class logit (LCL) RUM models in the additional presence of prior compensatory screening decision rules. The latter allow for continuous latent bounds that determine choice alternatives to be or not to be considered for decision making. It is also proposed to evaluate and to test each against the other ones in an application to home-to-work mode choice in the Paris region of France using 2002 data.

Details

Choice Modelling: The State-of-the-art and The State-of-practice
Type: Book
ISBN: 978-1-84950-773-8

Content available
Book part
Publication date: 29 January 2024

Abstract

Details

Digital Technology and Changing Roles in Managerial and Financial Accounting: Theoretical Knowledge and Practical Application
Type: Book
ISBN: 978-1-80455-973-4

Article
Publication date: 31 August 2005

Daniel Lemire, Harold Boley, Sean McGrath and Marcel Ball

Learning objects strive for reusability in e‐Learning to reduce cost and allow personalization of content. We show why learning objects require adapted Information Retrieval…

Abstract

Learning objects strive for reusability in e‐Learning to reduce cost and allow personalization of content. We show why learning objects require adapted Information Retrieval systems. In the spirit of the Semantic Web, we discuss the semantic description, discovery, and composition of learning objects. As part of our project, we tag learning objects with both objective (e.g., title, date, and author) and subjective (e.g., quality and relevance) metadata. We present the RACOFI (Rule‐Applying Collaborative Filtering) Composer prototype with its novel combination of two libraries and their associated engines: a collaborative filtering system and an inference rule system. We developed RACOFI to generate context‐aware recommendation lists. Context is handled by multidimensional predictions produced from a database‐driven scalable collaborative filtering algorithm. Rules are then applied to the predictions to customize the recommendations according to user profiles. The RACOFI Composer architecture has been developed into the contextaware music portal inDiscover.

Details

Interactive Technology and Smart Education, vol. 2 no. 3
Type: Research Article
ISSN: 1741-5659

Keywords

21 – 30 of over 5000