Search results

1 – 10 of over 1000
Article
Publication date: 12 May 2020

Shumpei Haginoya, Aiko Hanayama and Tamae Koike

The purpose of this paper was to compare the accuracy of linking crimes using geographical proximity between three distance measures: Euclidean (distance measured by the length of…

Abstract

Purpose

The purpose of this paper was to compare the accuracy of linking crimes using geographical proximity between three distance measures: Euclidean (distance measured by the length of a straight line between two locations), Manhattan (distance obtained by summing north-south distance and east-west distance) and the shortest route distances.

Design/methodology/approach

A total of 194 cases committed by 97 serial residential burglars in Aomori Prefecture in Japan between 2004 and 2015 were used in the present study. The Mann–Whitney U test was used to compare linked (two offenses committed by the same offender) and unlinked (two offenses committed by different offenders) pairs for each distance measure. Discrimination accuracy between linked and unlinked crime pairs was evaluated using area under the receiver operating characteristic curve (AUC).

Findings

The Mann–Whitney U test showed that the distances of the linked pairs were significantly shorter than those of the unlinked pairs for all distance measures. Comparison of the AUCs showed that the shortest route distance achieved significantly higher accuracy compared with the Euclidean distance, whereas there was no significant difference between the Euclidean and the Manhattan distance or between the Manhattan and the shortest route distance. These findings give partial support to the idea that distance measures taking the impact of environmental factors into consideration might be able to identify a crime series more accurately than Euclidean distances.

Research limitations/implications

Although the results suggested a difference between the Euclidean and the shortest route distance, it was small, and all distance measures resulted in outstanding AUC values, probably because of the ceiling effects. Further investigation that makes the same comparison in a narrower area is needed to avoid this potential inflation of discrimination accuracy.

Practical implications

The shortest route distance might contribute to improving the accuracy of crime linkage based on geographical proximity. However, further investigation is needed to recommend using the shortest route distance in practice. Given that the targeted area in the present study was relatively large, the findings may contribute especially to improve the accuracy of proactive comparative case analysis for estimating the whole picture of the distribution of serial crimes in the region by selecting more effective distance measure.

Social implications

Implications to improve the accuracy in linking crimes may contribute to assisting crime investigations and the earlier arrest of offenders.

Originality/value

The results of the present study provide an initial indication of the efficacy of using distance measures taking environmental factors into account.

Details

Journal of Criminological Research, Policy and Practice, vol. 7 no. 1
Type: Research Article
ISSN: 2056-3841

Keywords

Article
Publication date: 4 October 2018

Maha Al-Yahya

In the context of information retrieval, text genre is as important as its content, and knowledge of the text genre enhances the search engine features by providing customized…

Abstract

Purpose

In the context of information retrieval, text genre is as important as its content, and knowledge of the text genre enhances the search engine features by providing customized retrieval. The purpose of this study is to explore and evaluate the use of stylometric analysis, a quantitative analysis for the linguistics features of text, to support the task of automated text genre detection for Classical Arabic text.

Design/methodology/approach

Unsupervised clustering and supervised classification were applied on the King Saud University Corpus of Classical Arabic texts (KSUCCA) using the most frequent words in the corpus (MFWs) as stylometric features. Four popular distance measures established in stylometric research are evaluated for the genre detection task.

Findings

The results of the experiments show that stylometry-based genre clustering and classification align well with human-defined genre. The evidence suggests that genre style signals exist for Classical Arabic and can be used to support the task of automated genre detection.

Originality/value

This work targets the task of genre detection in Classical Arabic text using stylometric features, an approach that has only been previously applied to Arabic authorship attribution. The study also provides a comparison of four distance measures used in stylomtreic analysis on the KSUCCA, a corpus with over 50 million words of Classical Arabic using clustering and classification.

Details

The Electronic Library, vol. 36 no. 5
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 13 July 2015

Ioannis Tsimperidis, Vasilios Katos and Nathan Clarke

– The purpose of this paper is to investigate the feasibility of identifying the gender of an author by measuring the keystroke duration when typing a message.

Abstract

Purpose

The purpose of this paper is to investigate the feasibility of identifying the gender of an author by measuring the keystroke duration when typing a message.

Design/methodology/approach

Three classifiers were constructed and tested. The authors empirically evaluated the effectiveness of the classifiers by using empirical data. The authors used primary data as well as a publicly available dataset containing keystrokes from a different language to validate the language independence assumption.

Findings

The results of this paper indicate that it is possible to identify the gender of an author by analyzing keystroke durations with a probability of success in the region of 70 per cent.

Research limitations/implications

The proposed approach was validated with a limited number of participants and languages, yet the statistical tests show the significance of the results. However, this approach will be further tested with other languages.

Practical implications

Having the ability to identify the gender of an author of a certain piece of text has value in digital forensics, as the proposed method will be a source of circumstantial evidence for “putting fingers on keyboard” and for arbitrating cases where the true origin of a message needs to be identified.

Social implications

If the proposed method is included as part of a text-composing system (such as e-mail, and instant messaging applications), it could increase trust toward the applications that use it and may also work as a deterrent for crimes involving forgery.

Originality/value

The proposed approach combines and adapts techniques from the domains of biometric authentication and data classification.

Details

Information & Computer Security, vol. 23 no. 3
Type: Research Article
ISSN: 2056-4961

Keywords

Abstract

Details

Rutgers Studies in Accounting Analytics: Audit Analytics in the Financial Industry
Type: Book
ISBN: 978-1-78743-086-0

Article
Publication date: 1 September 2023

Shaghayegh Abolmakarem, Farshid Abdi, Kaveh Khalili-Damghani and Hosein Didehkhani

This paper aims to propose an improved version of portfolio optimization model through the prediction of the future behavior of stock returns using a combined wavelet-based long…

105

Abstract

Purpose

This paper aims to propose an improved version of portfolio optimization model through the prediction of the future behavior of stock returns using a combined wavelet-based long short-term memory (LSTM).

Design/methodology/approach

First, data are gathered and divided into two parts, namely, “past data” and “real data.” In the second stage, the wavelet transform is proposed to decompose the stock closing price time series into a set of coefficients. The derived coefficients are taken as an input to the LSTM model to predict the stock closing price time series and the “future data” is created. In the third stage, the mean-variance portfolio optimization problem (MVPOP) has iteratively been run using the “past,” “future” and “real” data sets. The epsilon-constraint method is adapted to generate the Pareto front for all three runes of MVPOP.

Findings

The real daily stock closing price time series of six stocks from the FTSE 100 between January 1, 2000, and December 30, 2020, is used to check the applicability and efficacy of the proposed approach. The comparisons of “future,” “past” and “real” Pareto fronts showed that the “future” Pareto front is closer to the “real” Pareto front. This demonstrates the efficacy and applicability of proposed approach.

Originality/value

Most of the classic Markowitz-based portfolio optimization models used past information to estimate the associated parameters of the stocks. This study revealed that the prediction of the future behavior of stock returns using a combined wavelet-based LSTM improved the performance of the portfolio.

Details

Journal of Modelling in Management, vol. 19 no. 2
Type: Research Article
ISSN: 1746-5664

Keywords

Content available
Book part
Publication date: 30 September 2020

Abstract

Details

Big Data Analytics and Intelligence: A Perspective for Health Care
Type: Book
ISBN: 978-1-83909-099-8

Article
Publication date: 1 July 2020

Maozeng Xu, Zhongya Mei, Siyu Luo and Yi Tan

This paper aims to analyze and provide insight on the algorithms for the optimization of construction site layout planning (CSLP). It resolves problems, such as the selection of…

1238

Abstract

Purpose

This paper aims to analyze and provide insight on the algorithms for the optimization of construction site layout planning (CSLP). It resolves problems, such as the selection of suitable algorithms, considering the optimality, optimization objectives and representation of layout solutions. The approaches for the better utilization of optimization algorithms are also presented.

Design/methodology/approach

To achieve the above, existing records (results = 200) were selected from three databases: Web of Science, ScienceDirect and Scopus. By implementing a systematic protocol, the articles related to the optimization algorithms for the CLSP (results = 75) were identified. Moreover, various related themes were collated and analyzed according to a coding structure.

Findings

The results indicate the consistent and increasing interest on the optimization algorithms for the CLSP, revealing that the trend in shifting to smart approaches in the construction industry is significant. Moreover, the interest in metaheuristic algorithms is dominant because 65.3% of the selected articles focus on these algorithms. The optimality, optimization objectives and solution representations are also important in algorithm selection. With the employment of other algorithms, self-developed applications and commercial software, optimization algorithms can be better utilized for solving CSLP problems. The findings also identify the gaps and directions for future research.

Research limitations/implications

The selection of articles in this review does not consider the industrial perspective and practical applications of commercial software. Further comparative analyses of major algorithms are necessary because this review only focuses on algorithm types.

Originality/value

This paper presents a comprehensive systematic review of articles published in the recent decade. It significantly contributes to the demonstration of the status and selection of CLSP algorithms and the benefit of using these algorithms. It also identifies the research gaps in knowledge and reveals potential improvements for future research.

Details

Engineering, Construction and Architectural Management, vol. 27 no. 8
Type: Research Article
ISSN: 0969-9988

Keywords

Article
Publication date: 1 June 2012

Amir H. Meghdadi and James F. Peters

The purpose of this paper is to demonstrate the effectiveness and advantages of using perceptual tolerance neighbourhoods in tolerance space‐based image similarity measures and…

Abstract

Purpose

The purpose of this paper is to demonstrate the effectiveness and advantages of using perceptual tolerance neighbourhoods in tolerance space‐based image similarity measures and its application in content‐based image classification and retrieval.

Design/methodology/approach

The proposed method in this paper is based on a set‐theoretic approach, where an image is viewed as a set of local visual elements. The method also includes a tolerance relation that detects the similarity between pairs of elements, if the difference between corresponding feature vectors is less than a threshold 2 (0,1).

Findings

It is shown that tolerance space‐based methods can be successfully used in a complete content‐based image retrieval (CBIR) system. Also, it is shown that perceptual tolerance neighbourhoods can replace tolerance classes in CBIR, resulting in more accuracy and less computations.

Originality/value

The main contribution of this paper is the introduction of perceptual tolerance neighbourhoods instead of tolerance classes in a new form of the Henry‐Peters tolerance‐based nearness measure (tNM) and a new neighbourhood‐based tolerance‐covering nearness measure (tcNM). Moreover, this paper presents a side – by – side comparison of the tolerance space based methods with other published methods on a test dataset of images.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 5 no. 2
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 24 June 2019

Nazanin Vafaei, Rita A. Ribeiro, Luis M. Camarinha-Matos and Leonilde Rocha Valera

Normalization is a crucial step in all decision models, to produce comparable and dimensionless data from heterogeneous data. As such, various normalization techniques are…

Abstract

Purpose

Normalization is a crucial step in all decision models, to produce comparable and dimensionless data from heterogeneous data. As such, various normalization techniques are available but their performance depends on a number of characteristics of the problem at hand. Thus, this study aims to introduce a recommendation framework for supporting users to select data normalization techniques that better fit the requirements in different application scenarios, based on multi-criteria decision methods.

Design/methodology/approach

Following the proposed approach, the authors compare six well-known normalization techniques applied to a case study of selecting suppliers in collaborative networks.

Findings

With this recommendation framework, the authors expect to contribute to improving the normalization of criteria in the evaluation and selection of suppliers and business partners in dynamic networked collaborative systems.

Originality/value

This is the first study about comparing normalization techniques for selecting the best normalization in dynamic multiple-criteria decision-making models in collaborative networks.

Details

Kybernetes, vol. 49 no. 4
Type: Research Article
ISSN: 0368-492X

Keywords

Article
Publication date: 3 April 2017

Pawel D. Domanski and Mateusz Gintrowski

This paper aims to present the results of the comparison between different approaches to the prediction of electricity prices. It is well-known that the properties of the data…

Abstract

Purpose

This paper aims to present the results of the comparison between different approaches to the prediction of electricity prices. It is well-known that the properties of the data generation process may prefer some modeling methods over the others. The data having an origin in social or market processes are characterized by unexpectedly wide realization space resulting in the existence of the long tails in the probabilistic density function. These data may not be easy in time series prediction using standard approaches based on the normal distribution assumptions. The electricity prices on the deregulated market fall into this category.

Design/methodology/approach

The paper presents alternative approaches, i.e. memory-based prediction and fractal approach compared with established nonlinear method of neural networks. The appropriate interpretation of results is supported with the statistical data analysis and data conditioning. These algorithms have been applied to the problem of the energy price prediction on the deregulated electricity market with data from Polish and Austrian energy stock exchanges.

Findings

The first outcome of the analysis is that there are several situations in the task of time series prediction, when standard modeling approach based on the assumption that each change is independent of the last following random Gaussian bell pattern may not be a true. In this paper, such a case was considered: price data from energy markets. Electricity prices data are biased by the human nature. It is shown that more relevant for data properties was Cauchy probabilistic distribution. Results have shown that alternative approaches may be used and prediction for both data memory-based approach resulted in the best performance.

Research limitations/implications

“Personalization” of the model is crucial aspect in the whole methodology. All available knowledge should be used on the forecasted phenomenon and incorporate it into the model. In case of the memory-based modeling, it is a specific design of the history searching routine that uses the understanding of the process features. Importance should shift toward methodology structure design and algorithm customization and then to parameter estimation. Such modeling approach may be more descriptive for the user enabling understanding of the process and further iterative improvement in a continuous striving for perfection.

Practical implications

Memory-based modeling can be practically applied. These models have large potential that is worth to be exploited. One disadvantage of this modeling approach is large calculation effort connected with a need of constant evaluation of large data sets. It was shown that a graphics processing unit (GPU) approach through parallel calculation on the graphical cards can improve it dramatically.

Social implications

The modeling of the electricity prices has big impact of the daily operation of the electricity traders and distributors. From one side, appropriate modeling can improve performance mitigating risks associated with the process. Thus, the end users should receive higher quality of services ultimately with lower prices and minimized risk of the energy loss incidents.

Originality/value

The use of the alternative approaches, such as memory-based reasoning or fractals, is very rare in the field of the electricity price forecasting. Thus, it gives a new impact for further research enabling development of better solutions incorporating all available process knowledge and customized hybrid algorithms.

Details

International Journal of Energy Sector Management, vol. 11 no. 1
Type: Research Article
ISSN: 1750-6220

Keywords

1 – 10 of over 1000