Search results

1 – 4 of 4
Open Access
Article
Publication date: 17 December 2019

Yingjie Yang, Sifeng Liu and Naiming Xie

The purpose of this paper is to propose a framework for data analytics where everything is grey in nature and the associated uncertainty is considered as an essential part in data…

1273

Abstract

Purpose

The purpose of this paper is to propose a framework for data analytics where everything is grey in nature and the associated uncertainty is considered as an essential part in data collection, profiling, imputation, analysis and decision making.

Design/methodology/approach

A comparative study is conducted between the available uncertainty models and the feasibility of grey systems is highlighted. Furthermore, a general framework for the integration of grey systems and grey sets into data analytics is proposed.

Findings

Grey systems and grey sets are useful not only for small data, but also big data as well. It is complementary to other models and can play a significant role in data analytics.

Research limitations/implications

The proposed framework brings a radical change in data analytics. It may bring a fundamental change in our way to deal with uncertainties.

Practical implications

The proposed model has the potential to avoid the mistake from a misleading data imputation.

Social implications

The proposed model takes the philosophy of grey systems in recognising the limitation of our knowledge which has significant implications in our way to deal with our social life and relations.

Originality/value

This is the first time that the whole data analytics is considered from the point of view of grey systems.

Details

Marine Economics and Management, vol. 2 no. 2
Type: Research Article
ISSN: 2516-158X

Keywords

Open Access
Article
Publication date: 3 February 2020

Kai Zheng, Xianjun Yang, Yilei Wang, Yingjie Wu and Xianghan Zheng

The purpose of this paper is to alleviate the problem of poor robustness and over-fitting caused by large-scale data in collaborative filtering recommendation algorithms.

Abstract

Purpose

The purpose of this paper is to alleviate the problem of poor robustness and over-fitting caused by large-scale data in collaborative filtering recommendation algorithms.

Design/methodology/approach

Interpreting user behavior from the probabilistic perspective of hidden variables is helpful to improve robustness and over-fitting problems. Constructing a recommendation network by variational inference can effectively solve the complex distribution calculation in the probabilistic recommendation model. Based on the aforementioned analysis, this paper uses variational auto-encoder to construct a generating network, which can restore user-rating data to solve the problem of poor robustness and over-fitting caused by large-scale data. Meanwhile, for the existing KL-vanishing problem in the variational inference deep learning model, this paper optimizes the model by the KL annealing and Free Bits methods.

Findings

The effect of the basic model is considerably improved after using the KL annealing or Free Bits method to solve KL vanishing. The proposed models evidently perform worse than competitors on small data sets, such as MovieLens 1 M. By contrast, they have better effects on large data sets such as MovieLens 10 M and MovieLens 20 M.

Originality/value

This paper presents the usage of the variational inference model for collaborative filtering recommendation and introduces the KL annealing and Free Bits methods to improve the basic model effect. Because the variational inference training denotes the probability distribution of the hidden vector, the problem of poor robustness and overfitting is alleviated. When the amount of data is relatively large in the actual application scenario, the probability distribution of the fitted actual data can better represent the user and the item. Therefore, using variational inference for collaborative filtering recommendation is of practical value.

Details

International Journal of Crowd Science, vol. 4 no. 1
Type: Research Article
ISSN: 2398-7294

Keywords

Open Access
Article
Publication date: 7 August 2017

Ali M. Abdulshahed, Andrew P. Longstaff and Simon Fletcher

The purpose of this paper is to produce an intelligent technique for modelling machine tool errors caused by the thermal distortion of Computer Numerical Control (CNC) machine…

1569

Abstract

Purpose

The purpose of this paper is to produce an intelligent technique for modelling machine tool errors caused by the thermal distortion of Computer Numerical Control (CNC) machine tools. A new metaheuristic method, the cuckoo search (CS) algorithm, based on the life of a bird family is proposed to optimize the GMC(1, N) coefficients. It is then used to predict thermal error on a small vertical milling centre based on selected sensors.

Design/methodology/approach

A Grey model with convolution integral GMC(1, N) is used to design a thermal prediction model. To enhance the accuracy of the proposed model, the generation coefficients of GMC(1, N) are optimized using a new metaheuristic method, called the CS algorithm.

Findings

The results demonstrate good agreement between the experimental and predicted thermal error. It can therefore be concluded that it is possible to optimize a Grey model using the CS algorithm, which can be used to predict the thermal error of a CNC machine tool.

Originality/value

An attempt has been made for the first time to apply CS algorithm for calibrating the GMC(1, N) model. The proposed CS-based Grey model has been validated and compared with particle swarm optimization (PSO) based Grey model. Simulations and comparison show that the CS algorithm outperforms PSO and can act as an alternative optmization algorithm for Grey models that can be used for thermal error compensation.

Details

Grey Systems: Theory and Application, vol. 7 no. 2
Type: Research Article
ISSN: 2043-9377

Keywords

Open Access
Article
Publication date: 11 August 2020

Hongfang Zhou, Xiqian Wang and Yao Zhang

Feature selection is an essential step in data mining. The core of it is to analyze and quantize the relevancy and redundancy between the features and the classes. In CFR feature…

1379

Abstract

Feature selection is an essential step in data mining. The core of it is to analyze and quantize the relevancy and redundancy between the features and the classes. In CFR feature selection method, they rarely consider which feature to choose if two or more features have the same value using evaluation criterion. In order to address this problem, the standard deviation is employed to adjust the importance between relevancy and redundancy. Based on this idea, a novel feature selection method named as Feature Selection Based on Weighted Conditional Mutual Information (WCFR) is introduced. Experimental results on ten datasets show that our proposed method has higher classification accuracy.

Details

Applied Computing and Informatics, vol. 20 no. 1/2
Type: Research Article
ISSN: 2634-1964

Keywords

Access

Only Open Access

Year

Content type

1 – 4 of 4