Search results
1 – 10 of 222K.V. Sheelavathy and V. Udaya Rani
Internet of Things (IoT) is a network, which provides the connection with various physical objects such as smart machines, smart home appliance and so on. The physical objects are…
Abstract
Purpose
Internet of Things (IoT) is a network, which provides the connection with various physical objects such as smart machines, smart home appliance and so on. The physical objects are allocated with a unique internet address, namely, Internet Protocol, which is used to perform the data broadcasting with the external objects using the internet. The sudden increment in the number of attacks generated by intruders, causes security-related problems in IoT devices while performing the communication. The main purpose of this paper is to develop an effective attack detection to enhance the robustness against the attackers in IoT.
Design/methodology/approach
In this research, the lasso regression algorithm is proposed along with ensemble classifier for identifying the IoT attacks. The lasso algorithm is used for the process of feature selection that modeled fewer parameters for the sparse models. The type of regression is analyzed for showing higher levels when certain parts of model selection is needed for parameter elimination. The lasso regression obtains the subset for predictors to lower the prediction error with respect to the quantitative response variable. The lasso does not impose a constraint for modeling the parameters caused the coefficients with some variables shrink as zero. The selected features are classified by using an ensemble classifier, that is important for linear and nonlinear types of data in the dataset, and the models are combined for handling these data types.
Findings
The lasso regression with ensemble classifier–based attack classification comprises distributed denial-of-service and Mirai botnet attacks which achieved an improved accuracy of 99.981% than the conventional deep neural network (DNN) methods.
Originality/value
Here, an efficient lasso regression algorithm is developed for extracting the features to perform the network anomaly detection using ensemble classifier.
Details
Keywords
Muralidhar Vaman Kamath, Shrilaxmi Prashanth, Mithesh Kumar and Adithya Tantri
The compressive strength of concrete depends on many interdependent parameters; its exact prediction is not that simple because of complex processes involved in strength…
Abstract
Purpose
The compressive strength of concrete depends on many interdependent parameters; its exact prediction is not that simple because of complex processes involved in strength development. This study aims to predict the compressive strength of normal concrete and high-performance concrete using four datasets.
Design/methodology/approach
In this paper, five established individual Machine Learning (ML) regression models have been compared: Decision Regression Tree, Random Forest Regression, Lasso Regression, Ridge Regression and Multiple-Linear regression. Four datasets were studied, two of which are previous research datasets, and two datasets are from the sophisticated lab using five established individual ML regression models.
Findings
The five statistical indicators like coefficient of determination (R2), mean absolute error, root mean squared error, Nash–Sutcliffe efficiency and mean absolute percentage error have been used to compare the performance of the models. The models are further compared using statistical indicators with previous studies. Lastly, to understand the variable effect of the predictor, the sensitivity and parametric analysis were carried out to find the performance of the variable.
Originality/value
The findings of this paper will allow readers to understand the factors involved in identifying the machine learning models and concrete datasets. In so doing, we hope that this research advances the toolset needed to predict compressive strength.
Details
Keywords
The purpose of the present study was to improve the fit of women’s bifurcated garments by developing an equation that can predict the crotch length accurately by using a few basic…
Abstract
Purpose
The purpose of the present study was to improve the fit of women’s bifurcated garments by developing an equation that can predict the crotch length accurately by using a few basic body measurements. This equation could provide a simple mass-customization approach to the design of bifurcated garments.
Design/methodology/approach
Demographic characteristics and easy-to-record body measurements available in the size USA database were used to predict the crotch length. Different methodologies including best subset regression, lasso regression and principal components regression were experimented with to identify the most important predictor variables and establish a relationship between the significant predictors and crotch length.
Findings
The lasso regression model provided the highest accuracy, required only five body dimensions and dealt with multicollinearity. The preliminary pattern preparation and garment fit tests indicated that by utilizing the proposed equation, patterns of customized garments could be successfully altered to match the crotch length of the customer, thereby, improving the precision and efficiency of the pattern making process.
Originality/value
Crotch length is a crucial measurement as it determines bifurcated garment comfort as well as aesthetic fit. The crotch length is usually estimated arbitrarily based on non-scientific methods while drafting patterns, and this increases the likelihood of dissatisfaction with the fit of the lower-body garments. The present study suggested an algorithm that could predict crotch length with 90.53% accuracy using the body dimensions height, hips, waist height, knee height and arm length.
Details
Keywords
Jiandong Zhou, Xiang Li, Xiande Zhao and Liang Wang
The purpose of this paper is to deal with the practical challenge faced by modern logistics enterprises to accurately evaluate driving performance with high computational…
Abstract
Purpose
The purpose of this paper is to deal with the practical challenge faced by modern logistics enterprises to accurately evaluate driving performance with high computational efficiency under the disturbance of road smoothness and to identify significantly associated performance influence factors.
Design/methodology/approach
The authors cooperate with a logistics server (G7) and establish a driving grading system by constructing real-time inertial navigation data-enabled indicators for both driving behaviour (times of aggressive speed change and times of lane change) and road smoothness (average speed and average vibration times of the vehicle body).
Findings
The developed driving grading system demonstrates highly accurate evaluations in practical use. Data analytics on the constructed indicators prove the significances of both driving behaviour heterogeneity and the road smoothness effect on objective driving grading. The methodologies are validated with real-life tests on different types of vehicles, and are confirmed to be quite effective in practical tests with 95% accuracy according to prior benchmarks. Data analytics based on the grading system validate the hypotheses of the driving fatigue effect, daily traffic periods impact and transition effect. In addition, the authors empirically distinguish the impact strength of external factors (driving time, rainfall and humidity, wind speed, and air quality) on driving performance.
Practical implications
This study has good potential for providing objective driving grading as required by the modern logistics industry to improve transparent management efficiency with real-time vehicle data.
Originality/value
This study contributes to the existing research by comprehensively measuring both road smoothness and driving performance in the driving grading system in the modern logistics industry.
Details
Keywords
Fan Chao, Weibin Wang and Guang Yu
In the era of big data, there is doubt about the significance of causal inference as a paramount scientific task in the social sciences. Meanwhile, data-mining techniques rooted…
Abstract
Purpose
In the era of big data, there is doubt about the significance of causal inference as a paramount scientific task in the social sciences. Meanwhile, data-mining techniques rooted in big data and artificial intelligence (AI) have infiltrated numerous aspects of social science research. This study aims to expound the criticality of discerning causal relationships – beyond mere correlations – and scrutinizes the ramifications of big data and AI in the identification of causality.
Design/methodology/approach
This study discusses the challenges and opportunities for causality identification in the era of big data under the framework of potential outcomes model and structural causal model.
Findings
First, even in the age of big data, correlations that lack interpretability, robustness and feasibility cannot substitute causality. Second, the richness of the sample size does not help solve the problem of systematic bias in the process of causal inference. Furthermore, current AI research targets correlations rather than causality, thus creating difficulties in advancing from observations to counterfactuals.
Originality/value
This study provides insights into the impact of big data era on causal inference in the social sciences, with a view toward enhancing the pool of theoretical concepts available to researchers in relevant fields and accurately guiding the direction of scientific research in these fields.
Details
Keywords
Jianghao Chu, Tae-Hwy Lee and Aman Ullah
In this chapter we consider the “Regularization of Derivative Expectation Operator” (Rodeo) of Lafferty and Wasserman (2008) and propose a modified Rodeo algorithm for…
Abstract
In this chapter we consider the “Regularization of Derivative Expectation Operator” (Rodeo) of Lafferty and Wasserman (2008) and propose a modified Rodeo algorithm for semiparametric single index models (SIMs) in big data environment with many regressors. The method assumes sparsity that many of the regressors are irrelevant. It uses a greedy algorithm, in that, to estimate the semiparametric SIM of Ichimura (1993), all coefficients of the regressors are initially set to start from near zero, then we test iteratively if the derivative of the regression function estimator with respect to each coefficient is significantly different from zero. The basic idea of the modified Rodeo algorithm for SIM (to be called SIM-Rodeo) is to view the local bandwidth selection as a variable selection scheme which amplifies the coefficients for relevant variables while keeping the coefficients of irrelevant variables relatively small or at the initial starting values near zero. For sparse semiparametric SIM, the SIM-Rodeo algorithm is shown to attain consistency in variable selection. In addition, the algorithm is fast to finish the greedy steps. We compare SIM-Rodeo with SIM-Lasso method in Zeng et al. (2012). Our simulation results demonstrate that the proposed SIM-Rodeo method is consistent for variable selection and show that it has smaller integrated mean squared errors (IMSE) than SIM-Lasso.
Details
Keywords
The main purpose of this study is to forecast inflation rates in the case of the Turkish economy with shrinkage methods of machine learning algorithms.
Abstract
Purpose
The main purpose of this study is to forecast inflation rates in the case of the Turkish economy with shrinkage methods of machine learning algorithms.
Design/methodology/approach
This paper compares the predictive ability of a set of machine learning techniques (ridge, lasso, ada lasso and elastic net) and a group of benchmark specifications (autoregressive integrated moving average (ARIMA) and multivariate vector autoregression (VAR) models) on the extensive dataset.
Findings
Results suggest that shrinkage methods perform better for variable selection. It is also seen that lasso and elastic net algorithms outperform conventional econometric methods in the case of Turkish inflation. These algorithms choose the energy production variables, construction-sector measure, reel effective exchange rate and money market indicators as the most relevant variables for inflation forecasting.
Originality/value
Turkish economy that is a typical emerging country has experienced two digit and high volatile inflation regime starting with the year 2017. This study contributes to the literature by introducing the machine learning techniques to forecast inflation in the Turkish economy. The study also compares the relative performance of machine learning techniques and different conventional methods to predict inflation in the Turkish economy and provide the empirical methodology offering the best predictive performance among their counterparts.
Details
Keywords
Suman Chhabri, Krishnendu Hazra, Amitava Choudhury, Arijit Sinha and Manojit Ghosh
Because of the mechanical properties of aluminium (Al), an accurate prediction of its properties has been challenging. Researchers are seeking reliable models for predicting the…
Abstract
Purpose
Because of the mechanical properties of aluminium (Al), an accurate prediction of its properties has been challenging. Researchers are seeking reliable models for predicting the mechanical strength of Al alloys owing to the continuous emergence of new Al alloys and their applications. There has been widespread use of empirical and statistical models for the prediction of different mechanical properties of Al and Al alloy, such as linear and nonlinear regression. Nevertheless, the development of these models requires laborious experimental work, and they may not produce accurate results depending on the relationship between the Al properties, mix of other compositions and curing conditions.
Design/methodology/approach
Numerous machine learning (ML) models have been proposed as alternative approaches for predicting the strengths of Al and its alloys. The hardness of Al alloys has been predicted by implementing various ML algorithms, such as linear regression, ridge regression, lasso regression and artificial neural network (ANN). This investigation critically analysed and discussed the application and performance of models generated by linear regression, ridge regression, lasso regression and ANN algorithms using different mechanical properties as training parameters.
Findings
Considering the definition of the problem, linear regression has been found to be the most suitable algorithm in predicting the hardness values of AA7XXX alloys as the model generated by it best fits the data set.
Originality/value
The work presented in this paper is original and not submitted anywhere else.
Details
Keywords
Philip Kostov, Thankom Arun and Samuel Annim
This paper aims to understand household’s latent behaviour decision-making in accessing financial services. In this analysis, the determinants of the choice of the pre-entry…
Abstract
Purpose
This paper aims to understand household’s latent behaviour decision-making in accessing financial services. In this analysis, the determinants of the choice of the pre-entry Mzansi account by consumers in South Africa is looked at.
Design/methodology/approach
In this study, 102 variables, grouped in the following categories: basic literacy, understanding financial terms, targets for financial advice, desired financial education and financial perception. Using a computationally efficient variable selection algorithm, variables that can satisfactorily explain the choice of a Mzansi account were studied.
Findings
The Mzansi intervention is appealing to individuals with basic but insufficient financial education. Aspirations seem to be very influential in revealing the choice of financial services, and, to this end, Mzansi is perceived as a pre-entry account not meeting the aspirations of individuals aiming to climb up the financial services ladder. It was found that Mzansi holders view the account mainly as a vehicle for receiving payments, but, on the other hand, are debt-averse and inclined to save. Hence, although there is at present no concrete evidence that the Mzansi intervention increases access to finance via diversification (i.e. by recruiting customers into higher-level accounts and services), this analysis shows that this is very likely to be the case.
Originality/value
The issue of demand-side constraints on access to finance have been largely been ignored in the theoretical and empirical literature. This paper undertakes some preliminary steps in addressing this gap.
Details
Keywords
Ahmad Mozaffari, Nasser Lashgarian Azad and Alireza Fathi
The purpose of this paper is to demonstrate the applicability of swarm and evolutionary techniques for regularized machine learning. Generally, by defining a proper penalty…
Abstract
Purpose
The purpose of this paper is to demonstrate the applicability of swarm and evolutionary techniques for regularized machine learning. Generally, by defining a proper penalty function, regularization laws are embedded into the structure of common least square solutions to increase the numerical stability, sparsity, accuracy and robustness of regression weights. Several regularization techniques have been proposed so far which have their own advantages and disadvantages. Several efforts have been made to find fast and accurate deterministic solvers to handle those regularization techniques. However, the proposed numerical and deterministic approaches need certain knowledge of mathematical programming, and also do not guarantee the global optimality of the obtained solution. In this research, the authors propose the use of constraint swarm and evolutionary techniques to cope with demanding requirements of regularized extreme learning machine (ELM).
Design/methodology/approach
To implement the required tools for comparative numerical study, three steps are taken. The considered algorithms contain both classical and swarm and evolutionary approaches. For the classical regularization techniques, Lasso regularization, Tikhonov regularization, cascade Lasso-Tikhonov regularization, and elastic net are considered. For swarm and evolutionary-based regularization, an efficient constraint handling technique known as self-adaptive penalty function constraint handling is considered, and its algorithmic structure is modified so that it can efficiently perform the regularized learning. Several well-known metaheuristics are considered to check the generalization capability of the proposed scheme. To test the efficacy of the proposed constraint evolutionary-based regularization technique, a wide range of regression problems are used. Besides, the proposed framework is applied to a real-life identification problem, i.e. identifying the dominant factors affecting the hydrocarbon emissions of an automotive engine, for further assurance on the performance of the proposed scheme.
Findings
Through extensive numerical study, it is observed that the proposed scheme can be easily used for regularized machine learning. It is indicated that by defining a proper objective function and considering an appropriate penalty function, near global optimum values of regressors can be easily obtained. The results attest the high potentials of swarm and evolutionary techniques for fast, accurate and robust regularized machine learning.
Originality/value
The originality of the research paper lies behind the use of a novel constraint metaheuristic computing scheme which can be used for effective regularized optimally pruned extreme learning machine (OP-ELM). The self-adaption of the proposed method alleviates the user from the knowledge of the underlying system, and also increases the degree of the automation of OP-ELM. Besides, by using different types of metaheuristics, it is demonstrated that the proposed methodology is a general flexible scheme, and can be combined with different types of swarm and evolutionary-based optimization techniques to form a regularized machine learning approach.
Details