Search results
1 – 7 of 7Ian Lenaers, Kris Boudt and Lieven De Moor
The purpose is twofold. First, this study aims to establish that black box tree-based machine learning (ML) models have better predictive performance than a standard linear…
Abstract
Purpose
The purpose is twofold. First, this study aims to establish that black box tree-based machine learning (ML) models have better predictive performance than a standard linear regression (LR) hedonic model for rent prediction. Second, it shows the added value of analyzing tree-based ML models with interpretable machine learning (IML) techniques.
Design/methodology/approach
Data on Belgian residential rental properties were collected. Tree-based ML models, random forest regression and eXtreme gradient boosting regression were applied to derive rent prediction models to compare predictive performance with a LR model. Interpretations of the tree-based models regarding important factors in predicting rent were made using SHapley Additive exPlanations (SHAP) feature importance (FI) plots and SHAP summary plots.
Findings
Results indicate that tree-based models perform better than a LR model for Belgian residential rent prediction. The SHAP FI plots agree that asking price, cadastral income, surface livable, number of bedrooms, number of bathrooms and variables measuring the proximity to points of interest are dominant predictors. The direction of relationships between rent and its factors is determined with SHAP summary plots. In addition to linear relationships, it emerges that nonlinear relationships exist.
Originality/value
Rent prediction using ML is relatively less studied than house price prediction. In addition, studying prediction models using IML techniques is relatively new in real estate economics. Moreover, to the best of the authors’ knowledge, this study is the first to derive insights of driving determinants of predicted rents from SHAP FI and SHAP summary plots.
Details
Keywords
Shilpa Sonawani and Kailas Patil
Indoor air quality monitoring is extremely important in urban, industrial areas. Considering the devastating effect of declining quality of air in major part of the countries like…
Abstract
Purpose
Indoor air quality monitoring is extremely important in urban, industrial areas. Considering the devastating effect of declining quality of air in major part of the countries like India and China, it is highly recommended to monitor the quality of air which can help people with respiratory diseases, children and elderly people to take necessary precautions and stay safe at their homes. The purpose of this study is to detect air quality and perform predictions which could be part of smart home automation with the use of newer technology.
Design/methodology/approach
This study proposes an Internet-of-Things (IoT)-based air quality measurement, warning and prediction system for ambient assisted living. The proposed ambient assisted living system consists of low-cost air quality sensors and ESP32 controller with new generation embedded system architecture. It can detect Indoor Air Quality parameters like CO, PM2.5, NO2, O3, NH3, temperature, pressure, humidity, etc. The low cost sensor data are calibrated using machine learning techniques for performance improvement. The system has a novel prediction model, multiheaded convolutional neural networks-gated recurrent unit which can detect next hour pollution concentration. The model uses a transfer learning (TL) approach for prediction when the system is new and less data available for prediction. Any neighboring site data can be used to transfer knowledge for early predictions for the new system. It can have a mobile-based application which can send warning notifications to users if the Indoor Air Quality parameters exceed the specified threshold values. This is all required to take necessary measures against bad air quality.
Findings
The IoT-based system has implemented the TL framework, and the results of this study showed that the system works efficiently with performance improvement of 55.42% in RMSE scores for prediction at new target system with insufficient data.
Originality/value
This study demonstrates the implementation of an IoT system which uses low-cost sensors and deep learning model for predicting pollution concentration. The system is tackling the issues of the low-cost sensors for better performance. The novel approach of pretrained models and TL work very well at the new system having data insufficiency issues. This study contributes significantly with the usage of low-cost sensors, open-source advanced technology and performance improvement in prediction ability at new systems. Experimental results and findings are disclosed in this study. This will help install multiple new cost-effective monitoring stations in smart city for pollution forecasting.
Details
Keywords
Feng Zhang, Youliang Wei and Tao Feng
GraphQL is a new Open API specification that allows clients to send queries and obtain data flexibly according to their needs. However, a high-complexity GraphQL query may lead to…
Abstract
Purpose
GraphQL is a new Open API specification that allows clients to send queries and obtain data flexibly according to their needs. However, a high-complexity GraphQL query may lead to an excessive data volume of the query result, which causes problems such as resource overload of the API server. Therefore, this paper aims to address this issue by predicting the response data volume of a GraphQL query statement.
Design/methodology/approach
This paper proposes a GraphQL response data volume prediction approach based on Code2Vec and AutoML. First, a GraphQL query statement is transformed into a path collection of an abstract syntax tree based on the idea of Code2Vec, and then the query is aggregated into a vector with the fixed length. Finally, the response result data volume is predicted by a fully connected neural network. To further improve the prediction accuracy, the prediction results of embedded features are combined with the field features and summary features of the query statement to predict the final response data volume by the AutoML model.
Findings
Experiments on two public GraphQL API data sets, GitHub and Yelp, show that the accuracy of the proposed approach is 15.85% and 50.31% higher than existing GraphQL response volume prediction approaches based on machine learning techniques, respectively.
Originality/value
This paper proposes an approach that combines Code2Vec and AutoML for GraphQL query response data volume prediction with higher accuracy.
Details
Keywords
Muralidhar Vaman Kamath, Shrilaxmi Prashanth, Mithesh Kumar and Adithya Tantri
The compressive strength of concrete depends on many interdependent parameters; its exact prediction is not that simple because of complex processes involved in strength…
Abstract
Purpose
The compressive strength of concrete depends on many interdependent parameters; its exact prediction is not that simple because of complex processes involved in strength development. This study aims to predict the compressive strength of normal concrete and high-performance concrete using four datasets.
Design/methodology/approach
In this paper, five established individual Machine Learning (ML) regression models have been compared: Decision Regression Tree, Random Forest Regression, Lasso Regression, Ridge Regression and Multiple-Linear regression. Four datasets were studied, two of which are previous research datasets, and two datasets are from the sophisticated lab using five established individual ML regression models.
Findings
The five statistical indicators like coefficient of determination (R2), mean absolute error, root mean squared error, Nash–Sutcliffe efficiency and mean absolute percentage error have been used to compare the performance of the models. The models are further compared using statistical indicators with previous studies. Lastly, to understand the variable effect of the predictor, the sensitivity and parametric analysis were carried out to find the performance of the variable.
Originality/value
The findings of this paper will allow readers to understand the factors involved in identifying the machine learning models and concrete datasets. In so doing, we hope that this research advances the toolset needed to predict compressive strength.
Details
Keywords
Boyi Li, Miao Tian, Xiaohan Liu, Jun Li, Yun Su and Jiaming Ni
The purpose of this study is to predict the thermal protective performance (TPP) of flame-retardant fabric more economically using machine learning and analyze the factors…
Abstract
Purpose
The purpose of this study is to predict the thermal protective performance (TPP) of flame-retardant fabric more economically using machine learning and analyze the factors affecting the TPP using model visualization.
Design/methodology/approach
A total of 13 machine learning models were trained by collecting 414 datasets of typical flame-retardant fabric from current literature. The optimal performance model was used for feature importance ranking and correlation variable analysis through model visualization.
Findings
Five models with better performance were screened, all of which showed R2 greater than 0.96 and root mean squared error less than 3.0. Heat map results revealed that the TPP of fabrics differed significantly under different types of thermal exposure. The effect of fabric weight was more apparent in the flame or low thermal radiation environment. The increase in fabric weight, fabric thickness, air gap width and relative humidity of the air gap improved the TPP of the fabric.
Practical implications
The findings suggested that the visual analysis method of machine learning can intuitively understand the change trend and range of second-degree burn time under the influence of multiple variables. The established models can be used to predict the TPP of fabrics, providing a reference for researchers to carry out relevant research.
Originality/value
The findings of this study contribute directional insights for optimizing the structure of thermal protective clothing, and introduce innovative perspectives and methodologies for advancing heat transfer modeling in thermal protective clothing.
Details
Keywords
Hossein Sohrabi and Esmatullah Noorzai
The present study aims to develop a risk-supported case-based reasoning (RS-CBR) approach for water-related projects by incorporating various uncertainties and risks in the…
Abstract
Purpose
The present study aims to develop a risk-supported case-based reasoning (RS-CBR) approach for water-related projects by incorporating various uncertainties and risks in the revision step.
Design/methodology/approach
The cases were extracted by studying 68 water-related projects. This research employs earned value management (EVM) factors to consider time and cost features and economic, natural, technical, and project risks to account for uncertainties and supervised learning models to estimate cost overrun. Time-series algorithms were also used to predict construction cost indexes (CCI) and model improvements in future forecasts. Outliers were deleted by the pre-processing process. Next, datasets were split into testing and training sets, and algorithms were implemented. The accuracy of different models was measured with the mean absolute percentage error (MAPE) and the normalized root mean square error (NRSME) criteria.
Findings
The findings show an improvement in the accuracy of predictions using datasets that consider uncertainties, and ensemble algorithms such as Random Forest and AdaBoost had higher accuracy. Also, among the single algorithms, the support vector regressor (SVR) with the sigmoid kernel outperformed the others.
Originality/value
This research is the first attempt to develop a case-based reasoning model based on various risks and uncertainties. The developed model has provided an approving overlap with machine learning models to predict cost overruns. The model has been implemented in collected water-related projects and results have been reported.
Details
Keywords
The purpose of this study was conducted at an enterprise that produces fasteners and is one of the leading companies in the sector in terms of market share. Possible defects in…
Abstract
Purpose
The purpose of this study was conducted at an enterprise that produces fasteners and is one of the leading companies in the sector in terms of market share. Possible defects in the coating of bolts and nuts either lead to products being scrapped or all of the coating process being repeated from beginning to end. In both cases, the enterprise faces a waste of time and excessive costs. Through this project, the six sigma theory and its means were effectively used to improve the efficiency and quality management of the company. The selection of the six sigma project has also contributed to the creation of various documents to be used for project screening and evaluation of financial results.
Design/methodology/approach
Six sigma is an optimization strategy that is used to improve the profitability of businesses, avoid waste, scrap and losses, reduce costs and improve the effectiveness of all activities to meet or exceed customers’ needs and expectations. Six sigma’s process improvement model, known as Definition-Measurement-Analysis-Improvement-Control, contributes to the economic and technical achievements of businesses. The normal distribution of a process should be within ±3 sigma of the mean. This represents a scale of 99.7% certainty. However, improving the process through the utilization of the six sigma rule, which accepts normal variabilities of processes twice as strict, will result in an error rate of 3.4 per million instead of 2,700 per million for each product or service.
Findings
Using six sigma practices to reduce the costs associated with low quality and to increase economic added value became a cultural practice. With this, the continuation of six sigma practices throughout the Company was intended. The annual cost reduction achieved with the utilization of six sigma practices can be up to $21,780. When time savings are also considered, a loss reduction of about $30,000 each year can be achieved. The coating thickness efficiency increased from 85% to 95% after the improvements made through the six sigma project. There is a significant increase in the efficiency of coating thickness. In addition, the coating thickness efficiency is also close to the target value of 95%–97%.
Originality/value
The results of the study were optimized with the help of deep learning. The performance of the model created in deep learning was quite close to the actual performance. This result implicates the validity of the improvement work. The results may act as a guide for the use of deep learning in new projects.
Details