Search results

1 – 7 of 7
Article
Publication date: 13 April 2023

Ian Lenaers, Kris Boudt and Lieven De Moor

The purpose is twofold. First, this study aims to establish that black box tree-based machine learning (ML) models have better predictive performance than a standard linear…

176

Abstract

Purpose

The purpose is twofold. First, this study aims to establish that black box tree-based machine learning (ML) models have better predictive performance than a standard linear regression (LR) hedonic model for rent prediction. Second, it shows the added value of analyzing tree-based ML models with interpretable machine learning (IML) techniques.

Design/methodology/approach

Data on Belgian residential rental properties were collected. Tree-based ML models, random forest regression and eXtreme gradient boosting regression were applied to derive rent prediction models to compare predictive performance with a LR model. Interpretations of the tree-based models regarding important factors in predicting rent were made using SHapley Additive exPlanations (SHAP) feature importance (FI) plots and SHAP summary plots.

Findings

Results indicate that tree-based models perform better than a LR model for Belgian residential rent prediction. The SHAP FI plots agree that asking price, cadastral income, surface livable, number of bedrooms, number of bathrooms and variables measuring the proximity to points of interest are dominant predictors. The direction of relationships between rent and its factors is determined with SHAP summary plots. In addition to linear relationships, it emerges that nonlinear relationships exist.

Originality/value

Rent prediction using ML is relatively less studied than house price prediction. In addition, studying prediction models using IML techniques is relatively new in real estate economics. Moreover, to the best of the authors’ knowledge, this study is the first to derive insights of driving determinants of predicted rents from SHAP FI and SHAP summary plots.

Details

International Journal of Housing Markets and Analysis, vol. 17 no. 1
Type: Research Article
ISSN: 1753-8270

Keywords

Content available
Article
Publication date: 4 January 2023

Shilpa Sonawani and Kailas Patil

Indoor air quality monitoring is extremely important in urban, industrial areas. Considering the devastating effect of declining quality of air in major part of the countries like…

Abstract

Purpose

Indoor air quality monitoring is extremely important in urban, industrial areas. Considering the devastating effect of declining quality of air in major part of the countries like India and China, it is highly recommended to monitor the quality of air which can help people with respiratory diseases, children and elderly people to take necessary precautions and stay safe at their homes. The purpose of this study is to detect air quality and perform predictions which could be part of smart home automation with the use of newer technology.

Design/methodology/approach

This study proposes an Internet-of-Things (IoT)-based air quality measurement, warning and prediction system for ambient assisted living. The proposed ambient assisted living system consists of low-cost air quality sensors and ESP32 controller with new generation embedded system architecture. It can detect Indoor Air Quality parameters like CO, PM2.5, NO2, O3, NH3, temperature, pressure, humidity, etc. The low cost sensor data are calibrated using machine learning techniques for performance improvement. The system has a novel prediction model, multiheaded convolutional neural networks-gated recurrent unit which can detect next hour pollution concentration. The model uses a transfer learning (TL) approach for prediction when the system is new and less data available for prediction. Any neighboring site data can be used to transfer knowledge for early predictions for the new system. It can have a mobile-based application which can send warning notifications to users if the Indoor Air Quality parameters exceed the specified threshold values. This is all required to take necessary measures against bad air quality.

Findings

The IoT-based system has implemented the TL framework, and the results of this study showed that the system works efficiently with performance improvement of 55.42% in RMSE scores for prediction at new target system with insufficient data.

Originality/value

This study demonstrates the implementation of an IoT system which uses low-cost sensors and deep learning model for predicting pollution concentration. The system is tackling the issues of the low-cost sensors for better performance. The novel approach of pretrained models and TL work very well at the new system having data insufficiency issues. This study contributes significantly with the usage of low-cost sensors, open-source advanced technology and performance improvement in prediction ability at new systems. Experimental results and findings are disclosed in this study. This will help install multiple new cost-effective monitoring stations in smart city for pollution forecasting.

Details

International Journal of Pervasive Computing and Communications, vol. 20 no. 1
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 8 March 2024

Feng Zhang, Youliang Wei and Tao Feng

GraphQL is a new Open API specification that allows clients to send queries and obtain data flexibly according to their needs. However, a high-complexity GraphQL query may lead to…

Abstract

Purpose

GraphQL is a new Open API specification that allows clients to send queries and obtain data flexibly according to their needs. However, a high-complexity GraphQL query may lead to an excessive data volume of the query result, which causes problems such as resource overload of the API server. Therefore, this paper aims to address this issue by predicting the response data volume of a GraphQL query statement.

Design/methodology/approach

This paper proposes a GraphQL response data volume prediction approach based on Code2Vec and AutoML. First, a GraphQL query statement is transformed into a path collection of an abstract syntax tree based on the idea of Code2Vec, and then the query is aggregated into a vector with the fixed length. Finally, the response result data volume is predicted by a fully connected neural network. To further improve the prediction accuracy, the prediction results of embedded features are combined with the field features and summary features of the query statement to predict the final response data volume by the AutoML model.

Findings

Experiments on two public GraphQL API data sets, GitHub and Yelp, show that the accuracy of the proposed approach is 15.85% and 50.31% higher than existing GraphQL response volume prediction approaches based on machine learning techniques, respectively.

Originality/value

This paper proposes an approach that combines Code2Vec and AutoML for GraphQL query response data volume prediction with higher accuracy.

Details

International Journal of Web Information Systems, vol. 20 no. 3
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 7 February 2022

Muralidhar Vaman Kamath, Shrilaxmi Prashanth, Mithesh Kumar and Adithya Tantri

The compressive strength of concrete depends on many interdependent parameters; its exact prediction is not that simple because of complex processes involved in strength…

Abstract

Purpose

The compressive strength of concrete depends on many interdependent parameters; its exact prediction is not that simple because of complex processes involved in strength development. This study aims to predict the compressive strength of normal concrete and high-performance concrete using four datasets.

Design/methodology/approach

In this paper, five established individual Machine Learning (ML) regression models have been compared: Decision Regression Tree, Random Forest Regression, Lasso Regression, Ridge Regression and Multiple-Linear regression. Four datasets were studied, two of which are previous research datasets, and two datasets are from the sophisticated lab using five established individual ML regression models.

Findings

The five statistical indicators like coefficient of determination (R2), mean absolute error, root mean squared error, Nash–Sutcliffe efficiency and mean absolute percentage error have been used to compare the performance of the models. The models are further compared using statistical indicators with previous studies. Lastly, to understand the variable effect of the predictor, the sensitivity and parametric analysis were carried out to find the performance of the variable.

Originality/value

The findings of this paper will allow readers to understand the factors involved in identifying the machine learning models and concrete datasets. In so doing, we hope that this research advances the toolset needed to predict compressive strength.

Details

Journal of Engineering, Design and Technology , vol. 22 no. 2
Type: Research Article
ISSN: 1726-0531

Keywords

Article
Publication date: 12 February 2024

Boyi Li, Miao Tian, Xiaohan Liu, Jun Li, Yun Su and Jiaming Ni

The purpose of this study is to predict the thermal protective performance (TPP) of flame-retardant fabric more economically using machine learning and analyze the factors…

Abstract

Purpose

The purpose of this study is to predict the thermal protective performance (TPP) of flame-retardant fabric more economically using machine learning and analyze the factors affecting the TPP using model visualization.

Design/methodology/approach

A total of 13 machine learning models were trained by collecting 414 datasets of typical flame-retardant fabric from current literature. The optimal performance model was used for feature importance ranking and correlation variable analysis through model visualization.

Findings

Five models with better performance were screened, all of which showed R2 greater than 0.96 and root mean squared error less than 3.0. Heat map results revealed that the TPP of fabrics differed significantly under different types of thermal exposure. The effect of fabric weight was more apparent in the flame or low thermal radiation environment. The increase in fabric weight, fabric thickness, air gap width and relative humidity of the air gap improved the TPP of the fabric.

Practical implications

The findings suggested that the visual analysis method of machine learning can intuitively understand the change trend and range of second-degree burn time under the influence of multiple variables. The established models can be used to predict the TPP of fabrics, providing a reference for researchers to carry out relevant research.

Originality/value

The findings of this study contribute directional insights for optimizing the structure of thermal protective clothing, and introduce innovative perspectives and methodologies for advancing heat transfer modeling in thermal protective clothing.

Details

International Journal of Clothing Science and Technology, vol. 36 no. 2
Type: Research Article
ISSN: 0955-6222

Keywords

Article
Publication date: 23 September 2022

Hossein Sohrabi and Esmatullah Noorzai

The present study aims to develop a risk-supported case-based reasoning (RS-CBR) approach for water-related projects by incorporating various uncertainties and risks in the…

Abstract

Purpose

The present study aims to develop a risk-supported case-based reasoning (RS-CBR) approach for water-related projects by incorporating various uncertainties and risks in the revision step.

Design/methodology/approach

The cases were extracted by studying 68 water-related projects. This research employs earned value management (EVM) factors to consider time and cost features and economic, natural, technical, and project risks to account for uncertainties and supervised learning models to estimate cost overrun. Time-series algorithms were also used to predict construction cost indexes (CCI) and model improvements in future forecasts. Outliers were deleted by the pre-processing process. Next, datasets were split into testing and training sets, and algorithms were implemented. The accuracy of different models was measured with the mean absolute percentage error (MAPE) and the normalized root mean square error (NRSME) criteria.

Findings

The findings show an improvement in the accuracy of predictions using datasets that consider uncertainties, and ensemble algorithms such as Random Forest and AdaBoost had higher accuracy. Also, among the single algorithms, the support vector regressor (SVR) with the sigmoid kernel outperformed the others.

Originality/value

This research is the first attempt to develop a case-based reasoning model based on various risks and uncertainties. The developed model has provided an approving overlap with machine learning models to predict cost overruns. The model has been implemented in collected water-related projects and results have been reported.

Details

Engineering, Construction and Architectural Management, vol. 31 no. 2
Type: Research Article
ISSN: 0969-9988

Keywords

Article
Publication date: 13 February 2023

Mehmet Altuğ

The purpose of this study was conducted at an enterprise that produces fasteners and is one of the leading companies in the sector in terms of market share. Possible defects in…

Abstract

Purpose

The purpose of this study was conducted at an enterprise that produces fasteners and is one of the leading companies in the sector in terms of market share. Possible defects in the coating of bolts and nuts either lead to products being scrapped or all of the coating process being repeated from beginning to end. In both cases, the enterprise faces a waste of time and excessive costs. Through this project, the six sigma theory and its means were effectively used to improve the efficiency and quality management of the company. The selection of the six sigma project has also contributed to the creation of various documents to be used for project screening and evaluation of financial results.

Design/methodology/approach

Six sigma is an optimization strategy that is used to improve the profitability of businesses, avoid waste, scrap and losses, reduce costs and improve the effectiveness of all activities to meet or exceed customers’ needs and expectations. Six sigma’s process improvement model, known as Definition-Measurement-Analysis-Improvement-Control, contributes to the economic and technical achievements of businesses. The normal distribution of a process should be within ±3 sigma of the mean. This represents a scale of 99.7% certainty. However, improving the process through the utilization of the six sigma rule, which accepts normal variabilities of processes twice as strict, will result in an error rate of 3.4 per million instead of 2,700 per million for each product or service.

Findings

Using six sigma practices to reduce the costs associated with low quality and to increase economic added value became a cultural practice. With this, the continuation of six sigma practices throughout the Company was intended. The annual cost reduction achieved with the utilization of six sigma practices can be up to $21,780. When time savings are also considered, a loss reduction of about $30,000 each year can be achieved. The coating thickness efficiency increased from 85% to 95% after the improvements made through the six sigma project. There is a significant increase in the efficiency of coating thickness. In addition, the coating thickness efficiency is also close to the target value of 95%–97%.

Originality/value

The results of the study were optimized with the help of deep learning. The performance of the model created in deep learning was quite close to the actual performance. This result implicates the validity of the improvement work. The results may act as a guide for the use of deep learning in new projects.

Details

International Journal of Lean Six Sigma, vol. 14 no. 7
Type: Research Article
ISSN: 2040-4166

Keywords

Access

Year

Last 6 months (7)

Content type

1 – 7 of 7