Search results

1 – 10 of over 29000
Book part
Publication date: 25 October 2023

Md Aminul Islam and Md Abu Sufian

This research navigates the confluence of data analytics, machine learning, and artificial intelligence to revolutionize the management of urban services in smart cities. The…

Abstract

This research navigates the confluence of data analytics, machine learning, and artificial intelligence to revolutionize the management of urban services in smart cities. The study thoroughly investigated with advanced tools to scrutinize key performance indicators integral to the functioning of smart cities, thereby enhancing leadership and decision-making strategies. Our work involves the implementation of various machine learning models such as Logistic Regression, Support Vector Machine, Decision Tree, Naive Bayes, and Artificial Neural Networks (ANN), to the data. Notably, the Support Vector Machine and Bernoulli Naive Bayes models exhibit robust performance with an accuracy rate of 70% precision score. In particular, the study underscores the employment of an ANN model on our existing dataset, optimized using the Adam optimizer. Although the model yields an overall accuracy of 61% and a precision score of 58%, implying correct predictions for the positive class 58% of the time, a comprehensive performance assessment using the Area Under the Receiver Operating Characteristic Curve (AUC-ROC) metrics was necessary. This evaluation results in a score of 0.475 at a threshold of 0.5, indicating that there's room for model enhancement. These models and their performance metrics serve as a key cog in our data analytics pipeline, providing decision-makers and city leaders with actionable insights that can steer urban service management decisions. Through real-time data availability and intuitive visualization dashboards, these leaders can promptly comprehend the current state of their services, pinpoint areas requiring improvement, and make informed decisions to bolster these services. This research illuminates the potential for data analytics, machine learning, and AI to significantly upgrade urban service management in smart cities, fostering sustainable and livable communities. Moreover, our findings contribute valuable knowledge to other cities aiming to adopt similar strategies, thus aiding the continued development of smart cities globally.

Details

Technology and Talent Strategies for Sustainable Smart Cities
Type: Book
ISBN: 978-1-83753-023-6

Keywords

Article
Publication date: 31 October 2023

Yangze Liang and Zhao Xu

Monitoring of the quality of precast concrete (PC) components is crucial for the success of prefabricated construction projects. Currently, quality monitoring of PC components…

Abstract

Purpose

Monitoring of the quality of precast concrete (PC) components is crucial for the success of prefabricated construction projects. Currently, quality monitoring of PC components during the construction phase is predominantly done manually, resulting in low efficiency and hindering the progress of intelligent construction. This paper presents an intelligent inspection method for assessing the appearance quality of PC components, utilizing an enhanced you look only once (YOLO) model and multi-source data. The aim of this research is to achieve automated management of the appearance quality of precast components in the prefabricated construction process through digital means.

Design/methodology/approach

The paper begins by establishing an improved YOLO model and an image dataset for evaluating appearance quality. Through object detection in the images, a preliminary and efficient assessment of the precast components' appearance quality is achieved. Moreover, the detection results are mapped onto the point cloud for high-precision quality inspection. In the case of precast components with quality defects, precise quality inspection is conducted by combining the three-dimensional model data obtained from forward design conversion with the captured point cloud data through registration. Additionally, the paper proposes a framework for an automated inspection platform dedicated to assessing appearance quality in prefabricated buildings, encompassing the platform's hardware network.

Findings

The improved YOLO model achieved a best mean average precision of 85.02% on the VOC2007 dataset, surpassing the performance of most similar models. After targeted training, the model exhibits excellent recognition capabilities for the four common appearance quality defects. When mapped onto the point cloud, the accuracy of quality inspection based on point cloud data and forward design is within 0.1 mm. The appearance quality inspection platform enables feedback and optimization of quality issues.

Originality/value

The proposed method in this study enables high-precision, visualized and automated detection of the appearance quality of PC components. It effectively meets the demand for quality inspection of precast components on construction sites of prefabricated buildings, providing technological support for the development of intelligent construction. The design of the appearance quality inspection platform's logic and framework facilitates the integration of the method, laying the foundation for efficient quality management in the future.

Details

Engineering, Construction and Architectural Management, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0969-9988

Keywords

Open Access
Article
Publication date: 9 October 2023

Aya Khaled Youssef Sayed Mohamed, Dagmar Auer, Daniel Hofer and Josef Küng

Data protection requirements heavily increased due to the rising awareness of data security, legal requirements and technological developments. Today, NoSQL databases are…

1807

Abstract

Purpose

Data protection requirements heavily increased due to the rising awareness of data security, legal requirements and technological developments. Today, NoSQL databases are increasingly used in security-critical domains. Current survey works on databases and data security only consider authorization and access control in a very general way and do not regard most of today’s sophisticated requirements. Accordingly, the purpose of this paper is to discuss authorization and access control for relational and NoSQL database models in detail with respect to requirements and current state of the art.

Design/methodology/approach

This paper follows a systematic literature review approach to study authorization and access control for different database models. Starting with a research on survey works on authorization and access control in databases, the study continues with the identification and definition of advanced authorization and access control requirements, which are generally applicable to any database model. This paper then discusses and compares current database models based on these requirements.

Findings

As no survey works consider requirements for authorization and access control in different database models so far, the authors define their requirements. Furthermore, the authors discuss the current state of the art for the relational, key-value, column-oriented, document-based and graph database models in comparison to the defined requirements.

Originality/value

This paper focuses on authorization and access control for various database models, not concrete products. This paper identifies today’s sophisticated – yet general – requirements from the literature and compares them with research results and access control features of current products for the relational and NoSQL database models.

Details

International Journal of Web Information Systems, vol. 20 no. 1
Type: Research Article
ISSN: 1744-0084

Keywords

Open Access
Article
Publication date: 22 May 2023

Edmund Baffoe-Twum, Eric Asa and Bright Awuku

Background: Geostatistics focuses on spatial or spatiotemporal datasets. Geostatistics was initially developed to generate probability distribution predictions of ore grade in the…

Abstract

Background: Geostatistics focuses on spatial or spatiotemporal datasets. Geostatistics was initially developed to generate probability distribution predictions of ore grade in the mining industry; however, it has been successfully applied in diverse scientific disciplines. This technique includes univariate, multivariate, and simulations. Kriging geostatistical methods, simple, ordinary, and universal Kriging, are not multivariate models in the usual statistical function. Notwithstanding, simple, ordinary, and universal kriging techniques utilize random function models that include unlimited random variables while modeling one attribute. The coKriging technique is a multivariate estimation method that simultaneously models two or more attributes defined with the same domains as coregionalization.

Objective: This study investigates the impact of populations on traffic volumes as a variable. The additional variable determines the strength or accuracy obtained when data integration is adopted. In addition, this is to help improve the estimation of annual average daily traffic (AADT).

Methods procedures, process: The investigation adopts the coKriging technique with AADT data from 2009 to 2016 from Montana, Minnesota, and Washington as primary attributes and population as a controlling factor (second variable). CK is implemented for this study after reviewing the literature and work completed by comparing it with other geostatistical methods.

Results, observations, and conclusions: The Investigation employed two variables. The data integration methods employed in CK yield more reliable models because their strength is drawn from multiple variables. The cross-validation results of the model types explored with the CK technique successfully evaluate the interpolation technique's performance and help select optimal models for each state. The results from Montana and Minnesota models accurately represent the states' traffic and population density. The Washington model had a few exceptions. However, the secondary attribute helped yield an accurate interpretation. Consequently, the impact of tourism, shopping, recreation centers, and possible transiting patterns throughout the state is worth exploring.

Details

Emerald Open Research, vol. 1 no. 5
Type: Research Article
ISSN: 2631-3952

Keywords

Article
Publication date: 3 September 2024

Biplab Bhattacharjee, Kavya Unni and Maheshwar Pratap

Product returns are a major challenge for e-businesses as they involve huge logistical and operational costs. Therefore, it becomes crucial to predict returns in advance. This…

Abstract

Purpose

Product returns are a major challenge for e-businesses as they involve huge logistical and operational costs. Therefore, it becomes crucial to predict returns in advance. This study aims to evaluate different genres of classifiers for product return chance prediction, and further optimizes the best performing model.

Design/methodology/approach

An e-commerce data set having categorical type attributes has been used for this study. Feature selection based on chi-square provides a selective features-set which is used as inputs for model building. Predictive models are attempted using individual classifiers, ensemble models and deep neural networks. For performance evaluation, 75:25 train/test split and 10-fold cross-validation strategies are used. To improve the predictability of the best performing classifier, hyperparameter tuning is performed using different optimization methods such as, random search, grid search, Bayesian approach and evolutionary models (genetic algorithm, differential evolution and particle swarm optimization).

Findings

A comparison of F1-scores revealed that the Bayesian approach outperformed all other optimization approaches in terms of accuracy. The predictability of the Bayesian-optimized model is further compared with that of other classifiers using experimental analysis. The Bayesian-optimized XGBoost model possessed superior performance, with accuracies of 77.80% and 70.35% for holdout and 10-fold cross-validation methods, respectively.

Research limitations/implications

Given the anonymized data, the effects of individual attributes on outcomes could not be investigated in detail. The Bayesian-optimized predictive model may be used in decision support systems, enabling real-time prediction of returns and the implementation of preventive measures.

Originality/value

There are very few reported studies on predicting the chance of order return in e-businesses. To the best of the authors’ knowledge, this study is the first to compare different optimization methods and classifiers, demonstrating the superiority of the Bayesian-optimized XGBoost classification model for returns prediction.

Details

Journal of Systems and Information Technology, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1328-7265

Keywords

Open Access
Article
Publication date: 26 August 2024

Sarath Radhakrishnan, Joan Calafell, Arnau Miró, Bernat Font and Oriol Lehmkuhl

Wall-modeled large eddy simulation (LES) is a practical tool for solving wall-bounded flows with less computational cost by avoiding the explicit resolution of the near-wall…

Abstract

Purpose

Wall-modeled large eddy simulation (LES) is a practical tool for solving wall-bounded flows with less computational cost by avoiding the explicit resolution of the near-wall region. However, its use is limited in flows that have high non-equilibrium effects like separation or transition. This study aims to present a novel methodology of using high-fidelity data and machine learning (ML) techniques to capture these non-equilibrium effects.

Design/methodology/approach

A precursor to this methodology has already been tested in Radhakrishnan et al. (2021) for equilibrium flows using LES of channel flow data. In the current methodology, the high-fidelity data chosen for training includes direct numerical simulation of a double diffuser that has strong non-equilibrium flow regions, and LES of a channel flow. The ultimate purpose of the model is to distinguish between equilibrium and non-equilibrium regions, and to provide the appropriate wall shear stress. The ML system used for this study is gradient-boosted regression trees.

Findings

The authors show that the model can be trained to make accurate predictions for both equilibrium and non-equilibrium boundary layers. In example, the authors find that the model is very effective for corner flows and flows that involve relaminarization, while performing rather ineffectively at recirculation regions.

Originality/value

Data from relaminarization regions help the model to better understand such phenomenon and to provide an appropriate boundary condition based on that. This motivates the authors to continue the research in this direction by adding more non-equilibrium phenomena to the training data to capture recirculation as well.

Details

International Journal of Numerical Methods for Heat & Fluid Flow, vol. 34 no. 8
Type: Research Article
ISSN: 0961-5539

Keywords

Open Access
Article
Publication date: 2 July 2024

Qingyun Fu, Shuxin Ding, Tao Zhang, Rongsheng Wang, Ping Hu and Cunlai Pu

To optimize train operations, dispatchers currently rely on experience for quick adjustments when delays occur. However, delay predictions often involve imprecise shifts based on…

Abstract

Purpose

To optimize train operations, dispatchers currently rely on experience for quick adjustments when delays occur. However, delay predictions often involve imprecise shifts based on known delay times. Real-time and accurate train delay predictions, facilitated by data-driven neural network models, can significantly reduce dispatcher stress and improve adjustment plans. Leveraging current train operation data, these models enable swift and precise predictions, addressing challenges posed by train delays in high-speed rail networks during unforeseen events.

Design/methodology/approach

This paper proposes CBLA-net, a neural network architecture for predicting late arrival times. It combines CNN, Bi-LSTM, and attention mechanisms to extract features, handle time series data, and enhance information utilization. Trained on operational data from the Beijing-Tianjin line, it predicts the late arrival time of a target train at the next station using multidimensional input data from the target and preceding trains.

Findings

This study evaluates our model's predictive performance using two data approaches: one considering full data and another focusing only on late arrivals. Results show precise and rapid predictions. Training with full data achieves a MAE of approximately 0.54 minutes and a RMSE of 0.65 minutes, surpassing the model trained solely on delay data (MAE: is about 1.02 min, RMSE: is about 1.52 min). Despite superior overall performance with full data, the model excels at predicting delays exceeding 15 minutes when trained exclusively on late arrivals. For enhanced adaptability to real-world train operations, training with full data is recommended.

Originality/value

This paper introduces a novel neural network model, CBLA-net, for predicting train delay times. It innovatively compares and analyzes the model's performance using both full data and delay data formats. Additionally, the evaluation of the network's predictive capabilities considers different scenarios, providing a comprehensive demonstration of the model's predictive performance.

Article
Publication date: 25 June 2024

Hua Huang, Weiwei Yu, Jiajing Yao and Peidong Yang

Aiming at solving the problems of low prediction accuracy and poor generalization caused by the difference in tool wear data distribution and the fixation of single global model…

Abstract

Purpose

Aiming at solving the problems of low prediction accuracy and poor generalization caused by the difference in tool wear data distribution and the fixation of single global model parameters, a hybrid prediction modeling method for tool wear based on joint distribution adaptation (JDA) is proposed.

Design/methodology/approach

Firstly, JDA is exploited to adapt the data features with different data distributions. Then, the adapted data features are identified by the KNN classifier. Finally, according to the tool state classification results, different regression prediction models are assigned to different wear stages to complete the whole tool wear prediction task.

Findings

The results of milling experiments show that the maximum prediction accuracy of this method is 95.13%, and it has good recognition accuracy and generalization performance. Through the application of the tool wear hybrid prediction modeling method, the prediction accuracy and generalization performance of the model are improved and the tool monitoring is realized.

Originality/value

The research results can provide solutions and a theoretical basis for the application of tool wear monitoring technology in practical industrial applications.

Details

Engineering Computations, vol. 41 no. 5
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 31 May 2024

Farzaneh Zarei and Mazdak Nik-Bakht

This paper aims to enrich the 3D urban models with data contributed by citizens to support data-driven decision-making in urban infrastructure projects. We introduced a new…

Abstract

Purpose

This paper aims to enrich the 3D urban models with data contributed by citizens to support data-driven decision-making in urban infrastructure projects. We introduced a new application domain extension to CityGML (social – input ADE) to enable citizens to store, classify and exchange comments generated by citizens regarding infrastructure elements. The main goal of social – input ADE is to add citizens’ feedback as semantic objects to the CityGML model.

Design/methodology/approach

Firstly, we identified the key functionalities of the suggested ADE and how to integrate them with existing 3D urban models. Next, we developed a high-level conceptual design outlining the main components and interactions within the social-input ADE. Then we proposed a package diagram for the social – input ADE to illustrate the organization of model elements and their dependencies. We also provide a detailed discussion of the functionality of different modules in the social-input ADE.

Findings

As a result of this research, it has seen that informative streams of information are generated via mining the stored data. The proposed ADE links the information of the built environment to the knowledge of end-users and enables an endless number of socially driven innovative solutions.

Originality/value

This work aims to provide a digital platform for aggregating, organizing and filtering the distributed end-users’ inputs and integrating them within the city’s digital twins to enhance city models. To create a data standard for integrating attributes of city physical elements and end-users’ social information and inputs in the same digital ecosystem, the open data model CityGML has been used.

Article
Publication date: 28 May 2024

Kuo-Yi Lin and Thitipong Jamrus

Motivated by recent research indicating the significant challenges posed by imbalanced datasets in industrial settings, this paper presents a novel framework for Industrial…

71

Abstract

Purpose

Motivated by recent research indicating the significant challenges posed by imbalanced datasets in industrial settings, this paper presents a novel framework for Industrial Data-driven Modeling for Imbalanced Fault Diagnosis, aiming to improve fault detection accuracy and reliability.

Design/methodology/approach

This study addressing the challenge of imbalanced datasets in predicting hard drive failures is both innovative and comprehensive. By integrating data enhancement techniques with cost-sensitive methods, the research pioneers a solution that directly targets the intrinsic issues posed by imbalanced data, a common obstacle in predictive maintenance and reliability analysis.

Findings

In real industrial environments, there is a critical demand for addressing the issue of imbalanced datasets. When faced with limited data for rare events or a heavily skewed distribution of categories, it becomes essential for models to effectively mine insights from the original imbalanced dataset. This involves employing techniques like data augmentation to generate new insights and rules, enhancing the model’s ability to accurately identify and predict failures.

Originality/value

Previous research has highlighted the complexity of diagnosing faults within imbalanced industrial datasets, often leading to suboptimal predictive accuracy. This paper bridges this gap by introducing a robust framework for Industrial Data-driven Modeling for Imbalanced Fault Diagnosis. It combines data enhancement and cost-sensitive methods to effectively manage the challenges posed by imbalanced datasets, further innovating with a bagging method to refine model optimization. The validation of the proposed approach demonstrates superior accuracy compared to existing methods, showcasing its potential to significantly improve fault diagnosis in industrial applications.

Details

Industrial Management & Data Systems, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0263-5577

Keywords

1 – 10 of over 29000