Search results

1 – 10 of over 1000
Article
Publication date: 22 March 2024

Mohd Mustaqeem, Suhel Mustajab and Mahfooz Alam

Software defect prediction (SDP) is a critical aspect of software quality assurance, aiming to identify and manage potential defects in software systems. In this paper, we have…

Abstract

Purpose

Software defect prediction (SDP) is a critical aspect of software quality assurance, aiming to identify and manage potential defects in software systems. In this paper, we have proposed a novel hybrid approach that combines Gray Wolf Optimization with Feature Selection (GWOFS) and multilayer perceptron (MLP) for SDP. The GWOFS-MLP hybrid model is designed to optimize feature selection, ultimately enhancing the accuracy and efficiency of SDP. Gray Wolf Optimization, inspired by the social hierarchy and hunting behavior of gray wolves, is employed to select a subset of relevant features from an extensive pool of potential predictors. This study investigates the key challenges that traditional SDP approaches encounter and proposes promising solutions to overcome time complexity and the curse of the dimensionality reduction problem.

Design/methodology/approach

The integration of GWOFS and MLP results in a robust hybrid model that can adapt to diverse software datasets. This feature selection process harnesses the cooperative hunting behavior of wolves, allowing for the exploration of critical feature combinations. The selected features are then fed into an MLP, a powerful artificial neural network (ANN) known for its capability to learn intricate patterns within software metrics. MLP serves as the predictive engine, utilizing the curated feature set to model and classify software defects accurately.

Findings

The performance evaluation of the GWOFS-MLP hybrid model on a real-world software defect dataset demonstrates its effectiveness. The model achieves a remarkable training accuracy of 97.69% and a testing accuracy of 97.99%. Additionally, the receiver operating characteristic area under the curve (ROC-AUC) score of 0.89 highlights the model’s ability to discriminate between defective and defect-free software components.

Originality/value

Experimental implementations using machine learning-based techniques with feature reduction are conducted to validate the proposed solutions. The goal is to enhance SDP’s accuracy, relevance and efficiency, ultimately improving software quality assurance processes. The confusion matrix further illustrates the model’s performance, with only a small number of false positives and false negatives.

Details

International Journal of Intelligent Computing and Cybernetics, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 18 April 2024

Vaishali Rajput, Preeti Mulay and Chandrashekhar Madhavrao Mahajan

Nature’s evolution has shaped intelligent behaviors in creatures like insects and birds, inspiring the field of Swarm Intelligence. Researchers have developed bio-inspired…

Abstract

Purpose

Nature’s evolution has shaped intelligent behaviors in creatures like insects and birds, inspiring the field of Swarm Intelligence. Researchers have developed bio-inspired algorithms to address complex optimization problems efficiently. These algorithms strike a balance between computational efficiency and solution optimality, attracting significant attention across domains.

Design/methodology/approach

Bio-inspired optimization techniques for feature engineering and its applications are systematically reviewed with chief objective of assessing statistical influence and significance of “Bio-inspired optimization”-based computational models by referring to vast research literature published between year 2015 and 2022.

Findings

The Scopus and Web of Science databases were explored for review with focus on parameters such as country-wise publications, keyword occurrences and citations per year. Springer and IEEE emerge as the most creative publishers, with indicative prominent and superior journals, namely, PLoS ONE, Neural Computing and Applications, Lecture Notes in Computer Science and IEEE Transactions. The “National Natural Science Foundation” of China and the “Ministry of Electronics and Information Technology” of India lead in funding projects in this area. China, India and Germany stand out as leaders in publications related to bio-inspired algorithms for feature engineering research.

Originality/value

The review findings integrate various bio-inspired algorithm selection techniques over a diverse spectrum of optimization techniques. Anti colony optimization contributes to decentralized and cooperative search strategies, bee colony optimization (BCO) improves collaborative decision-making, particle swarm optimization leads to exploration-exploitation balance and bio-inspired algorithms offer a range of nature-inspired heuristics.

Open Access
Article
Publication date: 23 January 2024

Luís Jacques de Sousa, João Poças Martins, Luís Sanhudo and João Santos Baptista

This study aims to review recent advances towards the implementation of ANN and NLP applications during the budgeting phase of the construction process. During this phase…

Abstract

Purpose

This study aims to review recent advances towards the implementation of ANN and NLP applications during the budgeting phase of the construction process. During this phase, construction companies must assess the scope of each task and map the client’s expectations to an internal database of tasks, resources and costs. Quantity surveyors carry out this assessment manually with little to no computer aid, within very austere time constraints, even though these results determine the company’s bid quality and are contractually binding.

Design/methodology/approach

This paper seeks to compile applications of machine learning (ML) and natural language processing in the architectural engineering and construction sector to find which methodologies can assist this assessment. The paper carries out a systematic literature review, following the preferred reporting items for systematic reviews and meta-analyses guidelines, to survey the main scientific contributions within the topic of text classification (TC) for budgeting in construction.

Findings

This work concludes that it is necessary to develop data sets that represent the variety of tasks in construction, achieve higher accuracy algorithms, widen the scope of their application and reduce the need for expert validation of the results. Although full automation is not within reach in the short term, TC algorithms can provide helpful support tools.

Originality/value

Given the increasing interest in ML for construction and recent developments, the findings disclosed in this paper contribute to the body of knowledge, provide a more automated perspective on budgeting in construction and break ground for further implementation of text-based ML in budgeting for construction.

Details

Construction Innovation , vol. 24 no. 7
Type: Research Article
ISSN: 1471-4175

Keywords

Open Access
Article
Publication date: 20 November 2023

Devesh Singh

This study aims to examine foreign direct investment (FDI) factors and develops a rational framework for FDI inflow in Western European countries such as France, Germany, the…

Abstract

Purpose

This study aims to examine foreign direct investment (FDI) factors and develops a rational framework for FDI inflow in Western European countries such as France, Germany, the Netherlands, Switzerland, Belgium and Austria.

Design/methodology/approach

Data for this study were collected from the World development indicators (WDI) database from 1995 to 2018. Factors such as economic growth, pollution, trade, domestic capital investment, gross value-added and the financial stability of the country that influence FDI decisions were selected through empirical literature. A framework was developed using interpretable machine learning (IML), decision trees and three-stage least squares simultaneous equation methods for FDI inflow in Western Europe.

Findings

The findings of this study show that there is a difference between the most important and trusted factors for FDI inflow. Additionally, this study shows that machine learning (ML) models can perform better than conventional linear regression models.

Research limitations/implications

This research has several limitations. Ideally, classification accuracies should be higher, and the current scope of this research is limited to examining the performance of FDI determinants within Western Europe.

Practical implications

Through this framework, the national government can understand how investors make their capital allocation decisions in their country. The framework developed in this study can help policymakers better understand the rationality of FDI inflows.

Originality/value

An IML framework has not been developed in prior studies to analyze FDI inflows. Additionally, the author demonstrates the applicability of the IML framework for estimating FDI inflows in Western Europe.

Details

Journal of Economics, Finance and Administrative Science, vol. 29 no. 57
Type: Research Article
ISSN: 2077-1886

Keywords

Article
Publication date: 20 September 2023

Hei-Chia Wang, Army Justitia and Ching-Wen Wang

The explosion of data due to the sophistication of information and communication technology makes it simple for prospective tourists to learn about previous hotel guests'…

Abstract

Purpose

The explosion of data due to the sophistication of information and communication technology makes it simple for prospective tourists to learn about previous hotel guests' experiences. They prioritize the rating score when selecting a hotel. However, rating scores are less reliable for suggesting a personalized preference for each aspect, especially when they are in a limited number. This study aims to recommend ratings and personalized preference hotels using cross-domain and aspect-based features.

Design/methodology/approach

We propose an aspect-based cross-domain personalized recommendation (AsCDPR), a novel framework for rating prediction and personalized customer preference recommendations. We incorporate a cross-domain personalized approach and aspect-based features of items from the review text. We extracted aspect-based feature vectors from two domains using bidirectional long short-term memory and then mapped them by a multilayer perceptron (MLP). The cross-domain recommendation module trains MLP to analyze sentiment and predict item ratings and the polarities of the aspect based on user preferences.

Findings

Expanded by its synonyms, aspect-based features significantly improve the performance of sentiment analysis on accuracy and the F1-score matrix. With relatively low mean absolute error and root mean square error values, AsCDPR outperforms matrix factorization, collaborative matrix factorization, EMCDPR and Personalized transfer of user preferences for cross-domain recommendation. These values are 1.3657 and 1.6682, respectively.

Research limitation/implications

This study assists users in recommending hotels based on their priority preferences. Users do not need to read other people's reviews to capture the key aspects of items. This model could enhance system reliability in the hospitality industry by providing personalized recommendations.

Originality/value

This study introduces a new approach that embeds aspect-based features of items in a cross-domain personalized recommendation. AsCDPR predicts ratings and provides recommendations based on priority aspects of each user's preferences.

Article
Publication date: 3 April 2024

Rizwan Ali, Jin Xu, Mushahid Hussain Baig, Hafiz Saif Ur Rehman, Muhammad Waqas Aslam and Kaleem Ullah Qasim

This study aims to endeavour to decode artificial intelligence (AI)-based tokens' complex dynamics and predictability using a comprehensive multivariate framework that integrates…

Abstract

Purpose

This study aims to endeavour to decode artificial intelligence (AI)-based tokens' complex dynamics and predictability using a comprehensive multivariate framework that integrates technical and macroeconomic indicators.

Design/methodology/approach

In this study we used advance machine learning techniques, such as gradient boosting regression (GBR), random forest (RF) and notably long short-term memory (LSTM) networks, this research provides a nuanced understanding of the factors driving the performance of AI tokens. The study’s comparative analysis highlights the superior predictive capabilities of LSTM models, as evidenced by their performance across various AI digital tokens such as AGIX-singularity-NET, Cortex and numeraire NMR.

Findings

This study finding shows that through an intricate exploration of feature importance and the impact of speculative behaviour, the research elucidates the long-term patterns and resilience of AI-based tokens against economic shifts. The SHapley Additive exPlanations (SHAP) analysis results show that technical and some macroeconomic factors play a dominant role in price production. It also examines the potential of these models for strategic investment and hedging, underscoring their relevance in an increasingly digital economy.

Originality/value

According to our knowledge, the absence of AI research frameworks for forecasting and modelling current aria-leading AI tokens is apparent. Due to a lack of study on understanding the relationship between the AI token market and other factors, forecasting is outstandingly demanding. This study provides a robust predictive framework to accurately identify the changing trends of AI tokens within a multivariate context and fill the gaps in existing research. We can investigate detailed predictive analytics with the help of modern AI algorithms and correct model interpretation to elaborate on the behaviour patterns of developing decentralised digital AI-based token prices.

Details

Journal of Economic Studies, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0144-3585

Keywords

Article
Publication date: 17 April 2024

Jahanzaib Alvi and Imtiaz Arif

The crux of this paper is to unveil efficient features and practical tools that can predict credit default.

Abstract

Purpose

The crux of this paper is to unveil efficient features and practical tools that can predict credit default.

Design/methodology/approach

Annual data of non-financial listed companies were taken from 2000 to 2020, along with 71 financial ratios. The dataset was bifurcated into three panels with three default assumptions. Logistic regression (LR) and k-nearest neighbor (KNN) binary classification algorithms were used to estimate credit default in this research.

Findings

The study’s findings revealed that features used in Model 3 (Case 3) were the efficient and best features comparatively. Results also showcased that KNN exposed higher accuracy than LR, which proves the supremacy of KNN on LR.

Research limitations/implications

Using only two classifiers limits this research for a comprehensive comparison of results; this research was based on only financial data, which exhibits a sizeable room for including non-financial parameters in default estimation. Both limitations may be a direction for future research in this domain.

Originality/value

This study introduces efficient features and tools for credit default prediction using financial data, demonstrating KNN’s superior accuracy over LR and suggesting future research directions.

Details

Kybernetes, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0368-492X

Keywords

Open Access
Article
Publication date: 31 July 2023

Daniel Šandor and Marina Bagić Babac

Sarcasm is a linguistic expression that usually carries the opposite meaning of what is being said by words, thus making it difficult for machines to discover the actual meaning…

2864

Abstract

Purpose

Sarcasm is a linguistic expression that usually carries the opposite meaning of what is being said by words, thus making it difficult for machines to discover the actual meaning. It is mainly distinguished by the inflection with which it is spoken, with an undercurrent of irony, and is largely dependent on context, which makes it a difficult task for computational analysis. Moreover, sarcasm expresses negative sentiments using positive words, allowing it to easily confuse sentiment analysis models. This paper aims to demonstrate the task of sarcasm detection using the approach of machine and deep learning.

Design/methodology/approach

For the purpose of sarcasm detection, machine and deep learning models were used on a data set consisting of 1.3 million social media comments, including both sarcastic and non-sarcastic comments. The data set was pre-processed using natural language processing methods, and additional features were extracted and analysed. Several machine learning models, including logistic regression, ridge regression, linear support vector and support vector machines, along with two deep learning models based on bidirectional long short-term memory and one bidirectional encoder representations from transformers (BERT)-based model, were implemented, evaluated and compared.

Findings

The performance of machine and deep learning models was compared in the task of sarcasm detection, and possible ways of improvement were discussed. Deep learning models showed more promise, performance-wise, for this type of task. Specifically, a state-of-the-art model in natural language processing, namely, BERT-based model, outperformed other machine and deep learning models.

Originality/value

This study compared the performance of the various machine and deep learning models in the task of sarcasm detection using the data set of 1.3 million comments from social media.

Details

Information Discovery and Delivery, vol. 52 no. 2
Type: Research Article
ISSN: 2398-6247

Keywords

Article
Publication date: 2 April 2024

R.S. Vignesh and M. Monica Subashini

An abundance of techniques has been presented so forth for waste classification but, they deliver inefficient results with low accuracy. Their achievement on various repositories…

Abstract

Purpose

An abundance of techniques has been presented so forth for waste classification but, they deliver inefficient results with low accuracy. Their achievement on various repositories is different and also, there is insufficiency of high-scale databases for training. The purpose of the study is to provide high security.

Design/methodology/approach

In this research, optimization-assisted federated learning (FL) is introduced for thermoplastic waste segregation and classification. The deep learning (DL) network trained by Archimedes Henry gas solubility optimization (AHGSO) is used for the classification of plastic and resin types. The deep quantum neural networks (DQNN) is used for first-level classification and the deep max-out network (DMN) is employed for second-level classification. This developed AHGSO is obtained by blending the features of Archimedes optimization algorithm (AOA) and Henry gas solubility optimization (HGSO). The entities included in this approach are nodes and servers. Local training is carried out depending on local data and updations to the server are performed. Then, the model is aggregated at the server. Thereafter, each node downloads the global model and the update training is executed depending on the downloaded global and the local model till it achieves the satisfied condition. Finally, local update and aggregation at the server is altered based on the average method. The Data tag suite (DATS_2022) dataset is used for multilevel thermoplastic waste segregation and classification.

Findings

By using the DQNN in first-level classification the designed optimization-assisted FL has gained an accuracy of 0.930, mean average precision (MAP) of 0.933, false positive rate (FPR) of 0.213, loss function of 0.211, mean square error (MSE) of 0.328 and root mean square error (RMSE) of 0.572. In the second level classification, by using DMN the accuracy, MAP, FPR, loss function, MSE and RMSE are 0.932, 0.935, 0.093, 0.068, 0.303 and 0.551.

Originality/value

The multilevel thermoplastic waste segregation and classification using the proposed model is accurate and improves the effectiveness of the classification.

Article
Publication date: 5 April 2024

Melike Artar, Yavuz Selim Balcioglu and Oya Erdil

Our proposed machine learning model contributes to improving the quality of Hire by providing a more nuanced and comprehensive analysis of candidate attributes. Instead of…

Abstract

Purpose

Our proposed machine learning model contributes to improving the quality of Hire by providing a more nuanced and comprehensive analysis of candidate attributes. Instead of focusing solely on obvious factors, such as qualifications and experience, our model also considers various dimensions of fit, including person-job fit and person-organization fit. By integrating these dimensions of fit into the model, we can better predict a candidate’s potential contribution to the organization, hence enhancing the Quality of Hire.

Design/methodology/approach

Within the scope of the investigation, the competencies of the personnel working in the IT department of one in the largest state banks of the country were used. The entire data collection includes information on 1,850 individual employees as well as 13 different characteristics. For analysis, Python’s “keras” and “seaborn” modules were used. The Gower coefficient was used to determine the distance between different records.

Findings

The K-NN method resulted in the formation of five clusters, represented as a scatter plot. The axis illustrates the cohesion that exists between things (employees) that are similar to one another and the separateness that exists between things that have their own individual identities. This shows that the clustering process is effective in improving both the degree of similarity within each cluster and the degree of dissimilarity between clusters.

Research limitations/implications

Employee competencies were evaluated within the scope of the investigation. Additionally, other criteria requested from the employee were not included in the application.

Originality/value

This study will be beneficial for academics, professionals, and researchers in their attempts to overcome the ongoing obstacles and challenges related to the securing the proper talent for an organization. In addition to creating a mechanism to use big data in the form of structured and unstructured data from multiple sources and deriving insights using ML algorithms, it contributes to the debates on the quality of hire in an entire organization. This is done in addition to developing a mechanism for using big data in the form of structured and unstructured data from multiple sources.

Details

Management Decision, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0025-1747

Keywords

1 – 10 of over 1000