Search results

1 – 8 of 8
Article
Publication date: 17 April 2020

Mahfooz Alam, Raza Abbas Haidri and Mohammad Shahid

Load balancing is an important issue for a heterogeneous distributed computing system environment that has been proven to be a nondeterministic polynomial time hard problem. This…

Abstract

Purpose

Load balancing is an important issue for a heterogeneous distributed computing system environment that has been proven to be a nondeterministic polynomial time hard problem. This paper aims to propose a resource-aware load balancing (REAL) model for a batch of independent tasks with a centralized load balancer to make the solution appropriate for a practical heterogeneous distributed environment having a migration cost with the objective of maximizing the level of load balancing considering bandwidth requirements for migration of the tasks.

Design/methodology/approach

To achieve the effective schedule, load balancing issues should be addressed and tackled through efficient workload distribution. In this approach, the migration has been carried out in two phases, namely, initial migration and best-fit migration. Using the best-fit policy in migrations helps in the possible performance improvement by minimizing the remaining idle slots on underloaded nodes that remain unentertained during the initial migration.

Findings

The experimental results reveal that the proposed model exhibits a superior performance among the other strategies on considered parameters such as makespan, average utilization and level of load balancing under study for a heterogeneous distributed environment.

Originality/value

Design of the REAL model and a comparative performance evaluation with LBSM and ITSLB have been conducted by using MATLAB 8.5.0.

Details

International Journal of Pervasive Computing and Communications, vol. 16 no. 2
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 5 October 2022

Sophiya Shiekh, Mohammad Shahid, Manas Sambare, Raza Abbas Haidri and Dileep Kumar Yadav

Cloud computing gives several on-demand infrastructural services by dynamically pooling heterogeneous resources to cater to users’ applications. The task scheduling needs to be…

67

Abstract

Purpose

Cloud computing gives several on-demand infrastructural services by dynamically pooling heterogeneous resources to cater to users’ applications. The task scheduling needs to be done optimally to achieve proficient results in a cloud computing environment. While satisfying the user’s requirements in a cloud environment, scheduling has been proven an NP-hard problem. Therefore, it leaves scope to develop new allocation models for the problem. The aim of the study is to develop load balancing method to maximize the resource utilization in cloud environment.

Design/methodology/approach

In this paper, the parallelized task allocation with load balancing (PTAL) hybrid heuristic is proposed for jobs coming from various users. These jobs are allocated on the resources one by one in a parallelized manner as they arrive in the cloud system. The novel algorithm works in three phases: parallelization, task allocation and task reallocation. The proposed model is designed for efficient task allocation, reallocation of resources and adequate load balancing to achieve better quality of service (QoS) results.

Findings

The acquired empirical results show that PTAL performs better than other scheduling strategies under various cases for different QoS parameters under study.

Originality/value

The outcome has been examined for the real data set to evaluate it with different state-of-the-art heuristics having comparable objective parameters.

Details

International Journal of Pervasive Computing and Communications, vol. 19 no. 5
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 22 March 2024

Mohd Mustaqeem, Suhel Mustajab and Mahfooz Alam

Software defect prediction (SDP) is a critical aspect of software quality assurance, aiming to identify and manage potential defects in software systems. In this paper, we have…

Abstract

Purpose

Software defect prediction (SDP) is a critical aspect of software quality assurance, aiming to identify and manage potential defects in software systems. In this paper, we have proposed a novel hybrid approach that combines Gray Wolf Optimization with Feature Selection (GWOFS) and multilayer perceptron (MLP) for SDP. The GWOFS-MLP hybrid model is designed to optimize feature selection, ultimately enhancing the accuracy and efficiency of SDP. Gray Wolf Optimization, inspired by the social hierarchy and hunting behavior of gray wolves, is employed to select a subset of relevant features from an extensive pool of potential predictors. This study investigates the key challenges that traditional SDP approaches encounter and proposes promising solutions to overcome time complexity and the curse of the dimensionality reduction problem.

Design/methodology/approach

The integration of GWOFS and MLP results in a robust hybrid model that can adapt to diverse software datasets. This feature selection process harnesses the cooperative hunting behavior of wolves, allowing for the exploration of critical feature combinations. The selected features are then fed into an MLP, a powerful artificial neural network (ANN) known for its capability to learn intricate patterns within software metrics. MLP serves as the predictive engine, utilizing the curated feature set to model and classify software defects accurately.

Findings

The performance evaluation of the GWOFS-MLP hybrid model on a real-world software defect dataset demonstrates its effectiveness. The model achieves a remarkable training accuracy of 97.69% and a testing accuracy of 97.99%. Additionally, the receiver operating characteristic area under the curve (ROC-AUC) score of 0.89 highlights the model’s ability to discriminate between defective and defect-free software components.

Originality/value

Experimental implementations using machine learning-based techniques with feature reduction are conducted to validate the proposed solutions. The goal is to enhance SDP’s accuracy, relevance and efficiency, ultimately improving software quality assurance processes. The confusion matrix further illustrates the model’s performance, with only a small number of false positives and false negatives.

Details

International Journal of Intelligent Computing and Cybernetics, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 15 February 2022

Minakshi Sharma, Rajneesh Kumar and Anurag Jain

During high demand for the virtualized resources in cloud environment, efficient task scheduling achieves the desired performance criteria by balancing the load in the system.

Abstract

Purpose

During high demand for the virtualized resources in cloud environment, efficient task scheduling achieves the desired performance criteria by balancing the load in the system.

Design/methodology/approach

It is a task scheduling approach used for load balancing in cloud environment. Task scheduling in such an environment is used for the task execution on a suitable resource by considering some parameters and constraints to achieve performance.

Findings

The presented mechanism is an extension of the previous proposed work quality of service (QoS)-enabled join minimum loaded queue (JMLQ) (Sharma et al., 2019c). The proposed approach has been tested in the CloudSim simulator, and the results show that the proposed approach achieves better results in comparison to QoS-enabled JMLQ and its other variants in the cloud environment.

Originality/value

90%

Details

International Journal of Pervasive Computing and Communications, vol. 19 no. 4
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 3 September 2018

Mohan Liyanage, Chii Chang and Satish Narayana Srirama

The distant data centre-centric Internet of Things (IoT) systems face the latency issue especially in the real-time-based applications, such as augmented reality, traffic…

Abstract

Purpose

The distant data centre-centric Internet of Things (IoT) systems face the latency issue especially in the real-time-based applications, such as augmented reality, traffic analytics and ambient assisted living. Recently, Fog computing models have been introduced to overcome the latency issue by using the proximity-based computational resources, such as the computers co-located with the cellular base station, grid router devices or computers in local business. However, the increasing users of Fog computing servers cause bottleneck issues and consequently the latency issue arises again. This paper aims to introduce the utilisation of Mist computing (Mist) model, which exploits the computational and networking resources from the devices at the very edge of the IoT networks.

Design/methodology/approach

This paper proposes a service-oriented mobile-embedded Platform as a Service (mePaaS) framework that allows the mobile device to provide a flexible platform for proximal users to offload their computational or networking program to mePaaS-based Mist computing node.

Findings

The prototype has been tested and performance has been evaluated on the real-world devices. The evaluation results have shown the promising nature of mePaaS.

Originality/value

The proposed framework supports resource-aware autonomous service configuration that can manage the availability of the functions provided by the Mist node based on the dynamically changing hardware resource availability. In addition, the framework also supports task distribution among a group of Mist nodes.

Details

International Journal of Pervasive Computing and Communications, vol. 14 no. 3/4
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 10 January 2023

Manpreet Singh, Urvashi Tandon and Amit Mittal

The purpose of this paper is to identify the antecedents of continued usage intentions in the connected devices ecosystem in health care by analyzing the users' and physicians'…

Abstract

Purpose

The purpose of this paper is to identify the antecedents of continued usage intentions in the connected devices ecosystem in health care by analyzing the users' and physicians' expectations in a new ecosystem where one prefers to connect digitally rather than physically.

Design/methodology/approach

This is a unique study in which data was collected from 242 doctors and 215 end-users to gauge the expectations from the connected devices in health care. Further, these responses were hypothesised using UTAUT-2 and ECT theories to analyze general users’ and professional users’ or doctors’ expectations for continued usage in connected devices ecosystem in the health-care ecosystem.

Findings

Performance expectancy, social influence, facilitating conditions and price value emerged as significant predictors of satisfaction in both user groups. But habit and hedonic motivation reflected an insignificant impact on user satisfaction. Surprisingly, effort expectancy emerged as a significant factor for end-user satisfaction, and this became insignificant for professional user satisfaction. Satisfaction was positively related to continued usage for both user groups, and app quality has a positive impact on all the predictors.

Practical implications

To the best of the authors’ knowledge, this is the first comparative study to understand the factors which influence consumer behavior leading to a holistic model and can be imbibed for creating a better customer experience in an era where we are more comfortable connecting digitally rather than physically.

Originality/value

This study has used the Unified Theory of Acceptance and Use of Technology-2 model and expectation confirmation theory to analyze the key factors influencing the intentions for continued usage of devices in the Internet of Medical Devices setup.

Details

Information Discovery and Delivery, vol. 51 no. 3
Type: Research Article
ISSN: 2398-6247

Keywords

Article
Publication date: 3 July 2020

Azra Nazir, Roohie Naaz Mir and Shaima Qureshi

The trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud…

274

Abstract

Purpose

The trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud as their resource giant. But this picture leads to underutilization of ever-increasing device pool of IoT that has already passed 15 billion mark in 2015. Thus, it is high time to explore a different approach to tackle this issue, keeping in view the characteristics and needs of the two fields. Processing at the Edge can boost applications with real-time deadlines while complementing security.

Design/methodology/approach

This review paper contributes towards three cardinal directions of research in the field of DL for IoT. The first section covers the categories of IoT devices and how Fog can aid in overcoming the underutilization of millions of devices, forming the realm of the things for IoT. The second direction handles the issue of immense computational requirements of DL models by uncovering specific compression techniques. An appropriate combination of these techniques, including regularization, quantization, and pruning, can aid in building an effective compression pipeline for establishing DL models for IoT use-cases. The third direction incorporates both these views and introduces a novel approach of parallelization for setting up a distributed systems view of DL for IoT.

Findings

DL models are growing deeper with every passing year. Well-coordinated distributed execution of such models using Fog displays a promising future for the IoT application realm. It is realized that a vertically partitioned compressed deep model can handle the trade-off between size, accuracy, communication overhead, bandwidth utilization, and latency but at the expense of an additionally considerable memory footprint. To reduce the memory budget, we propose to exploit Hashed Nets as potentially favorable candidates for distributed frameworks. However, the critical point between accuracy and size for such models needs further investigation.

Originality/value

To the best of our knowledge, no study has explored the inherent parallelism in deep neural network architectures for their efficient distribution over the Edge-Fog continuum. Besides covering techniques and frameworks that have tried to bring inference to the Edge, the review uncovers significant issues and possible future directions for endorsing deep models as processing engines for real-time IoT. The study is directed to both researchers and industrialists to take on various applications to the Edge for better user experience.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 13 no. 3
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 31 August 2021

Ibrahim Al-Shourbaji and Waleed Zogaan

The human resource (HR) allocation problem is one of the critical dimensions of the project management process. Due to this nature of the problem, researchers are continually…

Abstract

Purpose

The human resource (HR) allocation problem is one of the critical dimensions of the project management process. Due to this nature of the problem, researchers are continually optimizing one or more critical scheduling and allocation challenges in different ways. This study aims to optimize two goals, increasing customer satisfaction and reducing costs using the imperialist competitive algorithm.

Design/methodology/approach

Cloud-based e-commerce applications are preferred to conventional systems because they can save money in many areas, including resource use, running expenses, capital costs, maintenance and operation costs. In web applications, its core functionality of performance enhancement and automated device recovery is important. HR knowledge, expertise and competencies are becoming increasingly valuable carriers for organizational competitive advantage. As a result, HR management is becoming more relevant, as it seeks to channel all of the workers’ energy into meeting the organizational strategic objectives. The allocation of resources to maximize benefit or minimize cost is known as the resource allocation problem. Since discovering solutions in polynomial time is complicated, HR allocation in cloud-based e-commerce is an Nondeterministic Polynomial time (NP)-hard problem. In this paper, to promote the respective strengths and minimize the weaknesses, the imperialist competitive algorithm is suggested to solve these issues. The imperialist competitive algorithm is tested by comparing it to the literature’s novel algorithms using a simulation.

Findings

Empirical outcomes have illustrated that the suggested hybrid method achieves higher performance in discovering the appropriate HR allocation than some modern techniques.

Practical implications

The paper presents a useful method for improving HR allocation methods. The MATLAB-based simulation results have indicated that costs and waiting time have been improved compared to other algorithms, which cause the high application of this method in practical projects.

Originality/value

The main novelty of this paper is using an imperialist competitive algorithm for finding the best solution to the HR allocation problem in cloud-based e-commerce.

Details

Kybernetes, vol. 51 no. 6
Type: Research Article
ISSN: 0368-492X

Keywords

1 – 8 of 8