Search results
1 – 8 of 8Mahfooz Alam, Raza Abbas Haidri and Mohammad Shahid
Load balancing is an important issue for a heterogeneous distributed computing system environment that has been proven to be a nondeterministic polynomial time hard problem. This…
Abstract
Purpose
Load balancing is an important issue for a heterogeneous distributed computing system environment that has been proven to be a nondeterministic polynomial time hard problem. This paper aims to propose a resource-aware load balancing (REAL) model for a batch of independent tasks with a centralized load balancer to make the solution appropriate for a practical heterogeneous distributed environment having a migration cost with the objective of maximizing the level of load balancing considering bandwidth requirements for migration of the tasks.
Design/methodology/approach
To achieve the effective schedule, load balancing issues should be addressed and tackled through efficient workload distribution. In this approach, the migration has been carried out in two phases, namely, initial migration and best-fit migration. Using the best-fit policy in migrations helps in the possible performance improvement by minimizing the remaining idle slots on underloaded nodes that remain unentertained during the initial migration.
Findings
The experimental results reveal that the proposed model exhibits a superior performance among the other strategies on considered parameters such as makespan, average utilization and level of load balancing under study for a heterogeneous distributed environment.
Originality/value
Design of the REAL model and a comparative performance evaluation with LBSM and ITSLB have been conducted by using MATLAB 8.5.0.
Details
Keywords
Sophiya Shiekh, Mohammad Shahid, Manas Sambare, Raza Abbas Haidri and Dileep Kumar Yadav
Cloud computing gives several on-demand infrastructural services by dynamically pooling heterogeneous resources to cater to users’ applications. The task scheduling needs to be…
Abstract
Purpose
Cloud computing gives several on-demand infrastructural services by dynamically pooling heterogeneous resources to cater to users’ applications. The task scheduling needs to be done optimally to achieve proficient results in a cloud computing environment. While satisfying the user’s requirements in a cloud environment, scheduling has been proven an NP-hard problem. Therefore, it leaves scope to develop new allocation models for the problem. The aim of the study is to develop load balancing method to maximize the resource utilization in cloud environment.
Design/methodology/approach
In this paper, the parallelized task allocation with load balancing (PTAL) hybrid heuristic is proposed for jobs coming from various users. These jobs are allocated on the resources one by one in a parallelized manner as they arrive in the cloud system. The novel algorithm works in three phases: parallelization, task allocation and task reallocation. The proposed model is designed for efficient task allocation, reallocation of resources and adequate load balancing to achieve better quality of service (QoS) results.
Findings
The acquired empirical results show that PTAL performs better than other scheduling strategies under various cases for different QoS parameters under study.
Originality/value
The outcome has been examined for the real data set to evaluate it with different state-of-the-art heuristics having comparable objective parameters.
Details
Keywords
Mohd Mustaqeem, Suhel Mustajab and Mahfooz Alam
Software defect prediction (SDP) is a critical aspect of software quality assurance, aiming to identify and manage potential defects in software systems. In this paper, we have…
Abstract
Purpose
Software defect prediction (SDP) is a critical aspect of software quality assurance, aiming to identify and manage potential defects in software systems. In this paper, we have proposed a novel hybrid approach that combines Gray Wolf Optimization with Feature Selection (GWOFS) and multilayer perceptron (MLP) for SDP. The GWOFS-MLP hybrid model is designed to optimize feature selection, ultimately enhancing the accuracy and efficiency of SDP. Gray Wolf Optimization, inspired by the social hierarchy and hunting behavior of gray wolves, is employed to select a subset of relevant features from an extensive pool of potential predictors. This study investigates the key challenges that traditional SDP approaches encounter and proposes promising solutions to overcome time complexity and the curse of the dimensionality reduction problem.
Design/methodology/approach
The integration of GWOFS and MLP results in a robust hybrid model that can adapt to diverse software datasets. This feature selection process harnesses the cooperative hunting behavior of wolves, allowing for the exploration of critical feature combinations. The selected features are then fed into an MLP, a powerful artificial neural network (ANN) known for its capability to learn intricate patterns within software metrics. MLP serves as the predictive engine, utilizing the curated feature set to model and classify software defects accurately.
Findings
The performance evaluation of the GWOFS-MLP hybrid model on a real-world software defect dataset demonstrates its effectiveness. The model achieves a remarkable training accuracy of 97.69% and a testing accuracy of 97.99%. Additionally, the receiver operating characteristic area under the curve (ROC-AUC) score of 0.89 highlights the model’s ability to discriminate between defective and defect-free software components.
Originality/value
Experimental implementations using machine learning-based techniques with feature reduction are conducted to validate the proposed solutions. The goal is to enhance SDP’s accuracy, relevance and efficiency, ultimately improving software quality assurance processes. The confusion matrix further illustrates the model’s performance, with only a small number of false positives and false negatives.
Details
Keywords
Minakshi Sharma, Rajneesh Kumar and Anurag Jain
During high demand for the virtualized resources in cloud environment, efficient task scheduling achieves the desired performance criteria by balancing the load in the system.
Abstract
Purpose
During high demand for the virtualized resources in cloud environment, efficient task scheduling achieves the desired performance criteria by balancing the load in the system.
Design/methodology/approach
It is a task scheduling approach used for load balancing in cloud environment. Task scheduling in such an environment is used for the task execution on a suitable resource by considering some parameters and constraints to achieve performance.
Findings
The presented mechanism is an extension of the previous proposed work quality of service (QoS)-enabled join minimum loaded queue (JMLQ) (Sharma et al., 2019c). The proposed approach has been tested in the CloudSim simulator, and the results show that the proposed approach achieves better results in comparison to QoS-enabled JMLQ and its other variants in the cloud environment.
Originality/value
90%
Details
Keywords
Mohan Liyanage, Chii Chang and Satish Narayana Srirama
The distant data centre-centric Internet of Things (IoT) systems face the latency issue especially in the real-time-based applications, such as augmented reality, traffic…
Abstract
Purpose
The distant data centre-centric Internet of Things (IoT) systems face the latency issue especially in the real-time-based applications, such as augmented reality, traffic analytics and ambient assisted living. Recently, Fog computing models have been introduced to overcome the latency issue by using the proximity-based computational resources, such as the computers co-located with the cellular base station, grid router devices or computers in local business. However, the increasing users of Fog computing servers cause bottleneck issues and consequently the latency issue arises again. This paper aims to introduce the utilisation of Mist computing (Mist) model, which exploits the computational and networking resources from the devices at the very edge of the IoT networks.
Design/methodology/approach
This paper proposes a service-oriented mobile-embedded Platform as a Service (mePaaS) framework that allows the mobile device to provide a flexible platform for proximal users to offload their computational or networking program to mePaaS-based Mist computing node.
Findings
The prototype has been tested and performance has been evaluated on the real-world devices. The evaluation results have shown the promising nature of mePaaS.
Originality/value
The proposed framework supports resource-aware autonomous service configuration that can manage the availability of the functions provided by the Mist node based on the dynamically changing hardware resource availability. In addition, the framework also supports task distribution among a group of Mist nodes.
Details
Keywords
Manpreet Singh, Urvashi Tandon and Amit Mittal
The purpose of this paper is to identify the antecedents of continued usage intentions in the connected devices ecosystem in health care by analyzing the users' and physicians'…
Abstract
Purpose
The purpose of this paper is to identify the antecedents of continued usage intentions in the connected devices ecosystem in health care by analyzing the users' and physicians' expectations in a new ecosystem where one prefers to connect digitally rather than physically.
Design/methodology/approach
This is a unique study in which data was collected from 242 doctors and 215 end-users to gauge the expectations from the connected devices in health care. Further, these responses were hypothesised using UTAUT-2 and ECT theories to analyze general users’ and professional users’ or doctors’ expectations for continued usage in connected devices ecosystem in the health-care ecosystem.
Findings
Performance expectancy, social influence, facilitating conditions and price value emerged as significant predictors of satisfaction in both user groups. But habit and hedonic motivation reflected an insignificant impact on user satisfaction. Surprisingly, effort expectancy emerged as a significant factor for end-user satisfaction, and this became insignificant for professional user satisfaction. Satisfaction was positively related to continued usage for both user groups, and app quality has a positive impact on all the predictors.
Practical implications
To the best of the authors’ knowledge, this is the first comparative study to understand the factors which influence consumer behavior leading to a holistic model and can be imbibed for creating a better customer experience in an era where we are more comfortable connecting digitally rather than physically.
Originality/value
This study has used the Unified Theory of Acceptance and Use of Technology-2 model and expectation confirmation theory to analyze the key factors influencing the intentions for continued usage of devices in the Internet of Medical Devices setup.
Details
Keywords
Azra Nazir, Roohie Naaz Mir and Shaima Qureshi
The trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud…
Abstract
Purpose
The trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud as their resource giant. But this picture leads to underutilization of ever-increasing device pool of IoT that has already passed 15 billion mark in 2015. Thus, it is high time to explore a different approach to tackle this issue, keeping in view the characteristics and needs of the two fields. Processing at the Edge can boost applications with real-time deadlines while complementing security.
Design/methodology/approach
This review paper contributes towards three cardinal directions of research in the field of DL for IoT. The first section covers the categories of IoT devices and how Fog can aid in overcoming the underutilization of millions of devices, forming the realm of the things for IoT. The second direction handles the issue of immense computational requirements of DL models by uncovering specific compression techniques. An appropriate combination of these techniques, including regularization, quantization, and pruning, can aid in building an effective compression pipeline for establishing DL models for IoT use-cases. The third direction incorporates both these views and introduces a novel approach of parallelization for setting up a distributed systems view of DL for IoT.
Findings
DL models are growing deeper with every passing year. Well-coordinated distributed execution of such models using Fog displays a promising future for the IoT application realm. It is realized that a vertically partitioned compressed deep model can handle the trade-off between size, accuracy, communication overhead, bandwidth utilization, and latency but at the expense of an additionally considerable memory footprint. To reduce the memory budget, we propose to exploit Hashed Nets as potentially favorable candidates for distributed frameworks. However, the critical point between accuracy and size for such models needs further investigation.
Originality/value
To the best of our knowledge, no study has explored the inherent parallelism in deep neural network architectures for their efficient distribution over the Edge-Fog continuum. Besides covering techniques and frameworks that have tried to bring inference to the Edge, the review uncovers significant issues and possible future directions for endorsing deep models as processing engines for real-time IoT. The study is directed to both researchers and industrialists to take on various applications to the Edge for better user experience.
Details
Keywords
Ibrahim Al-Shourbaji and Waleed Zogaan
The human resource (HR) allocation problem is one of the critical dimensions of the project management process. Due to this nature of the problem, researchers are continually…
Abstract
Purpose
The human resource (HR) allocation problem is one of the critical dimensions of the project management process. Due to this nature of the problem, researchers are continually optimizing one or more critical scheduling and allocation challenges in different ways. This study aims to optimize two goals, increasing customer satisfaction and reducing costs using the imperialist competitive algorithm.
Design/methodology/approach
Cloud-based e-commerce applications are preferred to conventional systems because they can save money in many areas, including resource use, running expenses, capital costs, maintenance and operation costs. In web applications, its core functionality of performance enhancement and automated device recovery is important. HR knowledge, expertise and competencies are becoming increasingly valuable carriers for organizational competitive advantage. As a result, HR management is becoming more relevant, as it seeks to channel all of the workers’ energy into meeting the organizational strategic objectives. The allocation of resources to maximize benefit or minimize cost is known as the resource allocation problem. Since discovering solutions in polynomial time is complicated, HR allocation in cloud-based e-commerce is an Nondeterministic Polynomial time (NP)-hard problem. In this paper, to promote the respective strengths and minimize the weaknesses, the imperialist competitive algorithm is suggested to solve these issues. The imperialist competitive algorithm is tested by comparing it to the literature’s novel algorithms using a simulation.
Findings
Empirical outcomes have illustrated that the suggested hybrid method achieves higher performance in discovering the appropriate HR allocation than some modern techniques.
Practical implications
The paper presents a useful method for improving HR allocation methods. The MATLAB-based simulation results have indicated that costs and waiting time have been improved compared to other algorithms, which cause the high application of this method in practical projects.
Originality/value
The main novelty of this paper is using an imperialist competitive algorithm for finding the best solution to the HR allocation problem in cloud-based e-commerce.
Details