Search results
21 – 30 of over 35000Ivan Hanuliak and Peter Hanuliak
With the availability of powerful personal computers (PCs), workstations and networking devices, the recent trend in parallel computing is to connect a number of individual…
Abstract
Purpose
With the availability of powerful personal computers (PCs), workstations and networking devices, the recent trend in parallel computing is to connect a number of individual workstations (PC and PC symmetric multiprocessor systems (SMP)) to solve computation‐intensive tasks in parallel way on such clusters (networks of workstations (NOW), SMP and Grid). In this sense, it is not more true to consider traditionally evolved parallel computing and distributed computing as two separate research disciplines. Current trends in high performance computing are to use NOW (and SMP) as a cheaper alternative to traditionally used massively parallel multiprocessors or supercomputers and to profit from unifying of both mentioned disciplines. The purpose of this paper is to consider the individual workstations could be so single PC as parallel computers based on modern SMP implemented within workstation.
Design/methodology/approach
Such parallel systems (NOW and SMP), are connected through widely used communication standard networks and co‐operate to solve one large problem. Each workstation is threatened similarly to a processing element as in a conventional multiprocessor system. But, personal processors or multiprocessors as workstations are far more powerful and flexible than the processing elements in conventional multiprocessors. To make the whole system appear to the applications as a single parallel computing engine (a virtual parallel system), run‐time environments such as OpenMP, Java (SMP), message passing interface, Java (NOW) are used to provide an extra layer of abstraction.
Findings
To exploit the parallel processing capability of such cluster, the application program must be paralleled. The effective way how to do it for (parallelisation strategy) belongs to a most important step in developing effective parallel algorithm (optimisation). To behaviour analysis, all overheads that have the influence to performance of parallel algorithms (architecture, computation, communication, etc.) have to be taken into account. In this paper, such complex performance evaluation of iterative parallel algorithms (IPA) and their practical implementations are discussed (Jacobi and Gauss‐Seidel iteration). On real application example, the various influences in process of modelling and performance evaluation and the consequences of their distributed parallel implementations are demonstrated.
Originality/value
The paper usefully shows that better load balancing can be achieved among used network nodes (performance optimisation of parallel algorithm). Generally, it claims that the parallel algorithms or their parts (processes) with more communication (similar to analyzed Gauss‐Seidel parallel algorithm) will have better speed‐up values using modern SMP parallel system as its parallel implementation in NOW. For the algorithms or processes with small communication overheads (similar to analysed Jacobi parallel algorithm) the other network nodes can be used based on single processors.
Details
Keywords
Mahfooz Alam, Mahak, Raza Abbas Haidri and Dileep Kumar Yadav
Cloud users can access services at anytime from anywhere in the world. On average, Google now processes more than 40,000 searches every second, which is approximately 3.5 billion…
Abstract
Purpose
Cloud users can access services at anytime from anywhere in the world. On average, Google now processes more than 40,000 searches every second, which is approximately 3.5 billion searches per day. The diverse and vast amounts of data are generated with the development of next-generation information technologies such as cryptocurrency, internet of things and big data. To execute such applications, it is needed to design an efficient scheduling algorithm that considers the quality of service parameters like utilization, makespan and response time. Therefore, this paper aims to propose a novel Efficient Static Task Allocation (ESTA) algorithm, which optimizes average utilization.
Design/methodology/approach
Cloud computing provides resources such as virtual machine, network, storage, etc. over the internet. Cloud computing follows the pay-per-use billing model. To achieve efficient task allocation, scheduling algorithm problems should be interacted and tackled through efficient task distribution on the resources. The methodology of ESTA algorithm is based on minimum completion time approach. ESTA intelligently maps the batch of independent tasks (cloudlets) on heterogeneous virtual machines and optimizes their utilization in infrastructure as a service cloud computing.
Findings
To evaluate the performance of ESTA, the simulation study is compared with Min-Min, load balancing strategy with migration cost, Longest job in the fastest resource-shortest job in the fastest resource, sufferage, minimum completion time (MCT), minimum execution time and opportunistic load balancing on account of makespan, utilization and response time.
Originality/value
The simulation result reveals that the ESTA algorithm consistently superior performs under varying of batch independent of cloudlets and the number of virtual machines’ test conditions.
Details
Keywords
Fei Luo, Hai Jin, Xiaofei Liao and Qin Zhang
Peer‐to‐peer (P2P) communities have the capability to construct a powerful virtual supercomputer by assembling idle internet cycles. The purpose of this paper is to present the…
Abstract
Purpose
Peer‐to‐peer (P2P) communities have the capability to construct a powerful virtual supercomputer by assembling idle internet cycles. The purpose of this paper is to present the scheduling issues in an unstructured P2P‐based high performance computing (HPC) system to achieve high performance for applications.
Design/methodology/approach
A new application model is proposed for the system, where applications are parallelized in the program level. To address high performance for these applications, the system resources are controlled in a semi‐centralized 3‐layer network, where volunteers form many autonomous unstructured P2P domains. Furthermore, based on such a resource management policy, a job scheduling strategy is adopted, which is collaborated by global and domain scheduling. The global scheduling is responsible for the balance among domains, while the domain scheduling resolve workpiles' execution in a domain.
Findings
Theoretical analysis and a benchmark experiment show that the scheduling provides scalable and enormous computing capability in the P2P‐based HPC system.
Originality/value
The paper shows that scheduling helps P2HP (an unstructured P2P‐based HPC system) provide scalable and enormous computing capability for HPC applications.
Details
Keywords
Emmanuel Imuetinyan Aghimien, Lerato Millicent Aghimien, Olutomilayo Olayemi Petinrin and Douglas Omoregie Aghimien
This paper aims to present the result of a scientometric analysis conducted using studies on high-performance computing in computational modelling. This was done with a view to…
Abstract
Purpose
This paper aims to present the result of a scientometric analysis conducted using studies on high-performance computing in computational modelling. This was done with a view to showcasing the need for high-performance computers (HPC) within the architecture, engineering and construction (AEC) industry in developing countries, particularly in Africa, where the use of HPC in developing computational models (CMs) for effective problem solving is still low.
Design/methodology/approach
An interpretivism philosophical stance was adopted for the study which informed a scientometric review of existing studies gathered from the Scopus database. Keywords such as high-performance computing, and computational modelling were used to extract papers from the database. Visualisation of Similarities viewer (VOSviewer) was used to prepare co-occurrence maps based on the bibliographic data gathered.
Findings
Findings revealed the scarcity of research emanating from Africa in this area of study. Furthermore, past studies had placed focus on high-performance computing in the development of computational modelling and theory, parallel computing and improved visualisation, large-scale application software, computer simulations and computational mathematical modelling. Future studies can also explore areas such as cloud computing, optimisation, high-level programming language, natural science computing, computer graphics equipment and Graphics Processing Units as they relate to the AEC industry.
Research limitations/implications
The study assessed a single database for the search of related studies.
Originality/value
The findings of this study serve as an excellent theoretical background for AEC researchers seeking to explore the use of HPC for CMs development in the quest for solving complex problems in the industry.
Details
Keywords
Azra Nazir, Roohie Naaz Mir and Shaima Qureshi
The trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud…
Abstract
Purpose
The trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud as their resource giant. But this picture leads to underutilization of ever-increasing device pool of IoT that has already passed 15 billion mark in 2015. Thus, it is high time to explore a different approach to tackle this issue, keeping in view the characteristics and needs of the two fields. Processing at the Edge can boost applications with real-time deadlines while complementing security.
Design/methodology/approach
This review paper contributes towards three cardinal directions of research in the field of DL for IoT. The first section covers the categories of IoT devices and how Fog can aid in overcoming the underutilization of millions of devices, forming the realm of the things for IoT. The second direction handles the issue of immense computational requirements of DL models by uncovering specific compression techniques. An appropriate combination of these techniques, including regularization, quantization, and pruning, can aid in building an effective compression pipeline for establishing DL models for IoT use-cases. The third direction incorporates both these views and introduces a novel approach of parallelization for setting up a distributed systems view of DL for IoT.
Findings
DL models are growing deeper with every passing year. Well-coordinated distributed execution of such models using Fog displays a promising future for the IoT application realm. It is realized that a vertically partitioned compressed deep model can handle the trade-off between size, accuracy, communication overhead, bandwidth utilization, and latency but at the expense of an additionally considerable memory footprint. To reduce the memory budget, we propose to exploit Hashed Nets as potentially favorable candidates for distributed frameworks. However, the critical point between accuracy and size for such models needs further investigation.
Originality/value
To the best of our knowledge, no study has explored the inherent parallelism in deep neural network architectures for their efficient distribution over the Edge-Fog continuum. Besides covering techniques and frameworks that have tried to bring inference to the Edge, the review uncovers significant issues and possible future directions for endorsing deep models as processing engines for real-time IoT. The study is directed to both researchers and industrialists to take on various applications to the Edge for better user experience.
Details
Keywords
Introduction: With many new technologies requiring real-time data processing, cloud computing has become challenging to implement due to high bandwidth and high latency…
Abstract
Introduction: With many new technologies requiring real-time data processing, cloud computing has become challenging to implement due to high bandwidth and high latency requirements.
Purpose: To overcome this issue, edge computing is used to process data at the network’s edge. Edge computing is a distributed computing paradigm that brings computation and data storage closer to the location where it is needed. It is used to process time-sensitive data.
Methodology: The authors implemented the model using Linux Foundation’s open-source platform EdgeX Foundry to create an edge-computing device. The model involved getting data from an on-board sensor (on-board diagnostics (OBD-II)) and the GPS sensor of a car. The data are then observed and computed to the EdgeX server. The single server will send data to serve three real-life internet of things (IoT) use cases: auto insurance, supporting a smart city, and building a personal driving record.
Findings: The main aim of this model is to illustrate how edge computing can improve both latency and bandwidth usage needed for real-world IoT applications.
Details
Keywords
Nick Antonopoulos, Kyriakos Koukoumpetsos and Alex Shafarenko
The mobile software agent paradigm provides a generic, customisable foundation for the development of high performance distriubuted applications. An efficient, general‐purpose…
Abstract
The mobile software agent paradigm provides a generic, customisable foundation for the development of high performance distriubuted applications. An efficient, general‐purpose access control mechanism is required to support the development of a wide range of applications. This is achievable if the design of the access control system is based on the principles of simplicity, programmability (customisation) and reusability. However, existing mobile agent architectures either neglect this issue, or offer centralised schemes that do not support adaptive access control on a per‐agent basis and do not address the issues of secure knowledge sharing and reusing. In this paper a simple, distributed access control architecture is presented, based on the concept of distributed, active authorisation entities (lock cells), any combination of which can be referenced by an agent to provide input and/or output access control. It is demonstrated how these lock cells can be used to implement security domains and how they can be combined to create composite lock cells.
Details
Keywords
Shengpei Zhou, Zhenting Chang, Haina Song, Yuejiang Su, Xiaosong Liu and Jingfeng Yang
With the continuous technological development of automated driving and expansion of its application scope, the types of on-board equipment continue to be enriched and the computing…
Abstract
Purpose
With the continuous technological development of automated driving and expansion of its application scope, the types of on-board equipment continue to be enriched and the computing capabilities of on-board equipment continue to increase and corresponding applications become more diverse. As the applications need to run on on-board equipment, the requirements for the computing capabilities of on-board equipment become higher. Mobile edge computing is one of the effective methods to solve practical application problems in automated driving.
Design/methodology/approach
In this study, in accordance with practical requirements, this paper proposed an optimal resource management allocation method of autonomous-vehicle-infrastructure cooperation in a mobile edge computing environment and conducted an experiment in practical application.
Findings
The design of the road-side unit module and its corresponding real-time operating system task coordination in edge computing are proposed in the study, as well as the method for edge computing load integration and heterogeneous computing. Then, the real-time scheduling of highly concurrent computation tasks, adaptive computation task migration method and edge server collaborative resource allocation method is proposed. Test results indicate that the method proposed in this study can greatly reduce the task computing delay, and the power consumption generally increases with the increase of task size and task complexity.
Originality/value
The results showed that the proposed method can achieve lower power consumption and lower computational overhead while ensuring the quality of service for users, indicating a great application prospect of the method.
Details
Keywords
Stephan Olariu, Ismail Khalil and Mahmoud Abuelela
The past decade has witnessed a growing interest in vehicular networking and its myriad applications. The initial view of practitioners and researchers was that radio‐equipped…
Abstract
Purpose
The past decade has witnessed a growing interest in vehicular networking and its myriad applications. The initial view of practitioners and researchers was that radio‐equipped vehicles can keep the drivers informed about potential safety risks and can enhance their awareness of road conditions and traffic‐related events. This conceptual paper seeks to put forth a novel vision, namely that advances in vehicular networks, embedded devices, and cloud computing can be used to set up what are known as vehicular clouds (VCs).
Design/methodology/approach
The paper suggests that VCs are technologically feasible and that they are likely to have a significant societal impact.
Findings
The paper argues that at least in some of its manifestations, the ideas behind VCs are eminently implementable under present day technology. It is also expected that, once adopted and championed by municipalities and third‐party infrastructure providers, VCs will redefine the way in which pervasive computing and its myriad applications is thought of.
Research limitations/implications
This is a new concept for which a small‐scale prototype is being built. No large‐scale prototype exists at the moment.
Practical implications
VCs are a novel concept motivated by the realization of the fact that, most of the time, the tremendous amount of computing and communication resources available in vehicles is underutilized. Putting these resources to work in a meaningful way should have a significant societal impact.
Social implications
The main goal of this paper is to introduce and promote the concept of VCs, a non‐trivial extension, along several dimensions, of the by‐now “classic” cloud computing. The paper shows that the concept of VCs is feasible as of today – at least in some of its manifestations – and that it can have a significant societal impact in the years to come.
Originality/value
The idea of a VC is novel and so are the potential applications that are discussed in the paper.
Details