Search results

21 – 30 of over 35000
Article
Publication date: 16 March 2010

Ivan Hanuliak and Peter Hanuliak

With the availability of powerful personal computers (PCs), workstations and networking devices, the recent trend in parallel computing is to connect a number of individual…

Abstract

Purpose

With the availability of powerful personal computers (PCs), workstations and networking devices, the recent trend in parallel computing is to connect a number of individual workstations (PC and PC symmetric multiprocessor systems (SMP)) to solve computation‐intensive tasks in parallel way on such clusters (networks of workstations (NOW), SMP and Grid). In this sense, it is not more true to consider traditionally evolved parallel computing and distributed computing as two separate research disciplines. Current trends in high performance computing are to use NOW (and SMP) as a cheaper alternative to traditionally used massively parallel multiprocessors or supercomputers and to profit from unifying of both mentioned disciplines. The purpose of this paper is to consider the individual workstations could be so single PC as parallel computers based on modern SMP implemented within workstation.

Design/methodology/approach

Such parallel systems (NOW and SMP), are connected through widely used communication standard networks and co‐operate to solve one large problem. Each workstation is threatened similarly to a processing element as in a conventional multiprocessor system. But, personal processors or multiprocessors as workstations are far more powerful and flexible than the processing elements in conventional multiprocessors. To make the whole system appear to the applications as a single parallel computing engine (a virtual parallel system), run‐time environments such as OpenMP, Java (SMP), message passing interface, Java (NOW) are used to provide an extra layer of abstraction.

Findings

To exploit the parallel processing capability of such cluster, the application program must be paralleled. The effective way how to do it for (parallelisation strategy) belongs to a most important step in developing effective parallel algorithm (optimisation). To behaviour analysis, all overheads that have the influence to performance of parallel algorithms (architecture, computation, communication, etc.) have to be taken into account. In this paper, such complex performance evaluation of iterative parallel algorithms (IPA) and their practical implementations are discussed (Jacobi and Gauss‐Seidel iteration). On real application example, the various influences in process of modelling and performance evaluation and the consequences of their distributed parallel implementations are demonstrated.

Originality/value

The paper usefully shows that better load balancing can be achieved among used network nodes (performance optimisation of parallel algorithm). Generally, it claims that the parallel algorithms or their parts (processes) with more communication (similar to analyzed Gauss‐Seidel parallel algorithm) will have better speed‐up values using modern SMP parallel system as its parallel implementation in NOW. For the algorithms or processes with small communication overheads (similar to analysed Jacobi parallel algorithm) the other network nodes can be used based on single processors.

Article
Publication date: 17 September 2020

Mahfooz Alam, Mahak, Raza Abbas Haidri and Dileep Kumar Yadav

Cloud users can access services at anytime from anywhere in the world. On average, Google now processes more than 40,000 searches every second, which is approximately 3.5 billion…

Abstract

Purpose

Cloud users can access services at anytime from anywhere in the world. On average, Google now processes more than 40,000 searches every second, which is approximately 3.5 billion searches per day. The diverse and vast amounts of data are generated with the development of next-generation information technologies such as cryptocurrency, internet of things and big data. To execute such applications, it is needed to design an efficient scheduling algorithm that considers the quality of service parameters like utilization, makespan and response time. Therefore, this paper aims to propose a novel Efficient Static Task Allocation (ESTA) algorithm, which optimizes average utilization.

Design/methodology/approach

Cloud computing provides resources such as virtual machine, network, storage, etc. over the internet. Cloud computing follows the pay-per-use billing model. To achieve efficient task allocation, scheduling algorithm problems should be interacted and tackled through efficient task distribution on the resources. The methodology of ESTA algorithm is based on minimum completion time approach. ESTA intelligently maps the batch of independent tasks (cloudlets) on heterogeneous virtual machines and optimizes their utilization in infrastructure as a service cloud computing.

Findings

To evaluate the performance of ESTA, the simulation study is compared with Min-Min, load balancing strategy with migration cost, Longest job in the fastest resource-shortest job in the fastest resource, sufferage, minimum completion time (MCT), minimum execution time and opportunistic load balancing on account of makespan, utilization and response time.

Originality/value

The simulation result reveals that the ESTA algorithm consistently superior performs under varying of batch independent of cloudlets and the number of virtual machines’ test conditions.

Details

International Journal of Pervasive Computing and Communications, vol. 17 no. 3
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 12 October 2012

Fei Luo, Hai Jin, Xiaofei Liao and Qin Zhang

Peer‐to‐peer (P2P) communities have the capability to construct a powerful virtual supercomputer by assembling idle internet cycles. The purpose of this paper is to present the…

Abstract

Purpose

Peer‐to‐peer (P2P) communities have the capability to construct a powerful virtual supercomputer by assembling idle internet cycles. The purpose of this paper is to present the scheduling issues in an unstructured P2P‐based high performance computing (HPC) system to achieve high performance for applications.

Design/methodology/approach

A new application model is proposed for the system, where applications are parallelized in the program level. To address high performance for these applications, the system resources are controlled in a semi‐centralized 3‐layer network, where volunteers form many autonomous unstructured P2P domains. Furthermore, based on such a resource management policy, a job scheduling strategy is adopted, which is collaborated by global and domain scheduling. The global scheduling is responsible for the balance among domains, while the domain scheduling resolve workpiles' execution in a domain.

Findings

Theoretical analysis and a benchmark experiment show that the scheduling provides scalable and enormous computing capability in the P2P‐based HPC system.

Originality/value

The paper shows that scheduling helps P2HP (an unstructured P2P‐based HPC system) provide scalable and enormous computing capability for HPC applications.

Article
Publication date: 2 November 2020

Emmanuel Imuetinyan Aghimien, Lerato Millicent Aghimien, Olutomilayo Olayemi Petinrin and Douglas Omoregie Aghimien

This paper aims to present the result of a scientometric analysis conducted using studies on high-performance computing in computational modelling. This was done with a view to…

Abstract

Purpose

This paper aims to present the result of a scientometric analysis conducted using studies on high-performance computing in computational modelling. This was done with a view to showcasing the need for high-performance computers (HPC) within the architecture, engineering and construction (AEC) industry in developing countries, particularly in Africa, where the use of HPC in developing computational models (CMs) for effective problem solving is still low.

Design/methodology/approach

An interpretivism philosophical stance was adopted for the study which informed a scientometric review of existing studies gathered from the Scopus database. Keywords such as high-performance computing, and computational modelling were used to extract papers from the database. Visualisation of Similarities viewer (VOSviewer) was used to prepare co-occurrence maps based on the bibliographic data gathered.

Findings

Findings revealed the scarcity of research emanating from Africa in this area of study. Furthermore, past studies had placed focus on high-performance computing in the development of computational modelling and theory, parallel computing and improved visualisation, large-scale application software, computer simulations and computational mathematical modelling. Future studies can also explore areas such as cloud computing, optimisation, high-level programming language, natural science computing, computer graphics equipment and Graphics Processing Units as they relate to the AEC industry.

Research limitations/implications

The study assessed a single database for the search of related studies.

Originality/value

The findings of this study serve as an excellent theoretical background for AEC researchers seeking to explore the use of HPC for CMs development in the quest for solving complex problems in the industry.

Details

Journal of Engineering, Design and Technology , vol. 19 no. 5
Type: Research Article
ISSN: 1726-0531

Keywords

Abstract

Details

Marketing in Customer Technology Environments
Type: Book
ISBN: 978-1-83909-601-3

Article
Publication date: 3 July 2020

Azra Nazir, Roohie Naaz Mir and Shaima Qureshi

The trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud…

282

Abstract

Purpose

The trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud as their resource giant. But this picture leads to underutilization of ever-increasing device pool of IoT that has already passed 15 billion mark in 2015. Thus, it is high time to explore a different approach to tackle this issue, keeping in view the characteristics and needs of the two fields. Processing at the Edge can boost applications with real-time deadlines while complementing security.

Design/methodology/approach

This review paper contributes towards three cardinal directions of research in the field of DL for IoT. The first section covers the categories of IoT devices and how Fog can aid in overcoming the underutilization of millions of devices, forming the realm of the things for IoT. The second direction handles the issue of immense computational requirements of DL models by uncovering specific compression techniques. An appropriate combination of these techniques, including regularization, quantization, and pruning, can aid in building an effective compression pipeline for establishing DL models for IoT use-cases. The third direction incorporates both these views and introduces a novel approach of parallelization for setting up a distributed systems view of DL for IoT.

Findings

DL models are growing deeper with every passing year. Well-coordinated distributed execution of such models using Fog displays a promising future for the IoT application realm. It is realized that a vertically partitioned compressed deep model can handle the trade-off between size, accuracy, communication overhead, bandwidth utilization, and latency but at the expense of an additionally considerable memory footprint. To reduce the memory budget, we propose to exploit Hashed Nets as potentially favorable candidates for distributed frameworks. However, the critical point between accuracy and size for such models needs further investigation.

Originality/value

To the best of our knowledge, no study has explored the inherent parallelism in deep neural network architectures for their efficient distribution over the Edge-Fog continuum. Besides covering techniques and frameworks that have tried to bring inference to the Edge, the review uncovers significant issues and possible future directions for endorsing deep models as processing engines for real-time IoT. The study is directed to both researchers and industrialists to take on various applications to the Edge for better user experience.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 13 no. 3
Type: Research Article
ISSN: 1756-378X

Keywords

Book part
Publication date: 18 July 2022

Kamal Gulati and Pallavi Seth

Introduction: With many new technologies requiring real-time data processing, cloud computing has become challenging to implement due to high bandwidth and high latency…

Abstract

Introduction: With many new technologies requiring real-time data processing, cloud computing has become challenging to implement due to high bandwidth and high latency requirements.

Purpose: To overcome this issue, edge computing is used to process data at the network’s edge. Edge computing is a distributed computing paradigm that brings computation and data storage closer to the location where it is needed. It is used to process time-sensitive data.

Methodology: The authors implemented the model using Linux Foundation’s open-source platform EdgeX Foundry to create an edge-computing device. The model involved getting data from an on-board sensor (on-board diagnostics (OBD-II)) and the GPS sensor of a car. The data are then observed and computed to the EdgeX server. The single server will send data to serve three real-life internet of things (IoT) use cases: auto insurance, supporting a smart city, and building a personal driving record.

Findings: The main aim of this model is to illustrate how edge computing can improve both latency and bandwidth usage needed for real-world IoT applications.

Details

Big Data Analytics in the Insurance Market
Type: Book
ISBN: 978-1-80262-638-4

Keywords

Article
Publication date: 1 March 2001

Nick Antonopoulos, Kyriakos Koukoumpetsos and Alex Shafarenko

The mobile software agent paradigm provides a generic, customisable foundation for the development of high performance distriubuted applications. An efficient, general‐purpose…

Abstract

The mobile software agent paradigm provides a generic, customisable foundation for the development of high performance distriubuted applications. An efficient, general‐purpose access control mechanism is required to support the development of a wide range of applications. This is achievable if the design of the access control system is based on the principles of simplicity, programmability (customisation) and reusability. However, existing mobile agent architectures either neglect this issue, or offer centralised schemes that do not support adaptive access control on a per‐agent basis and do not address the issues of secure knowledge sharing and reusing. In this paper a simple, distributed access control architecture is presented, based on the concept of distributed, active authorisation entities (lock cells), any combination of which can be referenced by an agent to provide input and/or output access control. It is demonstrated how these lock cells can be used to implement security domains and how they can be combined to create composite lock cells.

Details

Internet Research, vol. 11 no. 1
Type: Research Article
ISSN: 1066-2243

Keywords

Article
Publication date: 14 June 2021

Shengpei Zhou, Zhenting Chang, Haina Song, Yuejiang Su, Xiaosong Liu and Jingfeng Yang

With the continuous technological development of automated driving and expansion of its application scope, the types of on-board equipment continue to be enriched and the computing

Abstract

Purpose

With the continuous technological development of automated driving and expansion of its application scope, the types of on-board equipment continue to be enriched and the computing capabilities of on-board equipment continue to increase and corresponding applications become more diverse. As the applications need to run on on-board equipment, the requirements for the computing capabilities of on-board equipment become higher. Mobile edge computing is one of the effective methods to solve practical application problems in automated driving.

Design/methodology/approach

In this study, in accordance with practical requirements, this paper proposed an optimal resource management allocation method of autonomous-vehicle-infrastructure cooperation in a mobile edge computing environment and conducted an experiment in practical application.

Findings

The design of the road-side unit module and its corresponding real-time operating system task coordination in edge computing are proposed in the study, as well as the method for edge computing load integration and heterogeneous computing. Then, the real-time scheduling of highly concurrent computation tasks, adaptive computation task migration method and edge server collaborative resource allocation method is proposed. Test results indicate that the method proposed in this study can greatly reduce the task computing delay, and the power consumption generally increases with the increase of task size and task complexity.

Originality/value

The results showed that the proposed method can achieve lower power consumption and lower computational overhead while ensuring the quality of service for users, indicating a great application prospect of the method.

Details

Assembly Automation, vol. 41 no. 3
Type: Research Article
ISSN: 0144-5154

Keywords

Article
Publication date: 5 April 2011

Stephan Olariu, Ismail Khalil and Mahmoud Abuelela

The past decade has witnessed a growing interest in vehicular networking and its myriad applications. The initial view of practitioners and researchers was that radio‐equipped…

1730

Abstract

Purpose

The past decade has witnessed a growing interest in vehicular networking and its myriad applications. The initial view of practitioners and researchers was that radio‐equipped vehicles can keep the drivers informed about potential safety risks and can enhance their awareness of road conditions and traffic‐related events. This conceptual paper seeks to put forth a novel vision, namely that advances in vehicular networks, embedded devices, and cloud computing can be used to set up what are known as vehicular clouds (VCs).

Design/methodology/approach

The paper suggests that VCs are technologically feasible and that they are likely to have a significant societal impact.

Findings

The paper argues that at least in some of its manifestations, the ideas behind VCs are eminently implementable under present day technology. It is also expected that, once adopted and championed by municipalities and third‐party infrastructure providers, VCs will redefine the way in which pervasive computing and its myriad applications is thought of.

Research limitations/implications

This is a new concept for which a small‐scale prototype is being built. No large‐scale prototype exists at the moment.

Practical implications

VCs are a novel concept motivated by the realization of the fact that, most of the time, the tremendous amount of computing and communication resources available in vehicles is underutilized. Putting these resources to work in a meaningful way should have a significant societal impact.

Social implications

The main goal of this paper is to introduce and promote the concept of VCs, a non‐trivial extension, along several dimensions, of the by‐now “classic” cloud computing. The paper shows that the concept of VCs is feasible as of today – at least in some of its manifestations – and that it can have a significant societal impact in the years to come.

Originality/value

The idea of a VC is novel and so are the potential applications that are discussed in the paper.

Details

International Journal of Pervasive Computing and Communications, vol. 7 no. 1
Type: Research Article
ISSN: 1742-7371

Keywords

21 – 30 of over 35000