To read this content please select one of the options below:

Exploring compression and parallelization techniques for distribution of deep neural networks over Edge–Fog continuum – a review

Azra Nazir (Department of Computer Science and Engineering, National Institute of Technology Srinagar, Srinagar, India)
Roohie Naaz Mir (Department of Computer Science and Engineering, National Institute of Technology Srinagar, Srinagar, India)
Shaima Qureshi (Department of Computer Science and Engineering, National Institute of Technology Srinagar, Srinagar, India)

International Journal of Intelligent Computing and Cybernetics

ISSN: 1756-378X

Article publication date: 3 July 2020

Issue publication date: 21 August 2020

274

Abstract

Purpose

The trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud as their resource giant. But this picture leads to underutilization of ever-increasing device pool of IoT that has already passed 15 billion mark in 2015. Thus, it is high time to explore a different approach to tackle this issue, keeping in view the characteristics and needs of the two fields. Processing at the Edge can boost applications with real-time deadlines while complementing security.

Design/methodology/approach

This review paper contributes towards three cardinal directions of research in the field of DL for IoT. The first section covers the categories of IoT devices and how Fog can aid in overcoming the underutilization of millions of devices, forming the realm of the things for IoT. The second direction handles the issue of immense computational requirements of DL models by uncovering specific compression techniques. An appropriate combination of these techniques, including regularization, quantization, and pruning, can aid in building an effective compression pipeline for establishing DL models for IoT use-cases. The third direction incorporates both these views and introduces a novel approach of parallelization for setting up a distributed systems view of DL for IoT.

Findings

DL models are growing deeper with every passing year. Well-coordinated distributed execution of such models using Fog displays a promising future for the IoT application realm. It is realized that a vertically partitioned compressed deep model can handle the trade-off between size, accuracy, communication overhead, bandwidth utilization, and latency but at the expense of an additionally considerable memory footprint. To reduce the memory budget, we propose to exploit Hashed Nets as potentially favorable candidates for distributed frameworks. However, the critical point between accuracy and size for such models needs further investigation.

Originality/value

To the best of our knowledge, no study has explored the inherent parallelism in deep neural network architectures for their efficient distribution over the Edge-Fog continuum. Besides covering techniques and frameworks that have tried to bring inference to the Edge, the review uncovers significant issues and possible future directions for endorsing deep models as processing engines for real-time IoT. The study is directed to both researchers and industrialists to take on various applications to the Edge for better user experience.

Keywords

Acknowledgements

This work is supported by Technical Education Quality Improvement Programme (TEQIP -III). The project is implemented by NPIU which is a unit of MHRD, Govt of India for implementation of World Bank Assisted Projects in Technical Education.

Citation

Nazir, A., Mir, R.N. and Qureshi, S. (2020), "Exploring compression and parallelization techniques for distribution of deep neural networks over Edge–Fog continuum – a review", International Journal of Intelligent Computing and Cybernetics, Vol. 13 No. 3, pp. 331-364. https://doi.org/10.1108/IJICC-04-2020-0038

Publisher

:

Emerald Publishing Limited

Copyright © 2020, Emerald Publishing Limited

Related articles