Search results
1 – 5 of 5The purpose of this paper is to investigate the existence and global exponential stability of periodic solution of memristor-based recurrent neural networks with time-varying…
Abstract
Purpose
The purpose of this paper is to investigate the existence and global exponential stability of periodic solution of memristor-based recurrent neural networks with time-varying delays and leakage delays.
Design/methodology/approach
The differential inequality theory and some novel mathematical analysis techniques are applied.
Findings
A set of sufficient conditions which guarantee the existence and global exponential stability of periodic solution of involved model is derived.
Practical implications
It plays an important role in designing the neural networks.
Originality/value
The obtained results of this paper are new and complement some previous studies. The innovation of this paper concludes two aspects: the analysis on the existence and global exponential stability of periodic solution of memristor-based recurrent neural networks with time-varying delays and leakage delays is first proposed; and it is first time to establish the sufficient criterion which ensures the existence and global exponential stability of periodic solution of memristor-based recurrent neural networks with time-varying delays and leakage delays.
Details
Keywords
Cheng-De Zheng, Ye Liu and Yan Xiao
The purpose of this paper is to develop a method for the existence, uniqueness and globally robust stability of the equilibrium point for Cohen–Grossberg neural networks with…
Abstract
Purpose
The purpose of this paper is to develop a method for the existence, uniqueness and globally robust stability of the equilibrium point for Cohen–Grossberg neural networks with time-varying delays, continuous distributed delays and a kind of discontinuous activation functions.
Design/methodology/approach
Based on the Leray–Schauder alternative theorem and chain rule, by using a novel integral inequality dealing with monotone non-decreasing function, the authors obtain a delay-dependent sufficient condition with less conservativeness for robust stability of considered neural networks.
Findings
It turns out that the authors’ delay-dependent sufficient condition can be formed in terms of linear matrix inequalities conditions. Two examples show the effectiveness of the obtained results.
Originality/value
The novelty of the proposed approach lies in dealing with a new kind of discontinuous activation functions by using the Leray–Schauder alternative theorem, chain rule and a novel integral inequality on monotone non-decreasing function.
Details
Keywords
Changjin Xu, Maoxin Liao and Peiluan Li
The purpose of this paper is to investigate the weighted pseudo-almost periodic solutions of shunting inhibitory cellular neural networks (SICNNs) with time-varying delays and…
Abstract
Purpose
The purpose of this paper is to investigate the weighted pseudo-almost periodic solutions of shunting inhibitory cellular neural networks (SICNNs) with time-varying delays and distributed delays.
Design/methodology/approach
The principle of weighted pseudo-almost periodic functions and some new mathematical analysis skills are applied.
Findings
A set of sufficient criteria which guarantee the existence and exponential stability of the weighted pseudo-almost periodic solutions of the considered SICNNs are established.
Originality/value
The derived results of this paper are new and complement some earlier works. The innovation of this paper concludes two points: a new sufficient criteria guaranteeing the existence and exponential stability of the weighted pseudo-almost periodic solutions of SICNNs are established; and the ideas of this paper can be applied to investigate some other similar neural networks.
Details
Keywords
Vaclav Snasel, Tran Khanh Dang, Josef Kueng and Lingping Kong
This paper aims to review in-memory computing (IMC) for machine learning (ML) applications from history, architectures and options aspects. In this review, the authors investigate…
Abstract
Purpose
This paper aims to review in-memory computing (IMC) for machine learning (ML) applications from history, architectures and options aspects. In this review, the authors investigate different architectural aspects and collect and provide our comparative evaluations.
Design/methodology/approach
Collecting over 40 IMC papers related to hardware design and optimization techniques of recent years, then classify them into three optimization option categories: optimization through graphic processing unit (GPU), optimization through reduced precision and optimization through hardware accelerator. Then, the authors brief those techniques in aspects such as what kind of data set it applied, how it is designed and what is the contribution of this design.
Findings
ML algorithms are potent tools accommodated on IMC architecture. Although general-purpose hardware (central processing units and GPUs) can supply explicit solutions, their energy efficiencies have limitations because of their excessive flexibility support. On the other hand, hardware accelerators (field programmable gate arrays and application-specific integrated circuits) win on the energy efficiency aspect, but individual accelerator often adapts exclusively to ax single ML approach (family). From a long hardware evolution perspective, hardware/software collaboration heterogeneity design from hybrid platforms is an option for the researcher.
Originality/value
IMC’s optimization enables high-speed processing, increases performance and analyzes massive volumes of data in real-time. This work reviews IMC and its evolution. Then, the authors categorize three optimization paths for the IMC architecture to improve performance metrics.
Details
Keywords
Lei Zhang, Huanbin Xue, Zeying Li and Yong Wei
The purpose of this paper is to study the dynamic behavior of complex-valued switched grey neural network models (SGNMs) with distributed delays when the system parameters and…
Abstract
Purpose
The purpose of this paper is to study the dynamic behavior of complex-valued switched grey neural network models (SGNMs) with distributed delays when the system parameters and external input are grey numbers.
Design/methodology/approach
Firstly, by using the properties of grey matrix, M-matrix theory and Homeomorphic mapping, the existence and uniqueness of equilibrium point of the SGNMs were discussed. Secondly, by constructing a proper Lyapunov functional and using the average dwell time approach and inequality technique, the robust exponential stability of the SGNMs under restricted switching was studied. Finally, a numerical example is given to verify the effectiveness of the proposed results.
Findings
Sufficient conditions for the existence and uniqueness of equilibrium point of the SGNMs have been established; sufficient conditions for guaranteeing the robust stability of the SGNMs under restricted switching have been obtained.
Originality/value
(1) Different from asymptotic stability, the exponential stability of SGNMs which include grey parameters and distributed time delays will be investigated in this paper, and the exponential convergence rate of the SGNMs can also be obtained; (2) the activation functions, self-feedback coefficients and interconnected matrices are with different forms in different subnetworks; and (3) the results obtained by LMIs approach are complicated, while the proposed sufficient conditions are straightforward, which are conducive to practical applications.
Details