location: Current position: Home >> Scientific Research >> Paper Publications

A remark on the error-backpropagation learning algorithm for spiking neural networks

Hits:

Indexed by:期刊论文

Date of Publication:2012-08-01

Journal:APPLIED MATHEMATICS LETTERS

Included Journals:SCIE、EI、Scopus

Volume:25

Issue:8

Page Number:1118-1120

ISSN No.:0893-9659

Key Words:Spiking neuron; Error-backpropagation; Differentiation of the firing time with respect to the state

Abstract:In the error-backpropagation learning algorithm for spiking neural networks, one has to differentiate the firing time t(alpha) as a functional of the state function x(t). But this differentiation is impossible to perform directly since t(alpha) cannot be formulated in a standard form as a functional of x(t). To overcome this difficulty, Bohte et al. (2002) (1] assume that there is a linear relationship between the firing time t(alpha) and the state x(t) around t = t(alpha). In terms of this assumption, the Frechet derivative of the functional is equal to the derivative of an ordinary function that can be computed directly and easily. Our contribution in this short note is to prove that this equality of differentiations is in fact mathematically correct, without the help of the linearity assumption. (C) 2012 Elsevier Ltd. All rights reserved.

Pre One:Negative effects of sufficiently small initialweights on back-propagation neural networks

Next One:A modified one-layer spiking neural network involves derivative of the state function at firing time