Hits:
Indexed by:期刊论文
Date of Publication:2018-01-01
Journal:Neural networks : the official journal of the International Neural Network Society
Included Journals:SCIE
Volume:103
Page Number:19-28
ISSN No.:1879-2782
Key Words:Spiking neural networks; SpikeProp; Smoothing L-1/2 regularization; Convergence; Sparsity
Abstract:Unlike the first and the second generation artificial neural networks, spiking neural networks (SNNs) model the human brain by incorporating not only synaptic state but also a temporal component into their operating model. However, their intrinsic properties require expensive computation during training. This paper presents a novel algorithm to SpikeProp for SNN by introducing smoothing L1∕2 regularization term into the error function. This algorithm makes the network structure sparse, with some smaller weights that can be eventually removed. Meanwhile, the convergence of this algorithm is proved under some reasonable conditions. The proposed algorithms have been tested for the convergence speed, the convergence rate and the generalization on the classical XOR-problem, Iris problem and Wisconsin Breast Cancer classification. Copyright © 2018 Elsevier Ltd. All rights reserved.