1,890
7
Essay, 36 pages (9000 words)

Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks

1. Introduction

Artificial neural networks (ANNs) were proposed as a first step toward bio-inspired computation by emulating the way the brain processes information with densely-interconnected neurons and synapses as computational and memory elements, respectively ( Rosenblatt, 1962 ; Bassett and Bullmore, 2006 ). In order to train ANNs, it is necessary to identify how much each neuron contributed to the output error, a problem referred to as the credit assignment ( Minsky, 1961 ). The backpropagation of error (BP) algorithm ( Rumelhart et al., 1986 ) allowed solving the credit assignment problem for multi-layer ANNs, thus enabling the development of deep networks for applications ranging from computer vision ( Krizhevsky et al., 2012 ; LeCun et al., 2015 ; He et al., 2016 ) to natural language processing ( Hinton et al., 2012 ; Amodei et al., 2016 ). However, two critical issues preclude BP from being biologically plausible.

First, BP requires symmetry between the forward and backward weights, which is known as the weight transport problem ( Grossberg, 1987 ). Beyond implying a perfect and instantaneous communication of parameters between the feedforward and feedback pathways, error backpropagation requires each layer to have full knowledge of all the weights in the downstream layers, making BP a non-local algorithm for both weight and error information. From a hardware efficiency point of view, the weight symmetry requirement also severely constrains memory access patterns ( Crafton et al., 2019 ). Therefore, there is an increasing interest in developing training algorithms that release this constraint, as it has been shown that weight symmetry is not mandatory to reach near-BP performance ( Liao et al., 2016 ). The feedback alignment (FA) algorithm ( Lillicrap et al., 2016 ), also called random backpropagation ( Baldi et al., 2018 ), demonstrates that using fixed random weights in the feedback pathway allows conveying useful error gradient information: the network learns to align the forward weights with the backward ones. Direct feedback alignment (DFA) ( Nøkland, 2016 ) builds on these results and directly propagates the error between the network predictions and the targets (i. e. one-hot-encoded labels) to each hidden layer through fixed random connectivity matrices. DFA demonstrates a limited accuracy penalty compared to BP on the MNIST ( LeCun et al., 1998 ) and CIFAR-10 ( Krizhevsky et al., 2009 ) datasets, while using the output error as a global modulator and keeping weight information local. Therefore, DFA bears important structural similarity with learning rules that are believed to take place in the brain ( Guerguiev et al., 2017 ; Neftci et al., 2017 ), known as three-factor synaptic plasticity rules, which rely on local pre- and post-synaptic spike-based activity together with a global modulation ( Urbanczik and Senn, 2014 ). Finally, another approach for solving the weight transport problem consists in computing targets for each layer instead of gradients. The target values can either be computed based on auto-encoders at each layer ( Lee et al., 2015 ) or generated by making use of the pre-activation of the current layer and the error of the next layer, propagated through a dedicated trainable feedback pathway ( Ororbia and Mali, 2019 ). The BP, FA and DFA algorithms are summarized in Figures 1A–C , respectively.

FIGURE 1 Learning Without Feedback: Fixed Random Learning Signals Allow for Feedforward Training of Deep Neural Networks Picture 1

The proposed direct random target projection algorithm builds on feedback-alignment-based algorithms to tackle the weight transport problem while further releasing update locking. Black arrows indicate the feedforward pathways and orange arrows the feedback pathways. In the k -th layer, the weighted sum of inputs y k −1 is denoted as z k , the bias as b k , the activation function as f k (·) and its derivative as f k (·), with k ∈ [1, K ], k ∈ ℕ, and K the number of layers. Trainable forward weight matrices are denoted as W k and fixed random connectivity matrices as B k . The input vector is denoted as x , the target vector as y * and the loss function as J (·). The estimated loss gradients for the outputs of the k -th hidden layer, denoted as δ y k , are provided for each training algorithm. The layer equations for z k , y k and δ z k , defined as the modulatory signals, are provided in the upper left corner, with ⊙ denoting the elementwise multiplication operator.(A)Backpropagation of error (BP) algorithm ( Rumelhart et al., 1986 ).(B)Feedback alignment (FA) algorithm ( Lillicrap et al., 2016 ).(C)Direct feedback alignment (DFA) algorithm ( Nøkland, 2016 ).(D)Proposed direct random target projection (DRTP) algorithm. Adapted from Nøkland (2016 ) and Czarnecki et al. (2017).

The second issue of BP is its requirement for a full forward pass before parameters can be updated during the backward pass, a phenomenon referred to as update locking ( Czarnecki et al., 2017 ; Jaderberg et al., 2017 ). Beyond making BP biologically implausible, update locking has critical implications for BP implementation as it requires buffering all the layer inputs and activations during the forward and backward passes in order to compute the weight updates, leading to a high memory overhead. As the previously-described FA and DFA solutions to the weight transport problem only tackle the weight locality aspect, specific techniques enabling local error handling or gradient approximation are required to tackle update locking. On the one hand, the error locality approach relies on layerwise loss functions ( Mostafa et al., 2018 ; Belilovsky et al., 2019 ; Nøkland and Eidnes, 2019 ; Kaiser et al., 2020 ), it enables training layers independently and without requiring a forward pass in the entire network. The generation of local errors can be achieved with auxiliary fixed random classifiers, allowing for near-BP performance on the MNIST and CIFAR-10 datasets ( Mostafa et al., 2018 ). This strategy has also been ported to a biologically-plausible spike-based three-factor synaptic plasticity rule ( Kaiser et al., 2020 ). Scaling to ImageNet ( Deng et al., 2009 ) requires either the use of two combined layerwise loss functions ( Nøkland and Eidnes, 2019 ) or a parallel optimization of a greedy objective using deeper auxiliary classifiers ( Belilovsky et al., 2019 ). However, the error locality approach still suffers from update locking at the layer scale as layerwise forward and backward passes are required. Beyond implying a computational overhead, the auxiliary classifiers also suffer from the weight transport problem, a requirement that can only be partially relaxed: in order to maintain performance, it is necessary to keep at least the weight sign information during the layerwise backward passes ( Mostafa et al., 2018 ). On the other hand, the synthetic gradients approach ( Czarnecki et al., 2017 ; Jaderberg et al., 2017 ) relies on layerwise predictors of subsequent network computation. However, training local gradient predictors still requires backpropagating gradient information from deeper layers.

In order to fully solve both the weight transport and the update locking problems, we propose the direct random target projection (DRTP) algorithm ( Figure 1D ). Compared to DFA, the targets are used in place of the output error and projected onto the hidden layers. We demonstrate both theoretically and experimentally that, in the framework of classification problems, the error sign information contained in the targets is sufficient to maintain feedback alignment with the loss gradients δ z k for the weighted sum of inputs in layer k , denoted as the modulatory signals in the subsequent text, and allows training multi-layer networks, leading to three key advantages. First, DRTP solves the weight transport problem by entirely removing the need for dedicated feedback pathways. Second, layers can be updated independently and without update locking as a full forward pass is not required, thus reducing memory requirements by releasing the need to buffer inputs and activations of each layer. Third, DRTP is a purely feedforward and low-cost algorithm whose updates rely on layerwise information that is immediately available upon computation of the layer outputs. Estimating the layerwise loss gradients δ y k only requires a label-dependent random vector selection, contrasting with the error locality and synthetic gradients approaches that require the addition of side networks for error or gradient prediction. DRTP even compares favorably to DFA, as the latter still requires a multiplication between the output error and a fixed random matrix.

Therefore, DRTP allows relaxing structural, memory and computational requirements, yet we demonstrate that, compared to BP, FA and DFA, DRTP is ideal for implementation in edge-computing devices, thus enabling adaptation to uncontrolled environments while meeting stringent power and resource constraints. Suitable applications for DRTP range from distributed smart sensor networks for the Internet-of-Things (IoT) ( Bol et al., 2015 ) to embedded systems and cognitive robotic agents ( Milde et al., 2017 ). The MNIST and CIFAR-10 datasets have thus been selected for benchmarking as they are representative of the complexity level required in autonomous always-on adaptive edge computing, which is not the case of larger and more challenging datasets such as ImageNet. This furthermore highlights that edge computing is an ideal use case for biologically-motivated algorithms, as an out-of-the-box application of feedback-alignment- and target-propagation-based algorithms currently does not scale to complex datasets (see Bartunov et al., 2018 for a recent review). We demonstrate this claim in Frenkel et al. (2020) with the design of an event-driven convolutional processor that requires only 16. 8–% power and 11. 8–% silicon area overheads for on-chip online learning, a record-low overhead that is specifically enabled by DRTP, thus highlighting its low cost for edge computing devices. Finally, as DRTP can also be formulated as a three-factor learning rule for biologically-plausible learning, it is suitable for embedded neuromorphic computing, in which high-density synaptic plasticity can currently not be achieved without compromising learning performance ( Frenkel et al., 2019b , c ).

2. Results

2. 1. Weight Updates Based Only on the Error Sign Provide Learning to Multi-Layer Networks

We demonstrate with two experiments, respectively on a regression task and a classification problem, that modulatory signals based only on the error sign are within 90° of those prescribed by BP, thus providing learning in multi-layer networks. To do so, we use an error-sign-based version of DFA, subsequently denoted as sDFA, in which the error vector is replaced by the error sign in the global feedback pathway.

2. 1. 1. Regression

This first experiment aims at demonstrating that the error sign provides useful modulatory signals to multi-layer networks by comparing training algorithms on a regression task. The objective is to approximate 10 non-linear functions T j ( x )= cos ( x ¯+ ϕ j ), where ϕ j = −π/2+/9 for j ∈ [0, 9], j ∈ ℕ 0 and x ¯ denotes the mean of x , a 256-dimensional vector whose entries are drawn from a normal distribution with a mean lying in [−π, π] (see section 4). A 256-100-100-10 fully-connected network is trained to approximate T (·) with five training algorithms: shallow learning (i. e. frozen random hidden layers and a trained output layer), BP, FA, DFA and sDFA.

The mean squared error (MSE) loss on the training set is shown in Figure 2A . While shallow learning fails to learn a meaningful approximation of T (·), sDFA and DFA show the fastest initial convergence due to the separate direct feedback pathway precluding gradients from vanishing, which is clearly an issue for BP and FA. Although this would be alleviated by using ReLU-based networks with batch normalization ( Ioffe and Szegedy, 2015 ), it highlights that direct-feedback-alignment-based methods do not need further techniques such as batch normalization to address this issue, ultimately leading to reduced hardware requirements. While DFA demonstrates the highest performance on this task, sDFA comes earlier to stagnation as it does not account for the output error magnitude reduction as training progresses, thus preventing a reduction of the effective learning rate in the hidden layers as the output error decreases. sDFA could therefore benefit from the use of a learning rate scheduler. Similar conclusions hold for the loss on the test set ( Figure 2B ). The angle between the modulatory signals prescribed by BP and by feedback-alignment-based algorithms is shown in Figures 2C, D for the first and second hidden layers, respectively. While all feedback-alignment-based algorithms lie close to each other within 90° of the BP modulatory signals, FA has a clear advantage during the first 100 epochs on the 5k-example training set. sDFA performs on par with DFA in the first hidden layer, while it surprisingly provides a better alignment than DFA in the second hidden layer, though not fully leveraged due to the absence of modulation in the magnitude of the updates from the output error.

FIGURE 2 Learning Without Feedback: Fixed Random Learning Signals Allow for Feedforward Training of Deep Neural Networks Picture 2

Error-sign-based direct feedback alignment (sDFA) provides useful modulatory signals in regression tasks. A 256-100-100-10 network with tanh hidden and output units is trained to learn cosine functions with five training algorithms: shallow learning, BP, FA, DFA and sDFA. With this simple setup, BP and FA suffer from the vanishing gradients problem, which would be alleviated by using ReLU-based networks with batch normalization. The scope of the figure is to highlight that sDFA provides useful modulatory signals for regression tasks, without any additional technique. As for other feedback-alignment-based algorithms, sDFA updates are within 90° of the backpropagation updates. The train and test losses and the alignment angles are monitored every 1k samples, error bars are one standard deviation over 10 runs. Angles have been smoothed by an exponentially-weighted moving average filter with a momentum coefficient of 0. 95.(A)Mean squared error loss on the 5k-example training set.(B)Mean squared error loss on the 1k-example test set.(C)Angle between the modulatory signals δ z k prescribed by BP and by feedback-alignment-based algorithms in the first hidden layer.(D)Angle between the modulatory signals δ z k prescribed by BP and by feedback-alignment-based algorithms in the second hidden layer.

2. 1. 2. Classification

With this second experiment, we demonstrate that, in addition to providing useful modulatory signals for regression problems, the error sign information allows training multi-layer networks to solve classification problems. The task consists in training a 256-500-500-10 network to solve a synthetic classification problem with 16×16-pixel images and 10 classes; the data to classify is generated automatically with the Python sklearn library ( Pedregosa et al., 2011 ) (see section 4). As for regression, the network is trained with shallow learning, BP, FA, DFA and sDFA.

Figure 3A shows that, after 500 epochs with a 25k-example training set, DFA provides the fastest and most accurate training with a classification error of 0. 05%, followed by BP, FA and sDFA with 0. 19, 0. 64, and 1. 54%, respectively. Shallow learning lags almost an order of magnitude behind with 8. 95%. However, Figure 3B shows that DFA also has a higher overfitting and lies close to sDFA on the test set, with 3. 48 and 4. 07%, respectively. The lowest classification errors are of 1. 85% for BP and 1. 81% for FA, while shallow learning lags behind at 9. 57%. The angle between the modulatory signals prescribed by BP and by feedback-alignment-based algorithms is shown in Figures 3C, D , for the first and second hidden layers, respectively. As for the regression task, all feedback-alignment-based algorithms exhibit alignments close to each other, while the convergence of BP and FA is slowed down by the vanishing gradients problem. Here, alignments tend to level off after 50 epochs, with the lowest angle provided by FA, followed by DFA and sDFA. As sDFA is always within 90° of the BP modulatory signals, it is able to train multi-layer networks.

FIGURE 3 Learning Without Feedback: Fixed Random Learning Signals Allow for Feedforward Training of Deep Neural Networks Picture 3

Error-sign-based direct feedback alignment (sDFA) provides useful modulatory signals in classification tasks. A 256-500-500-10 network with tanh hidden units and sigmoid output units is trained to classify a synthetic dataset of 16×16-pixel images into 10 classes with five training algorithms: shallow learning, BP, FA, DFA and sDFA. With this simple setup, BP and FA suffer from the vanishing gradients problem, which would be alleviated by using ReLU-based networks with batch normalization. The scope of the figure is to highlight that sDFA provides useful modulatory signals for classification tasks, without any additional technique. The update directions of the sDFA algorithm are within 90° of the backpropagation updates and are comparable to other feedback-alignment-based algorithms. The train and test losses and the alignment angles are monitored every 2. 5k samples, error bars are one standard deviation over 10 runs. Angles have been smoothed by an exponentially-weighted moving average filter with a momentum coefficient of 0. 95.(A)Error on the 25k-example training set, reaching on average 0. 19% for BP, 0. 64% for FA, 0. 05% for DFA, 1. 54% for sDFA and 8. 95% for shallow learning after 500 epochs.(B)Error on the test set, reaching on average 1. 85% for BP, 1. 81% for FA, 3. 48% for DFA, 4. 07% for sDFA and 9. 57% for shallow learning after 500 epochs.(C)Angle between the modulatory signals δ z k prescribed by BP and by feedback-alignment-based algorithms in the first hidden layer.(D)Angle between the modulatory signals δ z k prescribed by BP and by feedback-alignment-based algorithms in the second hidden layer.

2. 2. For Classification Problems, a Feedback Pathway Is No Longer Required as the Error Sign Is Known in Advance

In the framework of classification problems, training examples ( x , c *) consist of an input data sample to classify, denoted as x , and a label c * denoting the class x belongs to, among C possible classes. The target vector, denoted as y *, corresponds to the one-hot-encoded class label c *. The output layer non-linearity must be chosen as a sigmoid or a softmax function, yielding output values that are strictly bounded between 0 and 1. Denoting the output vector of a K -layer network as y K , the error vector is defined as e = y * y K . Under the aforementioned conditions, it results that the c -th entry of the C -dimensional error vector e , denoted e c , is defined as

e c ={ 1 y K c if c = c *, y K c otherwise .

As the entries of y K are strictly bounded between 0 and 1, the error sign is given by

e c ={ 1 if c = c *, 1 otherwise .

Due to the non-linearity in the output layer forcing the output values to remain strictly bounded between 0 and 1, the error sign is class-dependent and known in advance as training examples ( x , c *) already provide the error sign information with the label c *. A feedback pathway is thus no longer required as we have shown that the error sign allows providing useful modulatory signals to train multi-layer networks. Therefore, beyond being free from the weight transport problem as DFA, sDFA also allows releasing update locking and the associated memory overhead in classification problems.

2. 3. Direct Random Target Projection Delivers Useful Modulatory Signals for Classification

This section provides the grounds to show why the proposed direct random target projection (DRTP) algorithm delivers useful modulatory signals to multi-layer networks in the framework of classification problems. First, we show how DRTP can be viewed as a simplified version of sDFA in which the target vector y * is used as a surrogate for the error sign. Next, we demonstrate mathematically that, in a multi-layer network composed of linear hidden layers and a non-linear output layer, the modulatory signals prescribed by DRTP and BP are always within 90° of each other, thus providing learning in multi-layer networks.

2. 3. 1. DRTP Is a Simplified Version of Error-Sign-Based DFA

As we have shown that sDFA solves both the weight transport and the update locking problems in classification tasks, we propose the direct random target projection (DRTP) algorithm, illustrated in Figure 1D and written in pseudocode in Algorithm 1 , as a simplified version of sDFA that enhances both performance and computational efficiency. In sDFA, the feedback signal randomly projected to the hidden layers is the sign of the error vector e = y * y K , while in DRTP, this feedback signal is replaced by the target vector y *. Being a one-hot encoding of c *, y * has a single positive entry corresponding to the correct class and zero entries elsewhere:

y c *= 1 + sign ( e c ) 2 ={ 1 if c = c *, 0 otherwise . ALGORITHM 1 Learning Without Feedback: Fixed Random Learning Signals Allow for Feedforward Training of Deep Neural Networks Picture 4

Algorithm 1: Pseudocode for the direct random target projection (DRTP) algorithm. k ∈ [1, K ], k ∈ ℕ, denotes the layer index and W k , b k , B k and f k (·) denote the trainable forward weights and biases, the fixed random connectivity matrices and the activation function of the k -th hidden layer, respectively. The weighted sum of inputs or pre-activation is denoted as z k and the layer output or post-activation is denoted as y k , with y 0 corresponding to the input x . The one-hot-encoding of labels among C output classes is denoted as y * and the learning rate as η. The update for the weights and biases in the output layer are computed for sigmoid/softmax output units with a binary/categorical cross-entropy loss.

Thus, y * corresponds to a surrogate for the error sign vector used in sDFA, where shift and rescaling operations have been applied to sign( e ). As the connectivity matrices B k in the DRTP gradients δ y k = B k T y * are fixed and random ( Figure 1D ), they can be viewed as comprising the rescaling operation. Only the shift operation applied to sign( e ) makes a critical difference between DRTP and sDFA, which is favorable to DRTP for two reasons. First, DRTP is computationally cheaper than sDFA. Indeed, projecting the target vector y * to the hidden layers through fixed random connectivity matrices is equivalent to a label-dependent selection of a layerwise random vector. On the contrary, sDFA requires multiplying the error sign vector with the fixed random connectivity matrices for each training example, as all entries of the error sign vector are non-zero. Second, experiments on the MNIST and CIFAR-10 datasets show that DRTP systematically outperforms sDFA ( Supplementary Figures 1A , 2A , Supplementary Tables 1 , 2 ). Indeed, when the feedback information only relies on the error sign and no longer on its magnitude, the weight updates become less selective to the useful information: as all entries of the error sign vector have unit norm, the C −1 entries corresponding to incorrect classes outweigh the single entry associated to the correct class and degrade the alignment ( Supplementary Figures 1B , 2B ).

2. 3. 2. The Directions of the DRTP and BP Modulatory Signals Are Within 90° of Each Other

We provide a mathematical proof of alignment between the DRTP and BP modulatory signals. The structure of our proof is inspired from the FA proof of alignment in Lillicrap et al. (2016), which we expand in two ways. First, we extend this proof for the case of DRTP. Second, while Lillicrap et al. (2016) demonstrate the alignment with the BP modulatory signals for a network consisting of a single linear hidden layer, a linear output layer and a mean squared error loss, we demonstrate that alignment can be achieved for an arbitrary number of linear hidden layers, a non-linear output layer with sigmoid/softmax activation and a binary/categorical cross-entropy loss for classification problems. Both proofs are restricted to the case of a single training example. Under these conditions, it is possible to guarantee that the DRTP modulatory signals are aligned with those of BP. This comes from the fact that the prescribed weight updates lead to a soft alignment between the product of forward weight matrices and the fixed random connectivity matrices. The mathematical details, including the lemma and theorem proofs, have been abstracted out to the Supplementary Note 1 .

In the case of the multi-layer neural network composed of linear hidden layers shown in Figure 4 , the output of the k -th hidden layer is given by

y k = z k = W k y k 1 for k [ 1 , K 1 ],

where K is the number of layers, y 0 = x is the input vector, and the bias vector b k is omitted without loss of generality. The output layer is described by

z K = W K y K 1 , y K = σ ( z K ),

where σ(·) is either the sigmoid or the softmax activation function. The loss function J (·) is either the binary cross-entropy (BCE) loss for sigmoid output units or the categorical cross-entropy (CCE) loss for softmax output units, computed over the C output classes:

J BCE ( y K , y *)= 1 C c = 1 C ( y c * log ( y K c )+( 1 y c *) log ( 1 y K c )), J CCE ( y K , y *)= 1 C c = 1 C ( y c * log ( y K c )).

Lemma. In the case of zero-initialized weights, i. e. W k 0 = 0 for k ∈ [1, K ], k ∈ ℕ, and hence of zero-initialized hidden layer outputs, i. e. y k 0 = 0 for k ∈ [1, K −1] and z K 0 = 0 , considering a DRTP-based training performed recursively with a single element of the training set ( x, c *) and y * denoting the one-hot encoding of c *, at every discrete update step t , there are non-negative scalars s y k t and s W k t for k ∈ [1, K −1] and a C -dimensional vector s W K t such that

y k t = s y k t ( B k T y *) for k [ 1 , K 1 ] W 1 t = s W 1 t ( B 1 T y *) x T W k t = s W k t ( B k T y *)( B k 1 T y *) T for k [ 2 , K 1 ] W K t = s W K t ( B K 1 T y *) T .

Theorem. Under the same conditions as in the lemma and for the linear-hidden-layer network dynamics described above, the k -th layer modulatory signals prescribed by DRTP are always a negative scalar multiple of the Moore-Penrose pseudo-inverse of the product of forward matrices of layers k +1 to K , located in the feedback pathway between the output layer and the k -th hidden layer, multiplied by the error. That is, for k ∈ [1, K −1] and t > 0,

1 s k t ( i = K k + 1 W i t )+ e = B k T y * with s k t > 0 .

Alignment. In the framework of classification problems, as the coefficients s k t are strictly positive scalars for t > 0, it results from the theorem that the dot product between the BP and DRTP modulatory signals is strictly positive, i. e.

e T ( i = k + 1 K W i T ) T ( B k T y *)> 0 e T ( i = k + 1 K W i T ) T ( i = K k + 1 W i )+ I e s k t > 0 e T e s k t > 0 .

The BP and DRTP modulatory signals are thus within 90° of each other. □

FIGURE 4 Learning Without Feedback: Fixed Random Learning Signals Allow for Feedforward Training of Deep Neural Networks Picture 5

Network of DRTP-updated linear hidden layers considered in the context of the mathematical proof of alignment between the DRTP and BP modulatory signals. The same conventions as in Figure 1 are used.

2. 4. DRTP Learns to Classify MNIST and CIFAR-10 Images Without Feedback

In this section, we compare DRTP with BP and other feedback-alignment-based algorithms, namely FA and DFA, on the MNIST and CIFAR-10 datasets. Both datasets have 10 output classes, they respectively consist in classifying 28×28 grayscale images of handwritten digits for MNIST and 32×32 RGB images of vehicles and animals for CIFAR-10. The network topologies considered in our experiments are, on the one hand, fully-connected (FC) networks with one or two hidden layers, respectively denoted as FC1 and FC2, each hidden layer being constituted of either 500 or 1, 000 tanh units. On the other hand, convolutional (CONV) networks are used with either fixed random or trainable kernels. The CONV network for MNIST consists of one convolutional layer followed by a max-pooling layer and one fully-connected hidden layer, while for CIFAR-10 it consists of two convolutional layers, each followed by a max-pooling layer, and two fully-connected hidden layers (see section 4).

2. 4. 1. MNIST

The results on the MNIST dataset are summarized in Table 1 . In FC networks, BP, FA and DFA perform similarly, the accuracy degradation of FA and DFA is marginal. While there is a higher accuracy degradation for DRTP, it compares favorably to shallow learning, which suffers from a high accuracy penalty. It shows that DRTP allows training hidden layers to learn MNIST digit classification without feedback. The CONV network topology leads to the lowest error, highlighting that extracting spatial information, even with random kernels, is sufficient to solve the MNIST task. The accuracy slightly degrades along the FA, DFA and DRTP algorithms, with a higher gap for shallow learning. When kernels are trained, BP provides the highest improvement compared to the error obtained with random kernels, followed by DRTP, while no significant change can be observed for FA and DFA. This is likely due to the fact that there is not enough parameter redundancy in convolutional layers to allow for an efficient training with feedback-alignment-based algorithms, which is commonly referred to as a bottleneck effect (see section 3). Indeed, the angle between the BP loss gradients and the feedback-alignment-based ones is roughly 90°, leading to random updates ( Supplementary Figure 3 ). This improved performance of DRTP with trained kernels is thus unexpected. Regarding dropout, a positive impact is shown on BP, FA and DFA: a moderate dropout probability is beneficial for FC1 networks, while increasing it to 0. 25 can be used for FC2 networks. Dropout has no positive impact for CONV networks, while it degrades the accuracy obtained with DRTP and shallow learning in all cases.

TABLE 1 Learning Without Feedback: Fixed Random Learning Signals Allow for Feedforward Training of Deep Neural Networks Picture 6

Mean and standard deviation of the test error on the MNIST dataset over 10 trials.

2. 4. 2. CIFAR-10

The results on the CIFAR-10 dataset are summarized in Table 2 , highlighting conclusions similar to those already drawn for the MNIST dataset. Compared to BP, accuracy degrades along the FA, DFA and DRTP algorithms. The gap is higher for DRTP, yet it again compares favorably to shallow learning, demonstrating that DRTP also allows training hidden layers to learn CIFAR-10 image classification without feedback. For CONV networks, if kernels are trained, only BP is able to provide a significant advantage. Due to the bottleneck effect, FA only provides a slight improvement, while DFA and DRTP are negatively impacted. Regarding dropout, a moderate probability of 0. 1 works fairly well for BP, FA, DFA and DRTP, while a higher probability of 0. 25 rarely provides any advantage. Dropout always leads to an accuracy reduction for shallow learning. Finally, data augmentation (DA) improves the accuracy of all algorithms and is more effective than dropout.

TABLE 2 Learning Without Feedback: Fixed Random Learning Signals Allow for Feedforward Training of Deep Neural Networks Picture 7

Mean and standard deviation of the test error on the CIFAR-10 dataset over 10 trials.

3. Discussion

While the backpropagation of error algorithm allowed taking artificial neural networks to outperform humans on complex datasets such as ImageNet ( He et al., 2015 ), the key problems of weight transport and update locking highlight how aiming at breaking accuracy records on standard datasets has diverted attention from hardware efficiency considerations. While accuracy is the key driver for applications that can be backed by significant GPU and CPU resources, the development of decentralized adaptive smart sensors calls for keeping hardware requirements of learning algorithms to a minimum. Moreover, it has been shown that weight transport and update locking are not biologically plausible ( Grossberg, 1987 ; Baldi et al., 2018 ), following from the non-locality in both weight and gradient information. Therefore, there is currently an increasing interest in releasing these constraints in order to achieve higher hardware efficiency and to understand the mechanisms that could underlie biological synaptic plasticity.

The proposed DRTP algorithm successfully addresses both the weight transport and the update locking problems, which has only been partially demonstrated in previously-proposed approaches. Indeed, the FA and DFA algorithms only address the weight transport problem ( Lillicrap et al., 2016 ; Nøkland, 2016 ). The error locality approach still suffers from the weight transport problem in the local classifiers ( Mostafa et al., 2018 ; Nøkland and Eidnes, 2019 ; Kaiser et al., 2020 ), while the synthetic gradients approach requires backpropagating gradient information from deeper layers in order to train the layerwise gradient predictors ( Czarnecki et al., 2017 ; Jaderberg et al., 2017 ). Both the error locality and the synthetic gradients approaches also incur computational overhead by requiring the addition of side local networks for error or gradient prediction. On the contrary, DRTP is a strikingly simple rule that alleviates the two key BP issues by enabling each layer to be updated with local information as the forward evaluation proceeds. In order to estimate the layerwise loss gradients δ y k for each layer, the only operation required by DRTP is a label-dependent random vector selection ( Figure 1D ). Despite the absence of dedicated feedback pathways, we demonstrated on the MNIST and CIFAR-10 datasets that DRTP allows training hidden layers at low computational and memory costs, thus highlighting its suitability for deployment in adaptive smart sensors at the edge and for embedded systems in general. In terms of floating-point operations (FLOPs), the overhead of DRTP weight updates is approximately equal to the cost of the forward pass, assuming that (i) the number of classes of the problem is negligible compared to the number of units in the hidden layers, which is typical of edge computing tasks, and (ii) the learning rate is embedded in the magnitude of the random connectivity matrices B k T . Doubling the computational cost of shallow-learning networks (i. e. doubling the numbers of hidden units or hidden layers) does not allow recovering their performance gap compared to DRTP-updated networks ( Tables 1 , 2 ). Even more importantly when considering dedicated hardware implementations for edge computing, the memory requirements should be minimized so as to fit the whole network topology into on-chip memory resources. Indeed, accesses to off-chip DRAM memory are three orders of magnitude more expensive energy-wise than a 32-bit FLOP ( Horowitz, 2014 ). Therefore, as opposed to increasing the resources of shallow-trained networks, DRTP offers a low-overhead training algorithm operating on small network topologies, ideally suiting edge-computing hardware requirements. These claims are proven in silico in Frenkel et al. (2020), where implementing DRTP in an event-driven convolutional processor requires only 16. 8–% power and 11. 8–% silicon area overheads and allows demonstrating a favorable accuracy-power-area tradeoff compared to both on-chip online- and off-chip offline-trained conventional machine learning accelerators on the MNIST dataset.

By solving the weight transport and update locking problems, DRTP also releases key biological implausibility issues. Neurons in the brain separate forward and backward information in somatic and dendritic compartments, a property that is highlighted in the formulation of three-factor synaptic plasticity rules ( Urbanczik and Senn, 2014 ): pre-synaptic and post-synaptic activities are modulated by a third factor corresponding to a local dendritic voltage. Lillicrap et al. (2016) build on the idea that a separate dendritic compartment integrates higher-order feedback and generates local teaching signals, where the errors could be viewed as a mismatch between expected and actual perceptions or actions. This aspect is further emphasized in the subsequent work of Guerguiev et al. (2017) when framing DFA as a spike-based three-factor learning rule. In the case of DRTP, compared to DFA, the error signal is replaced by the targets, which could correspond to a modulation that bypasses the actual perceptions or realized actions, relying only on predictions or intentions. Furthermore, DRTP could come in line with recent findings in cortical areas that reveal the existence of output-independent target signals in the dendritic instructive pathways of intermediate-layer neurons ( Magee and Grienberger, 2020 ). Understanding the mechanisms of synaptic plasticity is critical in the field of neuromorphic engineering, which aims at porting biological computational principles to hardware toward higher energy efficiency ( Thakur et al., 2018 ; Rajendran et al., 2019 ). However, even simple local bio-inspired learning rules such as spike-timing-dependent plasticity (STDP) ( Bi and Poo, 1998 ) can lead to non-trivial hardware requirements, which currently hinders adaptive neuromorphic systems from reaching high-density large-scale integration ( Frenkel et al., 2019b ). While adaptations of STDP, such as spike-dependent synaptic plasticity (SDSP) ( Brader et al., 2007 ), release most of the STDP hardware constraints, their training performance is currently not sufficient to support deployability of neuromorphic hardware for real-world scenarios ( Frenkel et al., 2019b , c ). A three-factor formulation of DRTP would release the update locking problem in the spike-based three-factor formulations of DFA ( Guerguiev et al., 2017 ; Neftci et al., 2017 ), which currently imply memory and control overhead in their hardware implementations ( Detorakis et al., 2018 ; Park et al., 2019 ). Porting DRTP to neuromorphic hardware is thus a natural next step.

While DRTP relaxes structural, memory and computational requirements toward decentralized hardware deployment, the accuracy degradation over DFA comes from the fact that only the error sign is taken into account, not its class-dependent magnitude. This could be mitigated by keeping track of the error magnitude over the last samples in order to modulate the layerwise learning rates, at the expense of releasing the purely feedforward nature of DRTP. A learning rate scheduler could also be used. The DRTP algorithm was derived specifically for classification problems with sigmoid/softmax output units and a binary/categorical cross-entropy loss, yet hidden layer activations also play a key role in the learning dynamics of DRTP. As the estimated loss gradients δ y k computed from the targets have a constant sign and magnitude, the weights updates only change due to the previous layer outputs and the derivative of the activation function, as training progresses. When using activation functions such as tanh in the hidden layers, the network stops learning thanks to the activation function derivative, whose value vanishes as its input argument moves away from zero. This mechanism specific to DRTP is highlighted in Supplementary Figures 4 6 and could be exploited to generate networks whose activations can be binarized during inference, which we will investigate in future work. In return, only activation functions presenting this saturation property are expected to lead to satisfying performance when used in conjunction with DRTP, which for example excludes ReLU activations.

Finally, as for all other feedback-alignment-based algorithms, DRTP only slightly improves or even degrades the accuracy when applied to convolutional layers. Convolutional layers do not provide the parameter redundancy that can be found in fully-connected layers, a bottleneck effect that was first highlighted for FA ( Lillicrap et al., 2016 ) and has recently been studied for DFA ( Launay et al., 2019 ). Nevertheless, other training algorithms based either on a greedy layerwise learning ( Belilovsky et al., 2019 ) or on the alignment with local targets ( Ororbia and Mali, 2019 ) have proven to be successful in training convolutional layers at the expense of only partially solving the update locking problem. Indeed, the training algorithm proposed in Belilovsky et al. (2019) still suffers from update locking in the layerwise auxiliary networks while the one proposed in Ororbia and Mali (2019) relies on the backpropagation of the output error to compute the layerwise targets. If fixed random convolutional layers do not meet the performance requirements of the target application, a combination of DRTP for fully-connected layers together with error locality or synthetic gradients approaches for convolutional layers can be considered. This granularity in the selection of learning mechanisms, trading off accuracy and hardware efficiency, comes in accordance with the wide spectrum of plasticity mechanisms that are believed to operate in the brain ( Zenke et al., 2015 ).

4. Materials and Methods

The training on both the synthetic regression and classification tasks and the MNIST and CIFAR-10 datasets has been carried out with PyTorch ( Paszke et al., 2017 ), one of the numerous Python frameworks supporting deep learning. In all experiments, the reported update angles between feedback-alignment-based algorithms and BP were generated at each update step, where the BP update values were computed solely to assess the evolution of the alignment angle over the update steps carried out by FA, DFA, sDFA or DRTP.

4. 1. Regression

The examples in the training and test sets are denoted as ( x, y *). The 10-dimensional target vectors y * are generated using y j *= T j ( x )= cos ( x ¯+ ϕ j ), where ϕ j = −π/2+/9 for j ∈ [0, 9] and j ∈ ℕ 0 . x ¯ denotes the mean of x , a 256-dimensional vector whose entries are initialized from a normal distribution with a mean sampled from a uniform distribution between −π and π and with a unit variance. The training and test sets respectively contain 5k and 1k examples. The trained network has a 256-100-100-10 topology with tanh hidden and output units, whose forward weights are drawn from a He uniform distribution ( He et al., 2015 ) and are zero-initialized for feedback-alignment-based algorithms. The random connectivity matrices of feedback-alignment-based algorithms are also drawn from He uniform distributions. The weights are updated after each minibatch of 50 examples, and the network is trained for 500 epochs with a fixed learning rate η = 5 × 10 −4 for all training algorithms. As this is a regression task, the loss function is the mean squared error. The losses on the training and test sets and the alignment angles with BP updates are monitored every 1k samples. The experiment is repeated 10 times for each training algorithm, with different network initializations for each experiment run.

4. 2. Synthetic Data Classification

The examples in the training and test sets are generated using the make_classification function from the Python library sklearn ( Pedregosa et al., 2011 ). The main inputs required by this function are the number of samples to be generated, the number of features n in the input vectors x , the number of informative features n inf among the input vectors, the number of classes, the number of clusters per class and a factor class_sep which conditions the class separation. In this work, we have used n = 256 and n inf = 128, ten classes, five clusters per class and class_sep = 4. 5. Using this set of parameters, the make_classification function then generates examples by creating for each class clusters of points normally distributed about the vertices of an n inf -dimensional hypercube. The remaining features are filled with normally-distributed random noise. The generated examples are then separated into training and test sets of 25k and 5k examples, respectively. The trained network has a 256-500-500-10 topology with tanh hidden units and sigmoid output units. The forward and backward weights initialization, as well as the forward weight updates, are performed as for regression. The loss function is the binary cross-entropy loss. The network is trained for 500 epochs with a fixed learning rate η = 5 × 10 −4 . The losses on the training and test sets and the alignment angles with BP updates are monitored every 2. 5k samples. For each training algorithm, the experiment is repeated 10 times with different network initializations.

4. 3. MNIST and CIFAR-10 Images Classification

A fixed learning rate is selected based on a grid search for each training algorithm, dataset and network type ( Table 3 ). For both the MNIST and CIFAR-10 experiments, the chosen optimizer is Adam with default parameters. A sigmoid output layer and a binary cross-entropy loss are used for all training algorithms. The entries of the forward weight matrices W k are initialized with a He uniform distribution, as well as the entries of the fixed random connectivity matrices B k of feedback-alignment-based algorithms. When used, dropout is applied with the same probability to all fully-connected layers. For MNIST, the networks are trained for 100 epochs with a minibatch size of 60. The CONV network topology consists of a convolutional layer with 32 5×5 kernels, a stride of 1 and a padding of 2, a max-pooling layer with 2×2 kernels and a stride of 2, a fully-connected layer of 1, 000 tanh units and an output fully-connected layer of 10 units. For CIFAR-10, a minibatch size of 100 is used and early stopping is applied, with a maximum of 200 epochs. The CONV network topology consists of two convolutional layers with respectively 64 and 256 3×3 kernels, a stride and a padding of 1, both followed by a max-pooling layer with 2×2 kernels and a stride of 2, then two fully-connected layers of 1, 000 tanh units and an output fully-connected layer of 10 units. For all experiments, the test error is averaged over the last 10 epochs of training. The results reported in Tables 1 , 2 and Supplementary Tables 1 , 2 are the mean and standard deviation over 10 trials.

TABLE 3 Learning Without Feedback: Fixed Random Learning Signals Allow for Feedforward Training of Deep Neural Networks Picture 8

The learning rate values for the MNIST and CIFAR-10 datasets are selected based on a grid search.

Data Availability Statement

Publicly available datasets were analyzed in this study. The data can be found at: http://yann. lecun. com/exdb/mnist/ and https://www. cs. toronto. edu/~kriz/cifar. html .

Code Availability

The PyTorch code allowing to reproduce all results in this study is available open source under the Apache 2. 0 license at https://github. com/ChFrenkel/DirectRandomTargetProjection .

Author Contributions

CF developed the main idea. CF and ML derived the mathematical proofs and worked on the simulation experiments. CF, ML, and DB wrote the paper. All authors contributed to the article and approved the submitted version.

Funding

CF was with Université catholique de Louvain as a Research Fellow from the National Foundation for Scientific Research (FNRS) of Belgium, under grant number 1117116F-1117118F.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

An earlier version of this work appeared as a preprint in Frenkel et al. (2019a). The authors would like to thank Emre Neftci, Giacomo Indiveri, Marian Verhelst, Simon Carbonnelle and Vincent Schellekens for fruitful discussions and Christophe De Vleeschouwer for granting access to a deep learning workstation.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www. frontiersin. org/articles/10. 3389/fnins. 2021. 629892/full#supplementary-material

References

Amodei, D., Ananthanarayanan, S., Anubhai, R., Bai, J., Battenberg, E., Case, C., et al. (2016). “ Deep speech 2: end-to-end speech recognition in English and Mandarin,” in Proceedings of the 33rd International Conference on Machine Learning (New York, NY), 173–182.

Baldi, P., Sadowski, P., and Lu, Z. (2018). Learning in the machine: random backpropagation and the deep learning channel. Artif. Intell . 260, 1–35. doi: 10. 1016/j. artint. 2018. 03. 003

Bartunov, S., Santoro, A., Richards, B., Marris, L., Hinton, G. E., and Lillicrap, T. (2018). “ Assessing the scalability of biologically-motivated deep learning algorithms and architectures,” in Advances in Neural Information Processing Systems , eds S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (Montreal, QC: Curran Associates, Inc.), 9368–9378.

Bassett, D. S., and Bullmore, E. T. (2006). Small-world brain networks. Neuroscientist 12, 512–523. doi: 10. 1177/1073858406293182

Belilovsky, E., Eickenberg, M., and Oyallon, E. (2019). Decoupled greedy learning of CNNs. arXiv preprint arXiv: 1901. 08164 .

Bi, G.-Q., and Poo, M.-M. (1998). Synaptic modifications in cultured hippocampal neurons: dependence on spike timing, synaptic strength, and postsynaptic cell type. J. Neurosci . 18, 10464–10472. doi: 10. 1523/JNEUROSCI. 18-24-10464. 1998

Bol, D., de Streel, G., and Flandre, D. (2015). “ Can we connect trillions of IoT sensors in a sustainable way? A technology/circuit perspective,” in 2015 IEEE SOI-3D-Subthreshold Microelectronics Technology Unified Conference (S3S)(Sonoma Valley, CA: IEEE), 1–3. doi: 10. 1109/S3S. 2015. 7333500

Brader, J. M., Senn, W., and Fusi, S. (2007). Learning real-world stimuli in a neural network with spike-driven synaptic dynamics. Neural Comput . 19, 2881–2912. doi: 10. 1162/neco. 2007. 19. 11. 2881

Crafton, B., West, M., Basnet, P., Vogel, E., and Raychowdhury, A. (2019). “ Local learning in RRAM neural networks with sparse direct feedback alignment,” in 2019 IEEE/ACM International Symposium on Low Power Electronics and Design (ISLPED)(Lausanne: IEEE), 1–6. doi: 10. 1109/ISLPED. 2019. 8824820

Czarnecki, W. M., Świrszcz, G., Jaderberg, M., Osindero, S., Vinyals, O., and Kavukcuoglu, K. (2017). “ Understanding synthetic gradients and decoupled neural interfaces,” in Proceedings of the 34th International Conference on Machine Learning (Sydney, NSW), 904–912.

Deng, J., Dong, W., Socher, R., Li, L. J., Li, K., and Fei-Fei, L. (2009). “ ImageNet: A large-scale hierarchical image database,” in 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)(Miami, FL), 248–255. doi: 10. 1109/CVPR. 2009. 5206848

Detorakis, G., Sheik, S., Augustine, C., Paul, S., Pedroni, B. U., Dutt, N., et al. (2018). Neural and synaptic array transceiver: a brain-inspired computing framework for embedded learning. Front. Neurosci . 12: 583. doi: 10. 3389/fnins. 2018. 00583

Frenkel, C., Lefebvre, M., and Bol, D. (2019a). Learning without feedback: direct random target projection as a feedback-alignment algorithm with layerwise feedforward training. arXiv preprint arXiv: 1909. 01311 .

Frenkel, C., Lefebvre, M., Legat, J.-D., and Bol, D. (2019b). A 0. 086-mm 2 12. 7-pJ/SOP 64k-synapse 256-neuron online-learning digital spiking neuromorphic processor in 28-nm CMOS. IEEE Trans. Biomed. Circ. Syst . 13, 145–158. doi: 10. 1109/TBCAS. 2018. 2880425

Frenkel, C., Legat, J.-D., and Bol, D. (2019c). MorphIC: a 65-nm 738k-synapse/mm 2 quad-core binary-weight digital neuromorphic processor with stochastic spike-driven online learning. IEEE Trans. Biomed. Circ. Syst . 13, 999–1010. doi: 10. 1109/TBCAS. 2019. 2928793

Frenkel, C., Legat, J.-D., and Bol, D. (2020). “ A 28-nm convolutional neuromorphic processor enabling online learning with spike-based retinas,” in 2020 IEEE International Symposium on Circuits and Systems (ISCAS)(Sevilla: IEEE). doi: 10. 1109/ISCAS45731. 2020. 9180440

Grossberg, S. (1987). Competitive learning: From interactive activation to adaptive resonance. Cogn. Sci . 11, 23–63. doi: 10. 1111/j. 1551-6708. 1987. tb00862. x

Guerguiev, J., Lillicrap, T. P., and Richards, B. A. (2017). Towards deep learning with segregated dendrites. eLife 6: e22901. doi: 10. 7554/eLife. 22901

He, K., Zhang, X., Ren, S., and Sun, J. (2015). “ Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification,” in Proceedings of the IEEE International Conference on Computer Vision (Santiago), 1026–1034. doi: 10. 1109/ICCV. 2015. 123

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “ Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Las Vegas, NV), 770–778. doi: 10. 1109/CVPR. 2016. 90

Hinton, G. E., Deng, L., Yu, D., Dahl, G. E., Mohamed, A.-R., Jaitly, N., et al. (2012). Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. IEEE Signal Process. Mag . 29, 82–97. doi: 10. 1109/MSP. 2012. 2205597

Horowitz, M. (2014). “ Computing’s energy problem (and what we can do about it),” in IEEE International Solid-State Circuits Conference Digest of Technical Papers (ISSCC)(San Francisco, CA: IEEE), 10–14. doi: 10. 1109/ISSCC. 2014. 6757323

Ioffe, S., and Szegedy, C. (2015). Batch normalization: accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv: 1502. 03167 .

Jaderberg, M., Czarnecki, W. M., Osindero, S., Vinyals, O., Graves, A., Silver, D., et al. (2017). “ Decoupled neural interfaces using synthetic gradients,’ in Proceedings of the 34th International Conference on Machine Learning (Sydney, NSW), 1627–1635.

Kaiser, J., Mostafa, H., and Neftci, E. (2020). Synaptic plasticity dynamics for deep continuous local learning (DECOLLE). Front. Neurosci . 14: 424. doi: 10. 3389/fnins. 2020. 00424

Krizhevsky, A. Hinton, G. E., et al. (2009). Learning multiple layers of features from tiny images.

Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). “ ImageNet classification with deep convolutional neural networks,” in Advances in Neural Information Processing Systems , eds F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger (Lake Tahoe, NV: Curran Associates, Inc.), 1097–1105.

Launay, J., Poli, I., and Krzakala, F. (2019). Principled training of neural networks with direct feedback alignment. arXiv preprint arXiv: 1906. 04554 .

LeCun, Y., Bengio, Y., and Hinton, G. E. (2015). Deep learning. Nature 521, 436–444. doi: 10. 1038/nature14539

LeCun, Y., Cortes, C., and Burges, C. J. (1998). The MNIST database of handwritten digits.

Lee, D.-H., Zhang, S., Fischer, A., and Bengio, Y. (2015). “ Difference target propagation,” in Joint European Conference on Machine Learning and Knowledge Discovery in Databases (Würzburg), 498–515. doi: 10. 1007/978-3-319-23528-8_31

Liao, Q., Leibo, J. Z., and Poggio, T. (2016). “ How important is weight symmetry in backpropagation?” in Thirtieth AAAI Conference on Artificial Intelligence (Phoenix, AZ).

Lillicrap, T. P., Cownden, D., Tweed, D. B., and Akerman, C. J. (2016). Random synaptic feedback weights support error backpropagation for deep learning. Nat. Commun . 7, 1–10. doi: 10. 1038/ncomms13276

Magee, J. C., and Grienberger, C. (2020). Synaptic plasticity forms and functions. Annu. Rev. Neurosci . 43, 95–117. doi: 10. 1146/annurev-neuro-090919-022842

Milde, M. B., Blum, H., Dietmüller, A., Sumislawska, D., Conradt, J., Indiveri, G., et al. (2017). Obstacle avoidance and target acquisition for robot navigation using a mixed signal analog/digital neuromorphic processing system. Front. Neurorobot . 11: 28. doi: 10. 3389/fnbot. 2017. 00028

Minsky, M. (1961). Steps toward artificial intelligence. Proc. IRE 49, 8–30. doi: 10. 1109/JRPROC. 1961. 287775

Mostafa, H., Ramesh, V., and Cauwenberghs, G. (2018). Deep supervised learning using local errors. Front. Neurosci . 12: 608. doi: 10. 3389/fnins. 2018. 00608

Neftci, E. O., Augustine, C., Paul, S., and Detorakis, G. (2017). Event-driven random back-propagation: enabling neuromorphic deep learning machines. Front. Neurosci . 11: 324. doi: 10. 3389/fnins. 2017. 00324

Nøkland, A. (2016). “ Direct feedback alignment provides learning in deep neural networks,” in Advances in Neural Information Processing Systems , eds D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett (Barcelona: Curran Associates, Inc.) 1037–1045.

Nøkland, A., and Eidnes, L. H. (2019). “ Training neural networks with local error signals,” in Proceedings of the 36th International Conference on Machine Learning (Long Beach, CA), 4839–4850.

Ororbia, A. G., and Mali, A. (2019). “ Biologically motivated algorithms for propagating local target representations,” in Proceedings of the AAAI Conference on Artificial Intelligence (Honolulu, HI), 4651–4658. doi: 10. 1609/aaai. v33i01. 33014651

Park, J., Lee, J., and Jeon, D. (2019). A 65-nm neuromorphic image classification processor with energy-efficient training through direct spike-only feedback. IEEE J. Solid State Circ . 55, 108–119. doi: 10. 1109/JSSC. 2019. 2942367

Paszke, A., Gross, S., Chintala, S., Chanan, G., Yang, E., DeVito, Z., et al. (2017). “ Automatic differentiation in PyTorch,” in Proceedings of the 31st Conference of Neural Information Processing Systems (NIPS 2017)(Long Beach, CA).

Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., et al. (2011). Scikit-learn: machine learning in Python. J. Mach. Learn. Res . 12, 2825–2830.

Rajendran, B., Sebastian, A., Schmuker, M., Srinivasa, N., and Eleftheriou, E. (2019). Low-power neuromorphic hardware for signal processing applications: a review of architectural and system-level design approaches. IEEE Signal Process. Mag . 36, 97–110. doi: 10. 1109/MSP. 2019. 2933719

Rosenblatt, F. (1962). Principles of Neurodynamics: Perceptions and the Theory of Brain Mechanisms . Washington, DC: Spartan. doi: 10. 21236/AD0256582

Rumelhart, D. E., Hinton, G. E., and Williams, R. J. (1986). Learning representations by back-propagating errors. Nature 323, 533–536. doi: 10. 1038/323533a0

Thakur, C. S., Molin, J. L., Cauwenberghs, G., Indiveri, G., Kumar, K., Qiao, N., et al. (2018). Large-scale neuromorphic spiking array processors: a quest to mimic the brain. Front. Neurosci . 12: 891. doi: 10. 3389/fnins. 2018. 00891

Urbanczik, R., and Senn, W. (2014). Learning by the dendritic prediction of somatic spiking. Neuron 81, 521–528. doi: 10. 1016/j. neuron. 2013. 11. 030

Zenke, F., Agnes, E. J., and Gerstner, W. (2015). Diverse synaptic plasticity mechanisms orchestrated to form and retrieve memories in spiking neural networks. Nat. Commun . 6, 1–13. doi: 10. 1038/ncomms7922

Thank's for Your Vote!
Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks. Page 1
Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks. Page 2
Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks. Page 3
Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks. Page 4
Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks. Page 5
Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks. Page 6
Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks. Page 7
Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks. Page 8
Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks. Page 9

This work, titled "Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks" was written and willingly shared by a fellow student. This sample can be utilized as a research and reference resource to aid in the writing of your own work. Any use of the work that does not include an appropriate citation is banned.

If you are the owner of this work and don’t want it to be published on AssignBuster, request its removal.

Request Removal
Cite this Essay

References

AssignBuster. (2022) 'Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks'. 8 October.

Reference

AssignBuster. (2022, October 8). Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks. Retrieved from https://assignbuster.com/learning-without-feedback-fixed-random-learning-signals-allow-for-feedforward-training-of-deep-neural-networks/

References

AssignBuster. 2022. "Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks." October 8, 2022. https://assignbuster.com/learning-without-feedback-fixed-random-learning-signals-allow-for-feedforward-training-of-deep-neural-networks/.

1. AssignBuster. "Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks." October 8, 2022. https://assignbuster.com/learning-without-feedback-fixed-random-learning-signals-allow-for-feedforward-training-of-deep-neural-networks/.


Bibliography


AssignBuster. "Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks." October 8, 2022. https://assignbuster.com/learning-without-feedback-fixed-random-learning-signals-allow-for-feedforward-training-of-deep-neural-networks/.

Work Cited

"Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks." AssignBuster, 8 Oct. 2022, assignbuster.com/learning-without-feedback-fixed-random-learning-signals-allow-for-feedforward-training-of-deep-neural-networks/.

Get in Touch

Please, let us know if you have any ideas on improving Learning without feedback: fixed random learning signals allow for feedforward training of deep neural networks, or our service. We will be happy to hear what you think: [email protected]