From 5c942b31786548cc63aa0ac09be3ec0ba725f6c0 Mon Sep 17 00:00:00 2001 From: Jim Martens Date: Thu, 24 May 2018 12:48:08 +0200 Subject: [PATCH] [NN] Added actual comparison between methods Signed-off-by: Jim Martens --- neural-networks/seminarpaper.tex | 85 +++++++++++++++++++++++++------- 1 file changed, 66 insertions(+), 19 deletions(-) diff --git a/neural-networks/seminarpaper.tex b/neural-networks/seminarpaper.tex index f91946f..d8eda16 100644 --- a/neural-networks/seminarpaper.tex +++ b/neural-networks/seminarpaper.tex @@ -367,7 +367,7 @@ the respective synapse. \(M\) & Maximum neuromodulator sensitivity limit of the synapse \end{tabular} \caption{Parameters stored for each synapse. - Replication of Table 1 in Toutunji and Pasemann\cite{Toutounji2016}.} + Replication of Table 1 in Toutounji and Pasemann\cite{Toutounji2016}.} \label{tab:mrs-synapse} \end{table} @@ -424,7 +424,7 @@ current weight and the sampled value are within the interval \([W_i^{min}, W_i^{ w_i (t + 1) = w_i (t) + \Delta w_i \;\text{where}\; \Delta w_i \sim \mathcal{N}(0, \sigma^2) \] -Toutunji and Pasemann implemented a mechanism for disabling synapses +Toutounji and Pasemann implemented a mechanism for disabling synapses in the modulated gaussian walk as well but did not make use of it later and therefore they did not describe how it works. @@ -457,7 +457,7 @@ it when to learn. How does the actual learning happen? The weight change between two neurons is dependent on the activation of both neurons, the learning rate and the concentration -of neuromodulators. In short Hebbian learning is employed. +of neuromodulators. In short Hebbian learning is employed. \[ \Delta w_{ij} = \eta \cdot m_i \cdot a_i \cdot a_j @@ -474,7 +474,7 @@ by the other season. This results in a localized learning. In this section the three presented approaches for plasticity are compared with regard to their ability to mitigate or overcome catastrophic forgetting. For both the modulated random search and the modulated gaussian walk this aspect was -analyzed in the experiments conducted by Toutunji and Pasemann\cite{Toutounji2016}. +analyzed in the experiments conducted by Toutounji and Pasemann\cite{Toutounji2016}. Therefore the results of their work will be utilized for this comparison. Velez and Clune\cite{Velez2017} introduced the presented approach of localized learning to analyze its capability with respect to overcoming catastrophic @@ -526,27 +526,74 @@ Modulated gaussian walk is a clear improvement compared to the random search in the analyzed experiments. While all three approaches used diffusion-based neuromodulation the first two and -the third are quite different in their setup. For the future it would be interesting -to combine localized learning with gaussian walk on the experiments of Toutunji -and Pasemann. In particular the combined experiment might benefit from this as -localized learning could separate the learning for one task from the other -and gaussian walk could then improve the particular part that was problematic. +the third are quite different in their setup. First the general neuromodulation +architecture was different (each neuron can diffuse neuromodulators vs. two stationary +sources) and second the actual weight change was also different. Both modulated +gaussian walk and the shown localized learning are improving previous weights +instead of completely changing them. The gaussian walk is using a normal distribution +to get the weight change while the localized learning uses Hebbian learning and +therefore is dependent on the activations of two neurons and is directly incorporating +the neuromodulators in the weight change formula itself. + +It is important to note that the advantage of gaussian walk had nothing to do +with the architecture of the neuromodulation as that was identical for both +random search and gaussian walk. The improvement originated in the learning +rule. In the experiments of Toutounji and Pasemann they used a homogenous +diffusion but it would have been possible to use different diffusion strengths, +decays and so on for every neuron. + +For the future it would be interesting to +compare the LMNN architecture with the "sources" architecture of localized +learning to understand the impact of the neuromodulation architecture. +In addition the gaussian walk learning rule should be compared with the +Hebbian learning rule used by localized learning. + +For the E3 experiment used by Toutounji and Pasemann it is safe to assume that +the kind of a priori placement of neuromodulator sources won't work. The experiment +requires the robot to solve two tasks at the same time: It has +to approach the lights and avoid obstacles. Since both tasks need to be solved +at the same time and always it is not possible to devise two "seasons" or some +similar seperation of learning time. Therefore the LMNN architecture is likely +better suited. + +Nevertheless it would make sense to separate the learning for +these two tasks as a robot might already be very good at approaching lights but +only mediocre at avoiding obstacles. In that case the improvements for the second +task should not impact the first task. The Hebbian learning rule is likely +better suited to achieve this effect as it correlates the weight change with +the correlation of the connected neurons. Simply using a value sampled from a +normal distribution as the gaussian walk does it, probably does not result in +localized learning. On the other hand localized learning will likely only work +if it is possible to give the robot distinct feedback about its performance in +each task. If it only receives a combined feedback it is more difficult to +utilize localized learning as it is then not easy to find out which part (and +therefore which weights) performed bad. + +In situations where there is only one task to solve (E2) or the feedback is only +given as a total without distinct information about each sub task it is very +likely that localized learning won't work and therefore gaussian walk is better +suited than Hebbian learning. \section{Conclusion} \label{sec:concl} A second environmental feedback loop is important to tell autonomous systems -when to learn. But the method to learn is important as well to be any use -in a practical environment. The comparison has shown that localized learning -can overcome catastrophic forgetting for small networks in a very restricted -setup. Furthermore the comparison revealed that modulated random search is -not part of a solution to catastrophic forgetting and modulated gaussian -walk is significantly better in that regard. +when to learn. But it is important how this feedback loop is working. Furthermore +it is important how the learning actually works. The comparison has shown that +localized learning utilizing neuromodulator sources can overcome catastrophic +forgetting for small networks in a very restricted setup. Furthermore the +comparison revealed that modulated random search is not part of a solution to +catastrophic forgetting. In a more general case it is likely that the LMNN +architecture is better than the sources architecture and that Hebbian learning +is better suited for combined tasks and localized learning than modulated gaussian +walk. For single task environments or those where localized learning is not an +option modulated gaussian walk is likely better suited than Hebbian learning. -Future work should look into the effects of localized learning on the kind of -autonomous robot experiments that were conducted by Toutunji and Pasemann and -in general research the applicability to bigger problems for example in the area -of deep neural networks. +Future work should look into the guesses that were taken here and analyze which +network architecture is better and which learning rule is better for the kind of +autonomous robot experiments that were conducted by Toutounji and Pasemann. In +general the applicability of localized learning to bigger problems for example +in the area of deep neural networks should be researched. %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % hier werden - zum Ende des Textes - die bibliographischen Referenzen