1
0
mirror of https://github.com/2martens/uni.git synced 2026-05-06 19:36:26 +02:00

[NN] Improved reading flow

Signed-off-by: Jim Martens <github@2martens.de>
This commit is contained in:
2018-05-24 13:07:54 +02:00
parent 13aac2fac3
commit 13245b802f

View File

@ -215,7 +215,7 @@ maxnames=2
Autonomous robots need to adapt to new situations. They have a need to learn Autonomous robots need to adapt to new situations. They have a need to learn
for an entire life. In order to do this they need a second environmental feedback for an entire life. In order to do this they need a second environmental feedback
loop that tells them when to learn.\cite{Toutounji2016} loop that tells them when to learn\cite{Toutounji2016}.
The learning itself is also described as plasticity. In the context of this paper The learning itself is also described as plasticity. In the context of this paper
the definition of synaptic plasticity given by Citri\cite{Citri2008} will be used. the definition of synaptic plasticity given by Citri\cite{Citri2008} will be used.
@ -226,7 +226,9 @@ and backpropagation.
When a network has to adapt to new situations, it has to learn new tasks. Usually When a network has to adapt to new situations, it has to learn new tasks. Usually
the previously learned weights are largely forgotten. This phenomenon is called the previously learned weights are largely forgotten. This phenomenon is called
catastrophic forgetting.\cite{French1999,McCloskey1989} catastrophic forgetting\cite{French1999,McCloskey1989}. It is highly problematic,
because the weights encode the learning of a network. If they are forgotten or
rather overwritten the previously learned tasks cannot be fulfilled anymore.
Since catastrophic forgetting is a key problem for autonomous learning, it is Since catastrophic forgetting is a key problem for autonomous learning, it is
crucial to overcome it. In this paper I will present some approaches for crucial to overcome it. In this paper I will present some approaches for
@ -236,8 +238,9 @@ catastrophic forgetting.
\section{Catastrophic Forgetting} \section{Catastrophic Forgetting}
\label{sec:catastrophicforgetting} \label{sec:catastrophicforgetting}
This section presents the major research developments related to catastrophic French\cite{French1999} did a review of the existing research about catastrophic
forgetting and explains what it actually is. It follows the review of French\cite{French1999}. forgetting. The following paragraphs will follow this review and highlight the
major developments in research related to catastrophic forgetting.
McCloskey and Cohen\cite{McCloskey1989} originally discovered the problem of McCloskey and Cohen\cite{McCloskey1989} originally discovered the problem of
catastrophic forgetting, which was referred to as catastrophic interference. This catastrophic forgetting, which was referred to as catastrophic interference. This
@ -305,12 +308,13 @@ catastrophic forgetting at bay, has to be named.
\section{Plasticity} \section{Plasticity}
\label{sec:plasticity} \label{sec:plasticity}
Every neural network involves a learning aspect and hence plasticity, given our Catastrophic forgetting requires learned weights that can be forgotten.
Every neural network learns and therefore deals with plasticity, given our
definition of it. In this section three approaches for plasticity using diffusion-based definition of it. In this section three approaches for plasticity using diffusion-based
neuromodulation are presented in more detail. Modulated Random Search and neuromodulation are presented in more detail. Modulated Random Search and
Modulated Gaussian Walk are using linearly modulated neural networks. They Modulated Gaussian Walk are using linearly modulated neural networks. They
are taken from Toutounji and Pasemann\cite{Toutounji2016}. The third approach are taken from Toutounji and Pasemann\cite{Toutounji2016}. The third approach
was introduced by Velez and Clune\cite{Velez2017} uses diffusion-based neuromodulation was introduced by Velez and Clune\cite{Velez2017} and uses diffusion-based neuromodulation
for localized learning hence the name of the subsection here. for localized learning hence the name of the subsection here.
\subsection{Modulated Random Search} \subsection{Modulated Random Search}
@ -413,7 +417,7 @@ to find different network topologies (structure and weights combined).
\subsection{Modulated Gaussian Walk} \subsection{Modulated Gaussian Walk}
\label{subsec:mgw} \label{subsec:mgw}
Toutounji and Pasemann introduce the modulated gaussian walk. The key differences The modulated gaussian walk is introduced by Toutounji and Pasemann. The key differences
start with the parameters. There is no maximum sensitivity for the neuromodulator start with the parameters. There is no maximum sensitivity for the neuromodulator
concentration. When a weight change occurs the new weight is not chosen randomly concentration. When a weight change occurs the new weight is not chosen randomly
but rather the difference to be added to the current weight is sampled from a but rather the difference to be added to the current weight is sampled from a
@ -473,10 +477,10 @@ by the other season. This results in a localized learning.
\section{Comparison regarding catastrophic forgetting} \section{Comparison regarding catastrophic forgetting}
\label{sec:comparison} \label{sec:comparison}
In this section the three presented approaches for plasticity are compared with Given the presentations of the three approaches it is interesting to compare
regard to their ability to mitigate or overcome catastrophic forgetting. For both them with regard to their ability to mitigate or overcome catastrophic forgetting.
the modulated random search and the modulated gaussian walk this aspect was For both the modulated random search and the modulated gaussian walk this aspect
analyzed in the experiments conducted by Toutounji and Pasemann\cite{Toutounji2016}. was analyzed in the experiments conducted by Toutounji and Pasemann\cite{Toutounji2016}.
Therefore the results of their work will be utilized for this comparison. Therefore the results of their work will be utilized for this comparison.
Velez and Clune\cite{Velez2017} introduced the presented approach of localized Velez and Clune\cite{Velez2017} introduced the presented approach of localized
learning to analyze its capability with respect to overcoming catastrophic learning to analyze its capability with respect to overcoming catastrophic