1
0
mirror of https://github.com/2martens/uni.git synced 2026-05-06 19:36:26 +02:00

[NN] Explained approaches

Signed-off-by: Jim Martens <github@2martens.de>
This commit is contained in:
2018-05-10 14:54:47 +02:00
parent 9325badbe3
commit 1bcdb8d937

View File

@ -221,9 +221,9 @@ Kirkpatrick\cite{Kirkpatrick2017}, Velez\cite{Velez2017} and Shmelkov\cite{Shmel
In the context of this paper plasticity refers to synaptic plasticity as described In the context of this paper plasticity refers to synaptic plasticity as described
in Citri\cite{Citri2008}. The process of learning itself, changing the weights, in Citri\cite{Citri2008}. The process of learning itself, changing the weights,
is already considered plasticity. It is important however that the weights can is already considered plasticity. This can occur throughout the lifetime of a
be changed during the lifetime of the neural network. Ordinary neural networks network or during the training phase of networks using for example supervised learning
trained with backpropagation are not involving plasticity. and backpropagation.
\section{Catastrophic Forgetting} \section{Catastrophic Forgetting}
\label{sec:catastrophicforgetting} \label{sec:catastrophicforgetting}
@ -296,26 +296,139 @@ these newest approaches will not be described here.
\section{Plasticity} \section{Plasticity}
\label{sec:plasticity} \label{sec:plasticity}
Plasticity can be realized by various approaches. Here three approaches are Every neural network involves a learning aspect and hence plasticity, given our
presented, Modulated Random Search, Modulated Gaussian Walk and Diffusion based definition of it. In this section three approaches for plasticity using diffusion-based
neuromodulation. neuromodulation are presented in more detail. Modulated Random Search and
Modulated Gaussian Walk are using linearly modulated neural networks. They
are taken from Toutounji and Pasemann\cite{Toutounji2016}. The third approach
was introduced by Velez and Clune\cite{Velez2017} uses diffusion-based neuromodulation
for localized learning hence the name of the subsection here.
\subsection{Modulated Random Search} \subsection{Modulated Random Search}
\label{subsec:mrs} \label{subsec:mrs}
Does things. \subsubsection*{Modulated Neural Network}
Since both approaches from Toutounji and Pasemann are using linearly-modulated
neural networks the structure of these networks is described first. Linearly-modulated
neural networks (LMNN) are a specific variant of modulated neural
networks (MNN). Any artifical neural network (ANN) or simply neural network
in the context of Computer Science can become a modulated neural network by
adding a neuromodulator layer.
Toutounji and Pasemann describe a variant of this layer that uses neuromodulator
cells (NMCs). Each NMC produces a specific type of neuromodulator (NM) and
saves its own concentration level of it. The network wide concentration level at
a certain point in space and time can be obtained by summing up all concentration
levels saved in NMCs at the point in space. Produced neuromodulators usually
impact nearby network parts. This type of spatial impact requires a spatial
representation in the network where all network elements have a location in
the space.
There are a production and a reduction mode for the NMCs. During the production
mode the concentration of neuromodulator can be increased and during reduction
mode it can be decreased. A cell can enter production mode if it was stimulated
for some time while it falls back to reduction mode when this stimulation
does not happen for some time.
\subsubsection*{Linearly-Modulated Neural Network}
A linearly-modulated neural network uses discrete time and stimulates NMCs
with a simple linear model. Each NMC is connected to a carrier cell or neuron
which itself is part of a modulatory subnetwork. The NMC is stimulated if the
output of the carrier neuron is within a specified range
(\(\text{S}^{\text{min}}\), \(\text{S}^{\text{max}}\)). In every time step
is checked if the output of the carrier cell is high enough to stimulate the
NMC. If it is the stimulation level of the NMC increases. Otherwise it decreases.
Once the stimulation level reaches the threshold \(\text{T}^{\text{prod}}\)
the cell goes into the production mode. If it falls below \(\text{T}^{\text{red}}\)
the cell goes back into reduction mode.
Over time the neuromodulator diffuses to the surrounding control subnetwork
where it initiates plasticity that is dependent on the concentration of it at
the respective synapse.
\subsubsection*{Modulated Random Search}
\begin{table}
\begin{tabular}{l|l}
\textbf{Parameter} & \textbf{Description} \\
\(Type\) & The neuromodulator type the synapse is sensitive to \\
\(W\) & Weight change probability \\
\(D\) & Disable / enable probability \\
\(W^{min}, W^{max}\) & Minimum and maximum weight of synapse \\
\(M\) & Maximum neuromodulator sensitivity limit of the synapse
\end{tabular}
\caption{Parameters stored for each synapse.
Replication of Table 1 in Toutunji and Pasemann\cite{Toutounji2016}.}
\label{tab:mrs-synapse}
\end{table}
Modulated random search means essentially random weight changes. Each synapse
has some parameters that are used (see table \ref{tab:mrs-synapse}). The weight
change probability is the product of the intrinsic weight change probability
and the concentration of the neuromodulator the synapse is sensitive to
at its location. Additionally the maximum neuromodulator sensitivity is
the ceiling for the second part of that product. This means there is a maximum
weight change probability for each synapse. Should a weight change occur a new
weight is chosen randomly from the range of values described by the minimum and
maximum weight of the synapse.
Moreover a synapse can disable or enable itself. The actual disable/enable
probability is the product of the intrinsic value saved as parameter and
the neuromodulator concentration. The concentration is again ceiled by the
maximum sensitivity limit given as parameter. This means there is a maximum
disable/enable probability as well. A disabled synapse is treated as having
weight 0 but the actual value is stored so that it can be restored when the
synapse is enabled again.
Given a so called neural network structure or substrate this makes it easier
to find different network topologies (structure and weights combined).
\subsection{Modulated Gaussian Walk} \subsection{Modulated Gaussian Walk}
\label{subsec:mgw} \label{subsec:mgw}
Does things more efficient. Toutounji and Pasemann introduce the modulated gaussian walk. The key differences
start with the parameters. There is no maximum sensitivity for the neuromodulator
concentration. When a weight change occurs the new weight is not chosen randomly
but rather the difference to be added to the current weight is sampled from a
normal distribution with a mean of zero and \(\sigma^2\)-variance. The sampled
value could be infinitely large and hence the new weight outside of the given
bounds for it. Therefore the value is sampled until the sum of the
current weight and the sampled value are within the range.
Toutunji and Pasemann implemented a mechanism for disabling synapses
in the modulated gaussian walk as well but did not make use of it later and
therefore they did not describe how it works.
\subsection{Localized learning} \subsection{Localized learning}
\label{subsec:diffusion} \label{subsec:diffusion}
Velez describe another approach that employs Velez and Clune are using a small network to solve the foraging task. The network
modularity for the learning. Essentially this results in task-specific localized represents an agent that has a lifetime of three years. Each year consists of
learning and functional modules for each subtask. the seasons summer and winter. During each season the agent is presented with
food and has to either eat the food or not. Half of the food is nutritious
and the other poisonous. The target is a fitness value which is best if the
agent eats all nutritious food and none of the poisonous. The associations of
nutritious and poisonous are different between summer and winter but within
a season remain the same during the lifetime. Therefore a nutritious food in
summer will always be nutritious.
This setup makes it easy to measure if the agent is able to remember the learned
associations from the previous seasons.
The initial weights of the network are derived from an evolutionary algorithm.
All later learning uses neuromodulation. The neurons of the network are spatially
located and there are two sources of neuromodulators in the network - one on either
side. The sources are only active in their respective season and encode whether
the previously eaten food was nutritious (1) or poisonous (-1). If they are
not active their value is zero. As soon as the sources are activated the neuromodulators
fill a space within a radius of 1.5 units of distance from the source and potentially
trigger weight changes of neurons inside the radius. The strength of the neuromodulators
is decreasing with further distance from the source.
This explanation should suffice for the general understanding of their method.
The neurons within the vicinity of these sources only update their weights
in one of the seasons. Therefore they only learn for one season and are unaffected
by the other season. This results in a localized learning.
\section{Comparison regarding catastrophic forgetting} \section{Comparison regarding catastrophic forgetting}
\label{sec:comparison} \label{sec:comparison}