mirror of
https://github.com/2martens/uni.git
synced 2026-05-06 11:26:25 +02:00
[NN] Explained approaches
Signed-off-by: Jim Martens <github@2martens.de>
This commit is contained in:
@ -221,9 +221,9 @@ Kirkpatrick\cite{Kirkpatrick2017}, Velez\cite{Velez2017} and Shmelkov\cite{Shmel
|
|||||||
|
|
||||||
In the context of this paper plasticity refers to synaptic plasticity as described
|
In the context of this paper plasticity refers to synaptic plasticity as described
|
||||||
in Citri\cite{Citri2008}. The process of learning itself, changing the weights,
|
in Citri\cite{Citri2008}. The process of learning itself, changing the weights,
|
||||||
is already considered plasticity. It is important however that the weights can
|
is already considered plasticity. This can occur throughout the lifetime of a
|
||||||
be changed during the lifetime of the neural network. Ordinary neural networks
|
network or during the training phase of networks using for example supervised learning
|
||||||
trained with backpropagation are not involving plasticity.
|
and backpropagation.
|
||||||
|
|
||||||
\section{Catastrophic Forgetting}
|
\section{Catastrophic Forgetting}
|
||||||
\label{sec:catastrophicforgetting}
|
\label{sec:catastrophicforgetting}
|
||||||
@ -296,26 +296,139 @@ these newest approaches will not be described here.
|
|||||||
\section{Plasticity}
|
\section{Plasticity}
|
||||||
\label{sec:plasticity}
|
\label{sec:plasticity}
|
||||||
|
|
||||||
Plasticity can be realized by various approaches. Here three approaches are
|
Every neural network involves a learning aspect and hence plasticity, given our
|
||||||
presented, Modulated Random Search, Modulated Gaussian Walk and Diffusion based
|
definition of it. In this section three approaches for plasticity using diffusion-based
|
||||||
neuromodulation.
|
neuromodulation are presented in more detail. Modulated Random Search and
|
||||||
|
Modulated Gaussian Walk are using linearly modulated neural networks. They
|
||||||
|
are taken from Toutounji and Pasemann\cite{Toutounji2016}. The third approach
|
||||||
|
was introduced by Velez and Clune\cite{Velez2017} uses diffusion-based neuromodulation
|
||||||
|
for localized learning hence the name of the subsection here.
|
||||||
|
|
||||||
\subsection{Modulated Random Search}
|
\subsection{Modulated Random Search}
|
||||||
\label{subsec:mrs}
|
\label{subsec:mrs}
|
||||||
|
|
||||||
Does things.
|
\subsubsection*{Modulated Neural Network}
|
||||||
|
|
||||||
|
Since both approaches from Toutounji and Pasemann are using linearly-modulated
|
||||||
|
neural networks the structure of these networks is described first. Linearly-modulated
|
||||||
|
neural networks (LMNN) are a specific variant of modulated neural
|
||||||
|
networks (MNN). Any artifical neural network (ANN) or simply neural network
|
||||||
|
in the context of Computer Science can become a modulated neural network by
|
||||||
|
adding a neuromodulator layer.
|
||||||
|
|
||||||
|
Toutounji and Pasemann describe a variant of this layer that uses neuromodulator
|
||||||
|
cells (NMCs). Each NMC produces a specific type of neuromodulator (NM) and
|
||||||
|
saves its own concentration level of it. The network wide concentration level at
|
||||||
|
a certain point in space and time can be obtained by summing up all concentration
|
||||||
|
levels saved in NMCs at the point in space. Produced neuromodulators usually
|
||||||
|
impact nearby network parts. This type of spatial impact requires a spatial
|
||||||
|
representation in the network where all network elements have a location in
|
||||||
|
the space.
|
||||||
|
|
||||||
|
There are a production and a reduction mode for the NMCs. During the production
|
||||||
|
mode the concentration of neuromodulator can be increased and during reduction
|
||||||
|
mode it can be decreased. A cell can enter production mode if it was stimulated
|
||||||
|
for some time while it falls back to reduction mode when this stimulation
|
||||||
|
does not happen for some time.
|
||||||
|
|
||||||
|
\subsubsection*{Linearly-Modulated Neural Network}
|
||||||
|
|
||||||
|
A linearly-modulated neural network uses discrete time and stimulates NMCs
|
||||||
|
with a simple linear model. Each NMC is connected to a carrier cell or neuron
|
||||||
|
which itself is part of a modulatory subnetwork. The NMC is stimulated if the
|
||||||
|
output of the carrier neuron is within a specified range
|
||||||
|
(\(\text{S}^{\text{min}}\), \(\text{S}^{\text{max}}\)). In every time step
|
||||||
|
is checked if the output of the carrier cell is high enough to stimulate the
|
||||||
|
NMC. If it is the stimulation level of the NMC increases. Otherwise it decreases.
|
||||||
|
Once the stimulation level reaches the threshold \(\text{T}^{\text{prod}}\)
|
||||||
|
the cell goes into the production mode. If it falls below \(\text{T}^{\text{red}}\)
|
||||||
|
the cell goes back into reduction mode.
|
||||||
|
Over time the neuromodulator diffuses to the surrounding control subnetwork
|
||||||
|
where it initiates plasticity that is dependent on the concentration of it at
|
||||||
|
the respective synapse.
|
||||||
|
|
||||||
|
\subsubsection*{Modulated Random Search}
|
||||||
|
|
||||||
|
\begin{table}
|
||||||
|
\begin{tabular}{l|l}
|
||||||
|
\textbf{Parameter} & \textbf{Description} \\
|
||||||
|
\(Type\) & The neuromodulator type the synapse is sensitive to \\
|
||||||
|
\(W\) & Weight change probability \\
|
||||||
|
\(D\) & Disable / enable probability \\
|
||||||
|
\(W^{min}, W^{max}\) & Minimum and maximum weight of synapse \\
|
||||||
|
\(M\) & Maximum neuromodulator sensitivity limit of the synapse
|
||||||
|
\end{tabular}
|
||||||
|
\caption{Parameters stored for each synapse.
|
||||||
|
Replication of Table 1 in Toutunji and Pasemann\cite{Toutounji2016}.}
|
||||||
|
\label{tab:mrs-synapse}
|
||||||
|
\end{table}
|
||||||
|
|
||||||
|
Modulated random search means essentially random weight changes. Each synapse
|
||||||
|
has some parameters that are used (see table \ref{tab:mrs-synapse}). The weight
|
||||||
|
change probability is the product of the intrinsic weight change probability
|
||||||
|
and the concentration of the neuromodulator the synapse is sensitive to
|
||||||
|
at its location. Additionally the maximum neuromodulator sensitivity is
|
||||||
|
the ceiling for the second part of that product. This means there is a maximum
|
||||||
|
weight change probability for each synapse. Should a weight change occur a new
|
||||||
|
weight is chosen randomly from the range of values described by the minimum and
|
||||||
|
maximum weight of the synapse.
|
||||||
|
|
||||||
|
Moreover a synapse can disable or enable itself. The actual disable/enable
|
||||||
|
probability is the product of the intrinsic value saved as parameter and
|
||||||
|
the neuromodulator concentration. The concentration is again ceiled by the
|
||||||
|
maximum sensitivity limit given as parameter. This means there is a maximum
|
||||||
|
disable/enable probability as well. A disabled synapse is treated as having
|
||||||
|
weight 0 but the actual value is stored so that it can be restored when the
|
||||||
|
synapse is enabled again.
|
||||||
|
|
||||||
|
Given a so called neural network structure or substrate this makes it easier
|
||||||
|
to find different network topologies (structure and weights combined).
|
||||||
|
|
||||||
\subsection{Modulated Gaussian Walk}
|
\subsection{Modulated Gaussian Walk}
|
||||||
\label{subsec:mgw}
|
\label{subsec:mgw}
|
||||||
|
|
||||||
Does things more efficient.
|
Toutounji and Pasemann introduce the modulated gaussian walk. The key differences
|
||||||
|
start with the parameters. There is no maximum sensitivity for the neuromodulator
|
||||||
|
concentration. When a weight change occurs the new weight is not chosen randomly
|
||||||
|
but rather the difference to be added to the current weight is sampled from a
|
||||||
|
normal distribution with a mean of zero and \(\sigma^2\)-variance. The sampled
|
||||||
|
value could be infinitely large and hence the new weight outside of the given
|
||||||
|
bounds for it. Therefore the value is sampled until the sum of the
|
||||||
|
current weight and the sampled value are within the range.
|
||||||
|
|
||||||
|
Toutunji and Pasemann implemented a mechanism for disabling synapses
|
||||||
|
in the modulated gaussian walk as well but did not make use of it later and
|
||||||
|
therefore they did not describe how it works.
|
||||||
|
|
||||||
\subsection{Localized learning}
|
\subsection{Localized learning}
|
||||||
\label{subsec:diffusion}
|
\label{subsec:diffusion}
|
||||||
|
|
||||||
Velez describe another approach that employs
|
Velez and Clune are using a small network to solve the foraging task. The network
|
||||||
modularity for the learning. Essentially this results in task-specific localized
|
represents an agent that has a lifetime of three years. Each year consists of
|
||||||
learning and functional modules for each subtask.
|
the seasons summer and winter. During each season the agent is presented with
|
||||||
|
food and has to either eat the food or not. Half of the food is nutritious
|
||||||
|
and the other poisonous. The target is a fitness value which is best if the
|
||||||
|
agent eats all nutritious food and none of the poisonous. The associations of
|
||||||
|
nutritious and poisonous are different between summer and winter but within
|
||||||
|
a season remain the same during the lifetime. Therefore a nutritious food in
|
||||||
|
summer will always be nutritious.
|
||||||
|
This setup makes it easy to measure if the agent is able to remember the learned
|
||||||
|
associations from the previous seasons.
|
||||||
|
|
||||||
|
The initial weights of the network are derived from an evolutionary algorithm.
|
||||||
|
All later learning uses neuromodulation. The neurons of the network are spatially
|
||||||
|
located and there are two sources of neuromodulators in the network - one on either
|
||||||
|
side. The sources are only active in their respective season and encode whether
|
||||||
|
the previously eaten food was nutritious (1) or poisonous (-1). If they are
|
||||||
|
not active their value is zero. As soon as the sources are activated the neuromodulators
|
||||||
|
fill a space within a radius of 1.5 units of distance from the source and potentially
|
||||||
|
trigger weight changes of neurons inside the radius. The strength of the neuromodulators
|
||||||
|
is decreasing with further distance from the source.
|
||||||
|
|
||||||
|
This explanation should suffice for the general understanding of their method.
|
||||||
|
The neurons within the vicinity of these sources only update their weights
|
||||||
|
in one of the seasons. Therefore they only learn for one season and are unaffected
|
||||||
|
by the other season. This results in a localized learning.
|
||||||
|
|
||||||
\section{Comparison regarding catastrophic forgetting}
|
\section{Comparison regarding catastrophic forgetting}
|
||||||
\label{sec:comparison}
|
\label{sec:comparison}
|
||||||
|
|||||||
Reference in New Issue
Block a user