Updated introduction and made it more concise
Signed-off-by: Jim Martens <github@2martens.de>
This commit is contained in:
48
body.tex
48
body.tex
@ -8,34 +8,21 @@ Famous examples like the automatic soap dispenser which does not
|
|||||||
recognize the hand of a black person but dispenses soap when presented
|
recognize the hand of a black person but dispenses soap when presented
|
||||||
with a paper towel raise the question of bias in computer
|
with a paper towel raise the question of bias in computer
|
||||||
systems~\cite{Friedman1996}. Related to this ethical question regarding
|
systems~\cite{Friedman1996}. Related to this ethical question regarding
|
||||||
the design of so called algorithms, a term often used in public
|
the design of so called algorithms is the question of
|
||||||
discourse for applied neural networks, is the question of
|
|
||||||
algorithmic accountability~\cite{Diakopoulos2014}.
|
algorithmic accountability~\cite{Diakopoulos2014}.
|
||||||
|
|
||||||
The charm of supervised neural networks, that they can learn from
|
Supervised neural networks learn from input-output relations and
|
||||||
input-output relations and figure out by themselves what connections
|
figure out by themselves what connections are necessary for that.
|
||||||
are necessary for that, is also their Achilles heel. This feature
|
This feature is also their Achilles heel: it makes them effectively
|
||||||
makes them effectively black boxes. It is possible to question the
|
black boxes and prevents any answers to questions of causality.
|
||||||
training environment, like potential biases inside the data sets, or
|
|
||||||
the engineers constructing the networks but it is not really possible
|
|
||||||
to question the internal calculations made by a network. On the one
|
|
||||||
hand, one might argue, it is only math and nothing magical that
|
|
||||||
happens inside these networks. Clearly it is possible, albeit a chore,
|
|
||||||
to manually follow the calculations of any given trained network.
|
|
||||||
After all it is executed on a computer and at the lowest level only
|
|
||||||
uses basic math that does not differ between humans and computers. On
|
|
||||||
the other hand not everyone is capable of doing so and more
|
|
||||||
importantly it does not reveal any answers to questions of causality.
|
|
||||||
|
|
||||||
However, these questions of causility are of enormous consequence when
|
However, these questions of causility are of enormous consequence when
|
||||||
neural networks are used, for example, in predictive policing. Is a
|
results of neural networks are used to make life changing decisions:
|
||||||
correlation, a coincidence, enough to bring forth negative consequences
|
Is a correlation enough to bring forth negative consequences
|
||||||
for a particular person? And if so, what is the possible defence
|
for a particular person? And if so, what is the possible defence
|
||||||
against math? Similar questions can be raised when looking at computer
|
against math? Similar questions can be raised when looking at computer
|
||||||
vision networks that might be used together with so called smart
|
vision networks that might be used together with so called smart
|
||||||
CCTV cameras, for example, like those tested at the train station
|
CCTV cameras to discover suspicious activity.
|
||||||
Berlin Südkreuz. What if a network implies you committed suspicious
|
|
||||||
behaviour?
|
|
||||||
|
|
||||||
This leads to the need for neural networks to explain their results.
|
This leads to the need for neural networks to explain their results.
|
||||||
Such an explanation must come from the network or an attached piece
|
Such an explanation must come from the network or an attached piece
|
||||||
@ -103,17 +90,17 @@ representation of the input and has to find a decompression
|
|||||||
that reconstructs the input as accurate as possible. During
|
that reconstructs the input as accurate as possible. During
|
||||||
training these auto-encoders learn to reproduce a certain group
|
training these auto-encoders learn to reproduce a certain group
|
||||||
of object classes. The actual novelty detection takes place
|
of object classes. The actual novelty detection takes place
|
||||||
during testing. Given an image, and the output and loss of the
|
during testing: Given an image, and the output and loss of the
|
||||||
auto-encoder, a novelty score is calculated. A low novelty
|
auto-encoder, a novelty score is calculated. A low novelty
|
||||||
score signals a known object. The opposite is true for a high
|
score signals a known object. The opposite is true for a high
|
||||||
novelty score.
|
novelty score.
|
||||||
|
|
||||||
\subsection*{Research Question}
|
\subsection*{Research Question}
|
||||||
|
|
||||||
Given these two approaches to solve the explanation task of above,
|
Both presented approaches describe one way to solve the aforementioned
|
||||||
it comes down to performance. At the end of the day the best
|
problem of explanation. They can be differentiated by measuring
|
||||||
theoretical idea does not help in solving the task if it cannot
|
their performance: the best theoretical idea is useless if it does
|
||||||
be implemented in a performant way. Miller et al. have shown
|
not perform well. Miller et al. have shown
|
||||||
some success in using dropout sampling. However, the many forward
|
some success in using dropout sampling. However, the many forward
|
||||||
passes during testing for every image seem computationally expensive.
|
passes during testing for every image seem computationally expensive.
|
||||||
In comparison a single run through a trained auto-encoder seems
|
In comparison a single run through a trained auto-encoder seems
|
||||||
@ -124,11 +111,10 @@ use the work of Miller et al. as baseline to compare against.
|
|||||||
They use the SSD~\cite{Liu2016} network for object detection,
|
They use the SSD~\cite{Liu2016} network for object detection,
|
||||||
modified by added dropout layers, and the SceneNet
|
modified by added dropout layers, and the SceneNet
|
||||||
RGB-D~\cite{McCormac2017} data set using the MS COCO~\cite{Lin2014}
|
RGB-D~\cite{McCormac2017} data set using the MS COCO~\cite{Lin2014}
|
||||||
classes. Instead of dropout sampling my approach will use
|
classes. I will use a simple implementation of an auto-encoder and
|
||||||
an auto-encoder for novelty detection with all else, like
|
novelty detection to compare with the work of Miller et al.
|
||||||
using SSD for object detection and the SceneNet RGB-D data set,
|
SSD for the object detection and SceneNet RGB-D as the data
|
||||||
being equal. With respect to auto-encoders a recent implementation
|
set are used for both approaches.
|
||||||
of an adversarial auto-encoder~\cite{Pidhorskyi2018} will be used.
|
|
||||||
|
|
||||||
\paragraph{Hypothesis} Novelty detection using auto-encoders
|
\paragraph{Hypothesis} Novelty detection using auto-encoders
|
||||||
delivers similar or better object detection performance under open set
|
delivers similar or better object detection performance under open set
|
||||||
|
|||||||
Reference in New Issue
Block a user