Download Adam Arvay IGVA 1 Opening the black box of connectionist nets

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Recursive InterNetwork Architecture (RINA) wikipedia , lookup

Distributed firewall wikipedia , lookup

Zero-configuration networking wikipedia , lookup

Peering wikipedia , lookup

Computer network wikipedia , lookup

Cracking of wireless networks wikipedia , lookup

Piggybacking (Internet access) wikipedia , lookup

Network tap wikipedia , lookup

Airborne Networking wikipedia , lookup

List of wireless community networks by region wikipedia , lookup

Transcript
Opening the black box of connectionist nets: Some lessons from cognitive science
From Computer Standards & Interfaces. Impact factor 0.894
Visualization of neural network states is an important aspect of understanding the operation of the
network. One important benefit of being able to understand neural networks is due to their biological
basis. Understanding the way an artificial neural network operates can help with understanding how an
organic neural network operates.
The basic function of a neural network is to transform an input to a specific output. It achieves this
through a learning scheme. The network will learn the appropriate associations necessary to do the
transformation correctly through a series of learning trials. There are many different ways to set up an
artificial neural network along with many different learning algorithms but all neural networks have the
same goal, to transform an input into the appropriate output. The details of how the network performs
the transformation are abstract and can be described using very high dimensional math.
This paper represents an early attempt at understanding the internal workings of a neural network by
using visualization techniques. This paper was published in 1994 and the state of computer hardware at
the time required networks to remain relatively simple. At the same time, available hardware limited
the complexity of visualizations. Although this is an early paper, it gives an important historical
perspective on the development of visualization techniques used for neural networks and serves as a
useful starting point. Neural networking technology had advanced the point that off-the-shelf software
had been developed which could be used to build and train neural networks. This allowed groups to
focus more on the functionality of the network rather than understanding the operation of the network.
The black box behavior of these off the shelf networks made it possible to completely ignore the internal
activity of the network which may have contributed to the lack of visualization work being done around
this time.
This image is a representation of a hierarchy of types that
can be found in a neural network. This example shows how inputs of A and B can be represented in the
hidden layer as R1 while a set of inputs A and C can be represented in the hidden layer as R2. Then the
two representations can be represented as a single unit Rtop. In this way the 4 binary inputs are
transformed into a single output which is represented by Rtop.
The figure on the left shows a visualization of 3 different networks and the activation strengths of
connections required to produce the given output. This network is as simple as can possibly be and if
there were just a few more nodes and connections this representation would become far too cluttered.
The figure on the right shows a decision space representation of the networks. The vertices of the
square show all of the possible inputs and the line through the box shows the location of the correct
output (+1) in the network.
This figure shows a 3 dimensional version of a decision
space representation of 3 binary inputs with no hidden layer. The inputs are Bill, laughed, and Jean. The
function of this network is to check for valid sentences. With those three inputs there are only two
possible correct outputs: Bill laughed or Jean laughed. What this representation is supposed to
demonstrate is that the top plane which represents the “laughed” plane cannot distinguish between the
different varieties of sentences that contain the word laughed. No matter how you set up the vectors
you would end up with that solution surface which cannot be separated.
The main contribution of this paper is presented in this
figure. It shows a decision space diagram of two networks that learned a different way. It shows
visually how the two networks set up their respective networks. As far as network performance is
concerned, the two networks produce the same outputs so the one way to see the internal difference
between the networks is through this decision space diagram. The actual contents of the visualization is
not very clear but the key is that there is a difference that can be seen.
Neural networks are complicated mathematical structures that have become very useful tools in
understanding human cognition. The visualizations presented in this paper represent some of the first
steps to make the internal workings of the network comprehensible. The visualizations presented have
the advantage of being very simple but they suffer from failing to convey useful information. These
visualizations can be summed up as better than nothing, but plenty of room for improvement.