Download Quality – An Inherent Aspect of Agile Software Development

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Existential risk from artificial general intelligence wikipedia , lookup

Mathematical model wikipedia , lookup

Visual Turing Test wikipedia , lookup

Philosophy of artificial intelligence wikipedia , lookup

Perceptual control theory wikipedia , lookup

Neural modeling fields wikipedia , lookup

Machine learning wikipedia , lookup

Concept learning wikipedia , lookup

History of artificial intelligence wikipedia , lookup

Pattern recognition wikipedia , lookup

Catastrophic interference wikipedia , lookup

Convolutional neural network wikipedia , lookup

Hierarchical temporal memory wikipedia , lookup

Transcript
Hierarchical Temporal Memory
“The Classification of Un-preprocessed
Waveforms through the Application of the
Hierarchical Temporal Memory Model”
April 18, 2009
Version 1.0; 04/18/2009
John M. Casarella
Ivan G. Seidenberg School of CSIS, Pace University
Topics to be Discussed

Introduction

Intelligence

Artificial Intelligence

Neuroscience

Connectionism and Classical Neural Nets

Pattern Recognition, Feature Extraction & Signal Processing

Hierarchical Temporal Memory (Memory – Prediction)

Hypothesis

Research

Results
Introduction
"I was proceeding down the road. The trees
on the right were passing me in orderly
fashion at 60 miles per hour. Suddenly one
of them stepped in my path."
John von Neumann providing an
explanation for his automobile accident.
Intelligence

What is Intelligence?





A uniquely human quality?
“means the ability to solve hard problems”
Ability to create memories
Learning, language development, memory
formation (synaptic pattern creation)
The human ability to adapt to a changing
environment or the ability to change our
environment for survival
Alan Turing






“Can machines Think?”
Connectionism
Model digital computer like child’s mind, then
“educate” to obtain “adult”
“Unorganized Machines” : A network of neuronlike Boolean elements randomly connected
together
Proposed machines should be able to ‘learn by
experience’
The Turing Test - constrained and focused
research


Imitate human behavior
Evaluate AI only on the basis of behavioral response
Turing’s unorganized machine
Machine Intelligence

What is Artificial Intelligence?


The Objectives of AI




Create machines to do something which would require
intelligence if done by a human
To solve the problem of how to solve the problem
von Neumann and Shannon


The science and engineering of making intelligent machines,
especially intelligent computer program
Sequential processing vs. parallel
McCarthy (Helped define AI), Minsky (first dedicated AI lab
at MIT) and Zadeh (Fuzzy Logic)
Various Varieties



Expert Systems (rule based, fuzzy, frames)
Genetic Algorithms
Perceptrons (Classical Neural Networks)
Neural Nets

McCulloch and Pitts







Model of Neurons of the brain
Proposed a model of artificial Neurons
Cornerstone of neural computing and neural networks
Boolean nets of simple two-state ‘neurons’
Concept of ‘threshold’
No mechanism for learning
Hebb - Pattern recognition learned through by
changing the strength of the connection between
neurons
Classical Neural Networks

Rosenblatt
Perceptron Model - permitted mathematical analysis of
neural networks
 Based on McCulloch and Pitts
 Linear combiner followed by a hard limiter
 Activation and weight training
 Linear Separation - No XOR

Classical Neural Networks

Minsky and Papert, what where they thinking
Mathematical proof perceptron model of limited usefulness
 Classes of problems which perceptrons could not handle
 Negative impact on funding
 Narrow analysis of the model
 Incapable of learning the XOR - wrong
 Incorrectly postulating that multi-layer perceptrons would
be incapable of the XOR

Classical Neural Networks


The Quiet Years, Grossberg, Kohonen and
Anderson
Hopfield



Rumelhart and McClelland



Introduced non-linearities
ANNs could solve constrained optimization problems
Parallel Distributed Processing
Backpropagation
Interdisciplinary nature of neural net
research
Neuroscience



Structure of the neocortex
Learning, pattern recognition and synapses
Mountcastle

Columnar model of the Neocortex
Learning associated with construction of cell
assemblies related to the formation of pattern
associations


Neuroplasticity
Neuroscience

The Biology





Neocortex >50% of human brain
Locus of: perception, language, planned behavior,
declarative memory, imagination, planning
Extremely flexible/generic
Repetitive structure
The Neocortex




Hierarchy of cortical regions
Region - region connectivity
Cortical - thalamic connectivity
Cortical layers: cell types and connectivity
Neuroscience






Layer 1
Layer 2
Layer 3

A

B
Layer 4
Layer 5

A

B
Layer 6

A

B
Electrocardiogram






Electrocardiogram (ECG) records the
electrical activity of the heart over time
Breakthrough by Willem Einthoven in
1901
Electrodes placed as per a pattern
ECG displays the voltage between pairs
of these placed electrodes
Immediate results
Assigned letters P, Q, R, S and T to the
various deflections
Electrocardiogram
Measurement of the
flow of electrical
current as it moves
across the
conduction pathway
of the heart.
Recorded over time
Represents different
phases of the cardiac
cycle
Electrocardiogram
Hypothesis


Application of the HTM model, once
correctly designed and configured,
will provide a greater success rate in
the classification of complex
waveforms
Absent of pre-processing and
feature extraction, using a visual
process using actual images
Research

Task Description



Create an image dataset of each waveform
group for classification
Determine, through organized experiments,
an optimized HTM
Apply optimized HTM to the classification of
waveforms using images, devoid of any preprocessing or feature extraction
Hierarchical Temporal Memory
Overview


Each node performs similar
algorithm
Each node learns




“Names” of groups passed up



Hierarchy of memory nodes

1) Common spatial patterns
2) Common sequences of spatial
patterns
(use time to form groups of patterns
with a common cause)
- Many to one mapping, bottom to
top
- Stable patterns at top of hierarchy
Modeled as an extension of
Bayesian network with belief
propagation
Creates a hierarchical model (time
and space) of the world
Hierarchical Temporal Memory
Structure of an HTM
network for learning
invariant representations for
the binary images world.
 This network is organized
in 3 levels. Input is fed in at
the bottom level. Nodes are
shown as squares.
 The top level of the
network has one node, the
middle level has 16 nodes
and the bottom level has 64
nodes.
 The input image is of size
32 pixels by 32 pixels.
This image is divided into
adjoining patches of 4 pixels
by 4 pixels as shown.
Each bottom-level node’s
input corresponds to one
such 4x4 patch.

Hierarchical Temporal Memory
The fully learn ed node has 12 quantiz ation centers within its spatial po oler and 4 tempo ral
groups within its tempo ral po oler . The quantiz ation cente rs are shown as 4x4 pix el patc hes.
The tempo ral groups are shown in ter ms of their quantiz ation centers. The input to th e node
is a 4x 4 pix el patch .
Hierarchical Temporal Memory
QuickTime™ and a
decompressor
are needed to see this picture.
This figure illustrates how nodes operate in a hierarchy; we show a two-level network and its associated inputs for three
time steps. This network is constructed for illustrative purposes and is not the result of a real learning process. The outputs
of the nodes are represented using an array of rectangles. The number of rectangles in the array corresponds to the length
of the output vector. Filled rectangles represent ‘1’s and empty rectangles represent ‘0’s.
Hierarchical Temporal Memory
QuickTime™ and a
decompressor
are needed to see this picture.
This input sequence is for an “L” moving to the right. The level-2 node has already learned
one pattern before the beginning of this input sequence. The new input sequence
introduced one additional pattern to the level-2 node.
Hierarchical Temporal Memory




(A) An initial node that has not
started its learning process.
(B) The spatial pooler of the
node is in its learning phase
and has formed 2
quantization enters
(C) the spatial pooler has
finished its learning process
and is in the inference stage.
The temporal pooler is
receiving inputs and learning
the time-adjacency matrix.
(D) shows a fully learned
node where both the spatial
pooler and temporal pooler
have finished their learning
processes
Hierarchical Temporal Memory
Temporal Grouping
QuickTime™ and a
decompressor
are needed to see this picture.
Hidden Markov Model - A
Temporal Grouping
QuickTime™ and a
decompressor
are needed to see this picture.
HMM - 2
Temporal Grouping
QuickTime™ and a
decompressor
are needed to see this picture.
HMM - 3
Experimental Design

HTM Design, Parameters and Structure





Determine the number of hierarchies to be used
Image size in pixels influences the sensor layer
Pixel Image size broken down into primes to determine
layer 1 and layer 2 array configuration
Determine the number of “iterations” of viewed images
at each layer
Small learning and unknown datasets used
Hierarchical Temporal Memory





Memorization of the input patterns
Learning transition probabilities
Temporal Grouping
Degree of membership of input pattern
in temporal group
Belief Propagation
Experimental Design

Waveform Datasets





Individual beats broken down by classification and
grouped (SN, LBBB, RBBB)
Teaching and unknown dataset randomly created
Teaching sets of 50, 90 and 100 images used
Multiple sets created
Teaching vs. Unknown Datasets


Traditional ratios 1:1 or 2:1, teaching to unknown
With HTM model, the ratio was 1:3, teaching to
unknown
ECG Waveform Images - Sinus
ECG Waveform Images
Individual ECG Beat Images
Left Bundle Branch Block
Normal Sinus
Right Bundle Branch Block
All images were sized to 96 x 120 pixels
ECG Series Waveform Images
Left Bundle Branch Block
Right Bundle Branch Block
Sinus
Results – Individual Beat Classification

Learning





Smaller number of teaching images
Diverse images produce greater classification pct
Overtraining not evident, saturation may exist
RAM influences performance
Waveform Datasets




Diversity
Noise : approx. 87 pct w/o inclusion in teaching set
Average > 99 percent classification
Average differentiation of images by class approx. 99 pct
NUPIC Model Results
48 object categories producing 453 training images. Only 99.3 percent of the training
images were correctly classified. (32 x 32 pixels)
Of this “distorted set”, only 65.7 percent were correctly classified within their
categories
Individual Beat Results
Percent Classified by Model
HTM Model Series
Percent Classified
htm_100_17
98.5
htm_100_19
99.2
htm_100_21
98.8
htm_100_23
99.2
htm_100_25
99.2
htm_100_29
99.0
htm_100_33
98.8
htm_100_37
98.8
Results by Dataset
10 0
99 .8
99 .6
99 .4
99 .2
99
98 .8
98 .6
98 .4
98 .2
98
p dataset
q dataset
r dataset
s dataset
t dataset
v dataset
x dataset
Results by HTM Model
10 0
99 .8
99 .6
99 .4
99 .2
99
98 .8
98 .6
98 .4
98 .2
98
ht m_100 _17
ht m_100 _19
ht m_100 _21
ht m_100 _23
ht m_100 _25
ht m_100 _29
ht m_100 _33
ht m_100 _37
IR Spectra
Sample IR Spectra
Results – IR Spectra
Figure X: Left: IR Spectrum of an Alcohol; Right: IR Spectrum of a Hydrocarbon
Figure X: IR Spectrum of a Nitrile
Classification pct > 99
Gait Waveforms
ALS
Control
With a limited teaching and
unknown set > 98 pct
Huntingtons
References (Short List)







[4] Computational Models of the Neocortex. http://www.cs.brown.edu/people/tld/projects/cortex/
[6] Department of Computer Science, Colorado State University.
http://www.cs.colostate.edu/eeg/?Summary.
[8] George, Dileep and Jaros, Bobby. The HTM Learning Algoritims. Numenta, Inc. March 1, 2007.
[11] Hawkins, Jeff and Dileep, George. Hierarchical Temporal Memory, Concepts, Theory, and Terminology.
Numenta, Inc. 2006.
[12] Hawkins, Jeff. Learn like a Human. http://spectrum.ieee.org/apr07/4982.
[15] Swartz Center for Computational Neuroscience. University of California San Diego.
http://sccn.ucsd.edu/eeglab/downloadtoolbox.html
[16] Turing, A. M. “Computing Machinery and Intelligence”. Mind, New Series, Vol. 59, No. 236. (Oct.,
1950). pp 443 – 460.