Download Document

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Airborne Networking wikipedia , lookup

IEEE 1355 wikipedia , lookup

UniPro protocol stack wikipedia , lookup

Transcript
4.2 Data Input-Output Representation
1. Background ( Ref: Handbook of Neural Computation )
Saund: Key theme in AI is to discover good rep. for the problem at hand. A good rep. makes explicit
info. useful to the computation, it strips away obscuring clutter, it reduces info. to its essentials.
Output
Raw
Data
Input
Transf.
NN
Transf.
Final
Output
(1) Data Rep. is more important/critical than network topology
Goals of Data Rep. [Garbage In Garbage Out]
•
Feature Enhancement / Data Reduction for Separability
•
•
•
•
Similar (Diff) events  Similar (Diff) rep. for better interpolation
More elements for important features
Compactness without losing (or even enhancing) info. for fast learning
Preserve feature info. (clustering / metric info)
Ex. Binary coding could destroy metric
01111  10000 [Hamming metric]
Ex. Character Recognition
Raw data = 5 64bit binary vectors.
Representation ① raw data
② any code for five characters
③ shape features : horizontal and vertical spars, their ratio,
relative positions of the spars
Other extreme:
Wasserman : raw data may be more useful in cases where essential features are
unknown. NN can discover features in hidden neurons.
2. Data Preprocessing techniques
Data sets are plagued by Noise, Bias, Large variations in dynamic range . . .
(1) Normalize  Remove large dynamic variances over one or more
dimensions in data
Ex.
① Normalize gray scale image  invariant to light condition
② Normalize speech signal  invariant to absolute volume level
③ Normalize with respect to position and size - Character recognition
(2) Nomalization ① x'  x / l
Algorithms :
One Way to embed magnitude ( l ) info. :
( x1 ,, xD , l )  Normalize
② Row Norm. (2-D) : for each row, divide by the mean value
③ Column Norm. (2-D)
(3) Principal Component Analysis – Dim. Reduction
1
R
N
n
T
(
x

x
)(
x

x
)
 k
k
k 1
M
~
x   (e j  x)e j
j 1
M D
3. Case Study : Face Recognition
(profile)
Data Reduction
416 x 320

16x2
 23 (Discard High freq. components)
(= 133,120 pixels)
① Efficient tech. to extract high interest features.
② Method for data reduction with minimal info. loss
③ DCT applied to reduced vector descriptions enhances info. content, provides
invariance to small change and increased separability.
4. Actual Coding Schemes
(1) Local R O Y G B I V
(2) Distributed C2 C1 C0
1 0 0 0 0 0 0
0 0 1


1 1 1
0 0 0 0 0 0 1
In Local representation, each node may be [0 1] [-1 1] or even continuous when
more than one node can be active to indicate presence of two or more features.
(3) Coarse Distributed
Wider overlapping receptive fields
36 nodes
27 nodes
Students’ Questions from 2005:
While DCT in facial image processing helps data
reduction, does it also help face recognition ?
Since Human faces are all alike, its Fourier transform
will also be similar. Spatial features will be more
relevant to recognition.
Normalization or Coding will reduce data or help
classification. But, isn’t the process going to delay the
overall learning time ?
Coarse distributed coding will reduce the total number
of nodes. However, when a single node is represented
in overlapped fashion, isn’t the additional info. needed
such as the overlap position, etc. ?
When an NN technique is used for character or speech
recognition, how does its performance compare with
non-NN approaches ?
NN can be applied to many problems. Any application
where NN is hard to apply to ?
Is there any general measure to tell the importance of
information in feature extraction ?
If line search is used to find an optimal learning rate,
the number of steps may decrease but I am afraid the
overall processing time may increase.
Can better separation for classification result via data
representation ? Can the information content increase
via a good data rep. ?
5. Discrete Coding
(1) Simple Sum (fault tolerance, but requires many nodes for large numbers)
5 = 000011111 = 110000111 =
(2) Value Unit Encoding
1-3
4-6
7-9 10-12 13-15
1-3
4-6 7-9 10-12 13-15
2=
10 =
(3) Discrete Thermometer
x >0 x>3 x>6 x>9 x>12
2=
x >0 x>3 x>6 x>9 x>12
10 =
6. Continuous Coding
(1) Simple Analog
• For an activation ai range of [0,1] or [-1,1],
value in range [u, v] = (v - u) ai + u
• Logarithmic scale for data set with large dynamic range
(2) Continuous Thermometer
(3) Proportional Coarse Coding
- Pomerleau used a Gaussian Smearing function to represent steering directions in an
Autonomous Land Vehicle In a Neural Network (ALVINN)
Ref. “Neural Network Perception
For Mobile Robot Guidance,”
D. Pomerleau, Kluwer, 93.
Slight Right Turn