Download www.njfunk.com

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Neural modeling fields wikipedia , lookup

Visual search wikipedia , lookup

Neuroesthetics wikipedia , lookup

Metastability in the brain wikipedia , lookup

Convolutional neural network wikipedia , lookup

Neurocomputational speech processing wikipedia , lookup

Visual selective attention in dementia wikipedia , lookup

Attention wikipedia , lookup

Nervous system network models wikipedia , lookup

Feature detection (nervous system) wikipedia , lookup

Mathematical model wikipedia , lookup

Biological neuron model wikipedia , lookup

Agent-based model in biology wikipedia , lookup

Visual N1 wikipedia , lookup

Visual servoing wikipedia , lookup

Efficient coding hypothesis wikipedia , lookup

Transcript
A Model of Saliency-Based
Visual Attention
for Rapid Scene Analysis
Laurent Itti, Christof Koch, and Ernst Niebur
IEEE PAMI, 1998
What is Saliency?
●
Something is said to be salient if it stands
out
●
E.g. road signs should have high saliency
Introduction
●
Trying to model visual attention
●
Find locations of Focus of Attention in an
image
●
Use the idea of saliency as a basis for their
model
●
For primates focus of attention directed from:
●
Bottom-up: rapid, saliency driven, taskindependent
●
Top-down: slower, task dependent
Results of the Model
• Only considering “Bottom-up”
 task-independent
Model diagram
Model
●
Input: static images (640x480)
●
Each image at 8 different scales
(640x480, 320x240, 160x120, …)
●
Use different scales for computing “centresurround” differences (similar to assignment)
+
Fine scale
-
Course scale
Feature Maps
Intensity contrast (6 maps)
1.
●
Using “centre-surround”
●
Similar to neurons sensitive to dark centre,
bright surround, and opposite
Color (12 maps)
2.
●
Similar to intensity map, but using different
color channels
●
E.g. high response to centre red, surround
green
Feature Maps
Orientation maps (24 maps)
3.
●
Gabor filters at 0º, 45º, 90º, and 135º
●
Also at different scales
 Total of 42 feature maps are combined into
the saliency map
Saliency Map
●
Purpose: represent saliency at all locations
with a scalar quantity
●
Feature maps combined into three
“conspicuity maps”
●
●
Intensity (I)
●
Color (C)
●
Orientation (O)
Before they are combined they need to be
normalized
Normalization Operator
Example of operation
Leaky integrate-and-fire
neurons
“Inhibition of return”
Model diagram
Example of operation
• Using 2D “winnertake-all” neural
network at scale 4
• FOA shifts every 3070 ms
• Inhibition lasts 500900 ms
Results
Image
Saliency
Map
High saliency
Locations
(yellow circles)
Results
●
Tested on both synthetic and natural images
●
Typically finds objects of interest, e.g. traffic
signs, faces, flags, buildings…
●
Generally robust to noise (less to
multicoloured noise)
Uses
●
●
Real-time systems
●
Could be implemented in hardware
●
Great reduction of data volume
Video compression (Parkhurst & Niebur)
●
Compress less important parts of images
Summary
●
Basic idea:
●
Find multiple saliency measures in parallel
●
Normalize
●
Combine them to a single map
●
Use 2D integrate-and-fire layer of neurons to
determine position of FOA
●
Model appears to work accurately and
robustly (but difficult to evaluate)
●
Can be extended with other feature maps
References
●
Itti, Koch, and Niebur: “A Model of SaliencyBased Visual Attention for Rapid Scene Analysis”
IEEE PAMI Vol. 20, No. 11, November (1998)
●
Itti, Koch: “Computational Modeling of Visual
Attention”, Nature Reviews – Neuroscience Vol.
2 (2001)
●
Parkhurst, Law, Niebur: “Modeling the role of
salience in the allocation of overt visual
attention”, Vision Research 42 (2002)