Efficient neural models for visual attention Sylvain Chevallier, Nicolas Cuperlier and Philippe Gaussier ETIS - Neurocybernetic team Univ. Cergy-Pontoise – ENSEA – CNRS Cergy, France [email protected]

September, 22th. 2010

Framework

Outline

1

Framework Visual attention Neural models

2

Models and implementation Attentional architecture Implementations

3

Experimental results

S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

2 / 20

Framework

Visual attention

Change blindness

S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

3 / 20

Framework

Visual attention

Change blindness

S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

3 / 20

Framework

Visual attention

Change blindness

S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

3 / 20

Framework

Visual attention

Bio-inspired attentional vision systems Attentional spotlight metaphor Reduce the search space [Tsotsos, 90] Attentional architecture Feature extraction Combination on saliency map Focus selection through Winner-Take-All

[Itti & Koch, 98]

Applications Driver assistance [Michalke, 08] Retinal prostheses [Parikh, 10] Robotics [Frintrop, 06] S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

4 / 20

Framework

Neural models

Bio-inspired information coding Neurons exchange information through spikes

Spikes have little variations in amplitude and duration Spikes are fully characterized by their emission dates Level of description for neural models: Neuron level Temporal coding, precise spike timing Population level Rate coding, mean firing rate S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

5 / 20

Framework

Neural models

Neural models

Spiking Neuron Network Network of [1, . . . , i, . . . , N] spiking neurons: ( P P (s) dVi j∈Pre wij s∈Trainj δ(t − tj ) + I(t), if Vi < ϑ dt = −λi Vi (t) + else trigger a spike and Vi ← Vreset Frequency-based Neural Network Continuum neural field τ

∂u (x, t) = −u(x, t) + ∂t

S. Chevallier (ETIS)

Z

w(x − x0 )f [u(x0 , t)]dx0 + I(x, t) + h

Efficient neural models

September, 22th. 2010

6 / 20

Framework

Neural models

Goal of this paper

Question What is the most suited neural coding scheme for an efficient bio-inspired attentional architecture ?

Compare SNN and FNN Complexity analysis Quality of results Simple artificial images Natural images

S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

7 / 20

Models and implementation

Outline

1

Framework Visual attention Neural models

2

Models and implementation Attentional architecture Implementations

3

Experimental results

S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

8 / 20

Models and implementation

Attentional architecture

Preattentive visual architecture IOR

Input image

Low spatial frequencies

WTA

Saliency Input maps

High spatial frequencies

Multi-scale Features

FNN needs WTA to sort saliencies

Contrast of luminance (DOG) Orientations (Gabor) Color opponencies (DOG) S. Chevallier (ETIS)

Efficient neural models

SNN is an anytime process September, 22th. 2010

9 / 20

Models and implementation

Implementations

SNN implementation

DOG filter

details

Neural filter S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

10 / 20

Models and implementation

Implementations

Complexity analysis FNN Filtering cost: for f features, s spatial scales, filters of size M and N input image pixels WTA cost: O(N) with ARGMAX O(f × s × M × N) SNN Hybrid synchronous simulator, with time step ∆t Total cost = Spike propagation cost + neuron update cost cp × F × M × N + cu ×

A ∆t

F is mean firing rate, A is number of active neurons. cu is 10 FLOP. S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

11 / 20

Models and implementation

Implementations

Complexity analysis SNN computational cost depends on emitted spikes Is the number of spikes constant for processing different images ?

CPU cycles (106 )

2.5

1 patch 10 patchs 50 patchs 100 patchs

2 1.5 1 0.5 0

0

10

20 30 40 Simulated time (t)

50

60

For SNN, computational cost depends on the input image Rich images (w.r.t chosen filters) induce large number of spikes S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

12 / 20

Experimental results

Outline

1

Framework Visual attention Neural models

2

Models and implementation Attentional architecture Implementations

3

Experimental results

S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

13 / 20

Experimental results

Comparison on artificial images Pop-out artificial images

FNN Circle shows most salient region, winner of WTA (FNN) SNN Dots indicate the most salient pixels (SNN) Same salient items are found for FNN and SNN (20 images) S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

14 / 20

Experimental results

Natural images 19 webcam images of 160x120 pixels

Salient regions might not be extracted in the same order Measured computational cost (as CPU cycle): Constant for FNN SNN can find salient regions before FNN (1/4 of the images) S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

15 / 20

Experimental results

Conclusion and perspective Comparison of two neural models for an attentional system Frequency-based Neural Network: have a constant and lower computational cost, needs a WTA to sort saliencies Spiking Neuron Network: have a variable computational cost have anytime capabilities Perspective Formal analysis of spiking neuron processing Learning capability of neural network Attentional bias modulating salient regions Long term adaptation of input signal (slow variation of illumination) S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

16 / 20

Experimental results

Annex

S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

17 / 20

Experimental results

Input maps



dVi dt

= −λi Vi (t) + KLi , if Vi < ϑ else trigger a spike and Vi ← Vreset

with Li the considered pixel value Φi =

  1 λϑ ˆti = − ln 1 − i λi KLi

λi

= −



ln 1 −

back



S. Chevallier (ETIS)

1 ˆti

Efficient neural models

λi ϑ KLi



K Li ϑ

September, 22th. 2010

18 / 20

Experimental results

Input maps

S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

18 / 20

Experimental results

Integration maps



PPj = −λj Vj (t) + i=1 wij Si (t), if Vj < ϑ else trigger a spike and Vj ← Vreset dVj dt

Si (t) =

Ni X

δ(t − tif )

f =1

back

S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

19 / 20

Experimental results

Integration maps



PPj = −λj Vj (t) + i=1 wij Si (t), if Vj < ϑ else trigger a spike and Vj ← Vreset dVj dt

Vj (t) =

Pj X

wij

Ni X

e−λj (t−fˆti ) H(t, f ˆti )

f =1

i=1

Vj (Tj ) ≈

Pj X i=1

wij

1 − e−QNi /Li 1 − e−Q/Li

with Q =

λj ϑ K

back

S. Chevallier (ETIS)

Efficient neural models

September, 22th. 2010

19 / 20

Experimental results

Frequency coding

P1 P2 P3 P4

V

ϑ S t ISI

S. Chevallier (ETIS)

6 ms

4 ms

5 ms

Efficient neural models

5 ms

4 ms

September, 22th. 2010

20 / 20

Efficient neural models for visual attention

Reduce the search space [Tsotsos, 90]. Attentional architecture. Feature extraction. Combination on saliency map. Focus selection through. Winner-Take-All.

843KB Sizes 4 Downloads 284 Views

Recommend Documents

Efficient Neural Models for Visual Attention
process and thus reduce the complexity and the processing time of visual task. Artificial ..... the mean execution time is 0.62 sec, that is 1.6 frame per second.

Selective Visual Attention to Emotion
Jan 31, 2007 - signature of explicitly directed attention toward visual features, objects, and higher-order .... A 30 Hz digital low-pass filter was applied off-line to ...

Attention-Based Convolutional Neural Network for ...
the document are irrelevant for a given question. .... Feature maps for phrase representations pi and the max pooling steps that create sentence representations.

Efficient Training Strategies for Deep Neural Network ... -
many examples using the back-propagation algorithm and data selection ? Is care- ful initialization important ? How to speed-up training ? Can we benefit from ...

A Regularized Line Search Tunneling for Efficient Neural Network ...
Efficient Neural Network Learning. Dae-Won Lee, Hyung-Jun Choi, and Jaewook Lee. Department of Industrial Engineering,. Pohang University of Science and ...

Efficient Training Strategies for Deep Neural Network ... -
many examples using the back-propagation algorithm and data selection ? Is care- ful initialization important ? How to speed-up training ? Can we benefit from ...

Fast and Accurate Recurrent Neural Network Acoustic Models for ...
Jul 24, 2015 - the input signal, we first stack frames so that the networks sees multiple (e.g. 8) ..... guage Technology Workshop, 1994. [24] S. Fernández, A.

Models for Neural Spike Computation and ... - Research at Google
memories consistent with prior observations of accelerated time-reversed maze-running .... within traditional communications or computer science. Moreover .... the degree to which they contributed to each desired output strength of the.

Covert Attention with a Spiking Neural Network
Neural field. ▻ Spatio-temporal ... Neural field. ▻ Spatio-temporal ... Input Map. Focus Map. Error measure. ▻ Stimulus occupied a spatial position for a time ...

Covert Attention with a Spiking Neural Network
tions leading to place this work in a bio-inspired framework are explained in 1.1. The properties of the ... approaches offer a good framework for designing efficient and robust methods for extracting .... 5 have been obtained with an desktop Intel.

Rectifier Nonlinearities Improve Neural Network Acoustic Models
DNN acoustic models were initially thought to perform well because of unsupervised ... training data, 2) a comparison of rectifier variants, and. 3) a quantitative ..... Zeiler, M.D., Ranzato, M., Monga, R., Mao, M., Yang,. K., Le, Q.V., Nguyen, P., 

Efficient Indexing for Large Scale Visual Search
local feature-based image retrieval systems, “bag of visual terms” (BOV) model is ... with the same data structure, and answer a query in a few milliseconds [2].

Efficient Codebooks for Visual Concept Recognition by ...
neural network. This method aims to learn data features, without assuming any prior knowledge of their mathematical representation, such as Gabor or wavelet.

Efficient Visual Digital Signature Scheme for E- Governance ... - IJRIT
In this paper, we introduce a new digital signature scheme based on the concept ... A Visual Digital Signature scheme is a method to support visual verification of ...

Efficient Visual Digital Signature Scheme for E- Governance ... - IJRIT
Email Id: [email protected] , [email protected] ... Therefore, the sender (signer) has to depend on a computer to digitally sign a document. Also, the ...

Maximum Normalized Spacing for Efficient Visual ...
... distance met- ric learning method called Maximum Normalized Spacing (MNS) ... Data Clustering, Distance Metric Learning, Data Mining. Permission to ..... degrees apart as the object is rotated on a turntable and each object has 72 images.

An investigation of relationships among visual-attention ...
Jun 1, 2006 - focusing in time, in space and to objects, distributing attention across space to grasp global structures, shifting, multi-item tracking, and ...

Modeling Drivers' Visual Attention Allocation While ... - Semantic Scholar
are growing concerns over new in-vehicle technologies (IVTs), telematics, and “infotainment” ... vehicle status information, as well as many other wireless web or.

Visual stability based on remapping of attention pointers
Object files: A temporary representation combining an object's identity, ..... target's feature data are to be found in earlier visual cortices which are shown, highly.

Do Synesthetic Colors Grab Attention in Visual Search.pdf ...
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Do Synesthetic ...