In Proc. of IEEE Workshop on Neural Networks for Signal Processing, Sept. 1996

A Multi-Module Minimization Neural Network for Motion-Based Scene Segmentation Yen-Kuang Chen and S.Y. Kung Princeton University

Abstract–A competitive learning network, called Multi-Module Minimization (MMM) Neural Network, is proposed for unsupervised classification. Our objective is to provide a general framework to divide a set of input patterns into a number of clusters such that the patterns of the same cluster exhibit any pre-specified similarity measure (i.e. not limited only to RBF). As an example of non-RBF measure, let us look into a motion-based segmentation problem. The image frame can be divided into different regions (segments) each of which is characterized by a consistent affine motion. Algebraically, this leads to an LBF similarity criterion–because each region can be characterized by a 3-dimensional hyperplane. In order to apply the traditional RBF clustering techniques (e.g. VQ, k-mean), it requires a preprocessing step such as taking the Hough transform, which itself creates additional ambiguity. This problem is avoided in a direct approach such as the proposed MMM neural network. It allows us to directly cluster the tracked features into different moving objects by means of an LBF cost function. In general, the primary cost function should be carefully chosen to reflect the true physical model of the application. By minimizing the cost function, we can categorize a set of input patterns into a number of clusters. Because the primary similarity measure is no longer Euclidean type, it may become necessary to take spatial neighborhood into account as a secondary cost function. Still, a third cost function, reflecting the MDL type criterion, needs to be added so that noisy or spurious patterns will not be mistakenly modeled as a meaningful class. Accordingly, we have proposed an EM-type learning algorithm which uses all or part of the three cost functions mentioned above. A convergence proof for this algorithm is provided. Simulation results demonstrate that the MMM neural network does capture different motions and yield fairly accurate segmentation and motion-compensated frames.

1

Introduction

Various kinds of unsupervised clustering methods, including competitive learning networks, have been developed [5, 6, 7]. In a general unsupervised clustering problem, a set of input patterns X = fX1 : : :  XN g are grouped (according to some optimization criterion) into M clusters, parameterized by W = fW1  : : :  371

In Proc. of IEEE Workshop on Neural Networks for Signal Processing, Sept. 1996

WM g.

Mathematically, given the input minimizes

E=

N X

X , the problem is to find W 

error

j =1



which

(Xj W ) 

j

where  is an (unknown) mapping function: 

: ! j

j

where j 2 f1 : : :  N g, j 2 f1 : : :  M g, and error(Xj  Wj ) is the similarity measure chosen for the specific application. The most famous approaches are kmean (MacQueen) and Vector Quantization (Linde, Buzo, and Gray) (see e.g. [5, 7]) where the similarity measure is simply the Euclidean distance as error

(Xj Wi ) = jjXj ; Wi jj2 

For this radial-basis function (RBF) type clustering, the patterns are grouped into a same cluster if they are close enough. (This leads to a simple rule for the determination of fj g.) The optimal W which delivers the minimal error can be easily obtained by the arithmetic average of the input patterns of the same cluster. In this paper, we provide a more general framework (i.e. not limited only to RBF) for clustering feature patterns. Three cost functions are considered: 1. As a primary cost function, it is necessary to define a pre-specified similarity, error (Xj  Wi ), according to the physical constraints of the application. 2. Although the primary similarity measure is no longer Euclidean, it is sometimes necessary to take into spatial neighborhood into account as an auxiliary criterion. This necessitates the additional cost component concerning neighborhood sensitivity. The same concept may be extended to the “distance” in other features such as color, texture, etc. 3. The number of clusters may have to be limited, so that noisy or spurious patterns will not be mistakenly modeled as a meaningful class. For this purpose, a Minimal Description Length (MDL) type criterion proves generally effective.

2

Multi-Module Minimization Neural Network

The goal of this network is to find a set of weighting and  j which designates that Xj belongs to cluster j . In short, the optimization procedure is

min

N X

W j =1

error

372

(Xj W ) 

j

(1)

In Proc. of IEEE Workshop on Neural Networks for Signal Processing, Sept. 1996

The above can be decomposed into two strategies:

 min   min W 2.1

N X

error

j =1 N X

2 N X 4 (Xj W ) 8W , then min min  W j =1 2 N X 4 (Xj W ) 8 , then min min  W

error

j =1



j



j



3 (Xj W )5 3 (Xj W )5

error

error

j =1





j

j

(2)

(3)

Basic Learning Rule

Either of the above formulations would imply exhaustive searches on the product space fWg  fg, which is computationally costly. So, we adopt an EM-type iterative updating procedure consisting of two basic steps: Step (1) Given W , find the optimal mapping  for clustering the input patterns: In this step, we concentrate on minimizing the inner term of Equation (2):

min 

N X

error

j =1

where W is temporarily fixed.

(Xj W ), 

j

Step (2) Given , find the optimal weighting W : Again, we concentrate on minimizing the inner term of Equation (3):

min W

N X

error

j =1

(Xj W ), 

where  is temporarily fixed.

j

Converting these two steps into iterations, with (t) denoting the iteration number, we have: 

(t+1)

=

arg

W (t+1) =

arg

min  min W

N X

error

(Xj W(t))

error

(Xj W

j =1 N

X j =1





(4)

j

(t+1)

j

)

(5)

The iterations will be repeated until convergence is reached. Proof of Convergence for Fixed Size MMM Now, we can show that the total error will non-monotonically decrease in every iteration.

E(t) 

N X j =1

error

(Xj W(t) ) 

(t)

j

373

In Proc. of IEEE Workshop on Neural Networks for Signal Processing, Sept. 1996

Decision

σ : Winner-Net W1

W2

WM Weighting Adjustment

Xj Figure 1: Network structure for multi-module minimization. Every module targets at a cluster of input patterns. For every input pattern, every module will reckon a score. The winner-net will decide which module possesses the input pattern. Then, the module will accommodate its weighting for the input pattern.

 

N X

error

j =1

(by Equation (4))

(Xj W(t+1) ) = E(t+1)

(by Equation (5))



(t+1)

j

j =1 N

X

(Xj W(t) )

error



(t+1)

j

Note that the proof is valid for any similarity measure (or error function). The total error always decreases (non-monotonically) in every iteration of weight updating. If there is no trap (local minimum) between the initial weighting and the global minimum, then it will converge to the global minimum. To alleviate the adverse effect due to traps, multiple initial conditions should be tried and only the best converging solution will be selected. The basic structure, as shown in Figure 1, has one global winner-net and a number of small modules. An input pattern Xj is a sample point in the n-dimensional real or binary vector space. There are as many modules as the number of classes and each one represents a pattern category. The winner network selects the winner. (Here, the winner is the module which has the least error.) The competitive learning rule adopts the “winner-take-all” schema, i.e. a unit learns if and only if it wins the competition among all the other units.

2.2

Neighborhood/Color/Texture Correlation

P

So far, only the primary cost function is considered, i.e. error (Xj  Wj ). However, in many applications, the input patterns from the same class (or same object) exhibit high correlation in some other secondary factors, e.g. spatial, color, texture “distances.” To exploit information pertaining to such correlation factors, we 374

In Proc. of IEEE Workshop on Neural Networks for Signal Processing, Sept. 1996

incorporate them into the following cost function:

E = W min

0 N X @

error

j =1

(Xj W ) + 

j

X 8k6=j

r

(Xk Xj ) 

error

1 (Xk W )A 

j

(6)

where correlation factor r(Xk  Xj ) is a combined function of spatial-closeness and color/texture similarities, etc. The smaller the “distance” between input patterns Xk  Xj , the higher the correlation factor r(Xk  Xj ). The second term in Equation (6) is to help regularize the classification process [6]. Note that this extra term will increase if r(Xk  Xj ) is large (i.e. Xk  Xj are highly correlated) while they are being clustered into different modules. As a result, Xk  Xj are more likely to be put into a same class for the interest of reducing the total error function. Otherwise, a heavy penalty will have to be paid.

2.3

Minimal Description Length

So far, for simplicity, we have assumed that the total number of modules M is pre-specified. In practice, we would not be able to know M in advance. Note that if a large number of modules are adopted, then E can be substantially reduced. However, it also leads to an undesirable consequence that a single true class is represented by multiple modules. To alleviate such a problem, it is necessary to introduce one more optimization factor: i.e. the value of M . For example, Ayer and Sawhney [1] show that the adequate number of modules is automatically decided using the MDL principle that minimizes the encoding length of the model parameters and the residuals. Following the same MDL idea, we propose to introduce one more cost term, M increases. The optimization criterion now has the following form:

( ), which would increase when

C M

EMDL = E + ( ) C M

The optimal number of modules is decided by minimizing this cost function. In the event that there remain a small number of input patterns not covered by the existing modules (due to noises, say), then the newly introduced cost function could force we to simply ignore them (as noise), since it may become more costly by spending extra modules to model (cover) them.

3

Application to Motion-Based Scene Segmentation

In this section, we demonstrate how the multi-module minimization is applied to motion-based scene segmentation, which has many potential video applications, e.g. object-based video coding, scene analysis, and object recognition.

375

In Proc. of IEEE Workshop on Neural Networks for Signal Processing, Sept. 1996

There are two major (sometimes coexisting) motion factors in the real-life scene: motion of camera and (different) motions of objects. Camera motions fall into several categories: translation, zoom, and pan. On the other hand, the motions of objects are often modeled by 2D affine or 3D rigid-body motion. To cope with all these, the (2D or 3D) affine motion model (covering translation, rotation, and camera zooming) is often adopted.1 The following 2-D affine transformation is common in projective geometry:



0 0 y

x

 =

a11

a12

a21

a22

  +

x y

b1



b2

(7)

where (x y) and (x0  y0 ) denote the coordinates of a feature point (FP) in two consecutive frames. FPs from the same object (segment) should share a common affine motion (i.e. same parameters). Given many FPs in a same object (e.g. (xi yi ) & (x0i  yi0 ) for i = 1 : : :  K ), the x0i = a11xi + a12yi + b1 is a plane in 3-dimensional space. And, such parameters can be easily estimated by solving a least-square solution based on Equation (7).

3.1

Motion-Based Segmentation by MMM

FPs from different moving objects tend to form separate hyperplanes on the 3dimensional (x0  x y) space according to their own motion. So, FPs tracked by the tracker [3] can be classified into clusters such that the blocks in the same cluster share a common motion, i.e. they can be described by the same motion parameters. Hence, we can use choose a linear-basis function (LBF) as the error function, as shown here: error

(Xj Wi) = jjXTj Wi jj 

then our MMM NN is splitting the input patterns into clusters where the input patterns are close to a common hyperplane in the feature space. That is, the LBF MMM NN is used to cluster the tracked features into different moving objects. Figure 2 depicts the procedure of our motion-based segmentation scheme. As depicted in the flow chart, MMM NN plays the central role for motion classification. Most existing motion-based segmentation methods are 2-frame based approaches. We propose a novel technique, which can effectively integrate motions from multiple frames (more than 2 frames) so as to extract richer information. Because (y0  x y) will also form several 3-dimensional hyperplanes for different moving objects, and if we have multiple tracked frames, then (x00 x y) or (y 00  x y) can do too, we can fuse the results of different MMM NN to cluster the tracked features in a robust and efficient way. For example, we use

XTj = 

xj

yj

0

xj

0

yj

00

xj

00 1 ]

yj

1 Note that 3D affine model can cover the so-called para-perspective model [9] which is a linear approximation of the true perspective model. Sometimes, the perspective distortion is not negligible and yet a more general model has to be considered.

376

In Proc. of IEEE Workshop on Neural Networks for Signal Processing, Sept. 1996

and

2 66 66 Wi = 66 66 4

a11i

a21i

a12i

a22i

b1i

b2 i

;1

0 0 a12i a11i

0 0 0 ;1 0 0 0 ;1 0 0 0 0

b1i

3 77 0 77 0 77 0 77 ;1 5

0 0 a22i a21i

0

b2i

The image frame can thus be accordingly divided into different regions (segments) each of which is characterized by a consistent motion. Our simulation results support that our MMM NN does capture different motion objects, since it successfully leads to fairly accurate motion-based segmentation and motion-compensated frames. The results are shown in Figure 2. The “Hinging Hyperplanes” approach proposed by Breiman [2] is very similar to the MMM NN with LBF similarity measure, except for the limitation that Breiman’s procedure cannot be applied to situations with intersecting hyperplanes. Our procedure can.2

3.2

Comparing with Hough-Transform-then-VQ Approach

In order to apply the traditional RBF clustering techniques (e.g. VQ, k-mean) for this motion-based segmentation problem, it requires a Hough transform preprocessing step [4]. The affine parameters for every triplet (or quadruplet, or even more) of the tracked feature points are calculated and mapped onto a point in a new Euclidean space, so VQ can be applied. (The affine parameters for any triplet in the same hyperplane should be very close to each other in Euclidean distance.) Because the affine parameters are calculated for randomly chosen triplets, nosies may be created when the components of the triplet are selected from different moving objects. As a result, when the unsupervised classification VQ is performed, the signal-tonoise ratio (SNR) is lower than the SNR when we perform the LBF MMM directly. For example, if X1 X2 X3 should belong to one group and X 4  X5 X6 to another group, then it is clear that X1  X2 X5 should not be classified into any group. The point (in the new Euclidean space) is regarded as a noisy pattern, creating difficulty for VQ. This approach suffers from another problem in memory/quantization. Because we must calculate the affine parameters for any triplet, assuming there are n input points, there will be a total of C 3n (O(n3)) points in the transformed space, which is much greater than original n points. Extra memory is required, or, in order to reduce the memory requirement, some information has to be truncated, causing degraded accuracy. 2 However, under the more restricted condition, Breiman’s convergence proof does show it reaches a global minimum.

377

In Proc. of IEEE Workshop on Neural Networks for Signal Processing, Sept. 1996

One possible approach to solve these problems is to use a neighborhood sensitivity to pre-select possible triplets of points from small local regions for Hough transform [11]. So, the memory requirement is greatly reduced and so is the possibility of cross object error. Another approach is by a more direct approach such as the proposed MMM NN, where no extra memory is required. The partitioning scheme also avoids altogether the above-mentioned noisy patterns. From these points of view, the LBF MMM is preferable to the indirect approaches.

4

Extension to a Probabilistic MMM Formulation If we define

(Xj )  =

p

i

( (

 arg

min ( k )

error

(Xj Wk )) ) 

i

(8)

 j  i

where  (x y) is a unit impulse function, i.e. the Kronecker delta function, and j is defined in an obvious manner. Now, Equation (1) can be rewritten as minimizing the following equation:

E =

N X M X

error

j =1 i=1

(Xj Wi) (Xj ) 

p

(9)

i

which may be further linked to an energy function proposed by Durbin and Willshaw (see e.g. [7]):

E=;

T

X N M X j =1

log

(

exp

i=1

error

(Xj Wi ) ) ; 

!

T

(10)

where the temperature T moderates the influence of error distance. (For example, when T ! 0, Equation (10) becomes the same as Equation (9).) If T is large, only a huge error can make a significantly difference in the energy function. If T is small, any tiny error can make a big difference. The advantage of Equation (10) is that it is differentiable. Therefore, the network can be trained by gradient-decent whenever any input pattern is presented. So far, the p(Xj  i) is restricted to binary value, but it does not have to be so restrictive. It may be generalized into a probability of X j belonging to cluster i: for example,

(Xj ) = PM (;

p

i

exp

(Xj Wi ) ) (; (Xj Wi) )

error

i=1 exp

error



=T



(11)

=T

(Again, in the limiting case, as T ! 0, Equation (11) become Equation (8). p(Xj  i) error (Xj  Wi ) is the minimum, and p(Xj  i) = 0, otherwise.) Note that,

= 1 if

378

In Proc. of IEEE Workshop on Neural Networks for Signal Processing, Sept. 1996 in Equation (11), the smaller the error(Xj  Wi ), the greater the chance that Xj belongs to cluster i, and so the larger the probability p(X j  i). This probabilistic version provides some fuzzy decision capabilities, useful for dealing with ambiguous patterns. According to our own observation, however, the convergence of this probabilistic version is somewhat slower than that of binary partitioning. Their performance comparison will be reported in further publication.

References [1] S. Ayer and H. S. Sawhney, “Layered Representation of Motion Video using Robust Maximum-Likelihood Estimation of Mixture Models and MDL Encoding,” IEEE International Conference on Computer Vision, June 1995, pp. 777-784. [2] L. Breiman, “Hinging Hyperplanes for Regression, Classification, and Function Approximation,” IEEE Transactions on Information Theory, Vol. 39, No. 3, May 1993, pp. 999-1013. [3] Y.-K. Chen, Y.-T. Lin, and S. Y. Kung, “A Feature Tracking Algorithm Using Neighborhood Relaxation with Multi-Candidate Pre-Screening,” in IEEE International Conference on Image Processing, Lausanne, Switzerland, Sept. 1996. [4] N. Guil, J. Villalba, and E. L. Zapata, “A Fast Hough Transform for Segment Detection,” IEEE Transactions on Image Processing, Vol. 4, No. 11, Nov. 1995, pp. 1541-1548. [5] S. Haykin, Neural Networks, Macmillan College: NY, 1994. [6] V. Kumar and E. S. Manolakos, “Recurrent Architectures for Object Recognition by Parameter Estimation of Hierarchical Mixtures,” in 30th Conference on Information Sciences and Systems, Princeton, NJ, March 1996. [7] S. Y. Kung, Digital Neural Networks, Prentice Hall: NJ, 1993. [8] S. Y. Kung, Y.-T. Lin, and Y.-K. Chen, “Motion-Based Segmentation by Principal Singular Vector (PSV) Clustering Method,” Proceedings of ICASSP ’96, Atlanta, May 1996, pp. 3410-3413. [9] C. J. Poelman and T. Kanade, “A Paraperspective Factorization Method for Shape and Motion Recovery,” Technical Report CMU-CS-92-208, CMU, 1992. [10] C. Tomasi and T. Kanade, “Shape and Motion from Image Streams: a Factorization Method–Part 3, Detection and Tracking of Point Features,” Technical Report CMU-CS-91-132, CMU, Apr. 1991. [11] J. Y.-A. Wang and E. H. Adelson, “Representing Moving Images with Layers,” IEEE Transactions on Image Processing, Vol. 3, No. 5, Sept. 1994, pp. 625638. [12] Y. Weiss and E. H. Adelson, “Motion estimation and segmentation using a recurrent mixture of experts architecture,” IEEE Workshop on Neural Network for Signal Processing, Aug. 1995, pp. 293-302.

379

In Proc. of IEEE Workshop on Neural Networks for Signal Processing, Sept. 1996

Neighborhood Relaxation BMA Tracking

MMM Clustering for Initial Segmentation

Affine Motion Estimation, Compensation, and Secene Segmentation

Figure 2: A flow chart of a motion-based segmentation by multi-module minimization (MMM) clustering method. First, several frames of a video sequence are input to the feature tracker. After the tracking of moving features is obtained, MMM NN is used to separate the tracked features into several clusters. Finally, affine motion estimation and affine motion compensation test for each clusters are applied so that the scene is segmented and affine reconstructed image is obtained.

380

A Multi-Module Minimization Neural Network for Motion ...

Abstract–A competitive learning network, called Multi-Module Mini- mization (MMM) Neural ... not be mistakenly modeled as a meaningful class. Accordingly, we.

226KB Sizes 2 Downloads 332 Views

Recommend Documents

Neural Network Toolbox
3 Apple Hill Drive. Natick, MA 01760-2098 ...... Joan Pilgram for her business help, general support, and good cheer. Teri Beale for running the show .... translation of spoken language, customer payment processing systems. Transportation.

LONG SHORT TERM MEMORY NEURAL NETWORK FOR ...
a variant of recurrent networks, namely Long Short Term ... Index Terms— Long-short term memory, LSTM, gesture typing, keyboard. 1. ..... services. ACM, 2012, pp. 251–260. [20] Bryan Klimt and Yiming Yang, “Introducing the enron corpus,” .

A programmable neural network hierarchical ...
PNN realizes a neural sub-system fully controllable (pro- grammable) behavior ...... comings of strong modularity, but it also affords flex- ible and plausible ...

A Review on Neural Network for Offline Signature Recognition ... - IJRIT
Based on Fusion of Grid and Global Features Using Neural Networks. ... original signatures using the identity and four Gabor transforms, the second step is to ...

A Deep Convolutional Neural Network for Anomalous Online Forum ...
as releasing of NSA hacking tools [1], card cloning services [24] and online ... We propose a methodology that employs a neural network to learn deep features.

Development and Optimizing of a Neural Network for Offline Signature ...
Computer detection of forgeries may be divided into two classes, the on-line ... The signature recognition has been done by using optimum neural network ...

Neural Network Toolbox
[email protected] .... Simulation With Concurrent Inputs in a Dynamic Network . ... iii. Incremental Training (of Adaptive and Other Networks) . . . . 2-20.

A Regenerating Spiking Neural Network
allow the design of hardware and software devices capable of re-growing damaged ..... It is interesting to analyse how often a single mutilation can affect the ...

A Regularized Line Search Tunneling for Efficient Neural Network ...
Efficient Neural Network Learning. Dae-Won Lee, Hyung-Jun Choi, and Jaewook Lee. Department of Industrial Engineering,. Pohang University of Science and ...

A Novel Three-Phase Algorithm for RBF Neural Network Center ...
Network Center Selection. Dae-Won Lee and Jaewook Lee. Department of Industrial Engineering,. Pohang University of Science and Technology,. Pohang ...

Neural Network Toolbox
to the government's use and disclosure of the Program and Documentation, and ...... tool for industry, education and research, a tool that will help users find what .... Once there, you can download the TRANSPARENCY MASTERS with a click.

A Neural Network for Global Second Level Trigger
•Calorimeter parameters (5 values). •TRT parameters (2 values). •Preshower parameters (3 values). •Muon detector (1 value). From these 12 parameters one ...

A Cyclostationary Neural Network Model for the ...
fects the health of the community and directly influences the sustainability ... in order to enable the development of tools for the management and reduction of pollution ..... [10] G. Masters, Introduction to Environmental Engineering and Science.

Validation of a constraint satisfaction neural network for ...
In addition, the effect of missing data was evaluated in more detail. Medical Imaging .... Receiver Operating Characteristics (ROC) analysis. We used the ROCKIT ...

Convolutional Neural Network Committees For Handwritten Character ...
Abstract—In 2010, after many years of stagnation, the ... 3D objects, natural images and traffic signs [2]–[4], image denoising .... #Classes. MNIST digits. 60000. 10000. 10. NIST SD 19 digits&letters ..... sull'Intelligenza Artificiale (IDSIA),

A Review on Neural Network for Offline Signature ...
This paper represents a brief review on various approaches used in signature verification systems. Keywords: Signature, Biometric, Artificial Neural Networks, Off-line Signature Recognition and Verification. I. INTRODUCTION. Biometrics are technologi

A Simple Feedforward Neural Network for the PM10 ... - Springer Link
Dec 23, 2008 - A Simple Feedforward Neural Network for the PM10. Forecasting: Comparison with a Radial Basis Function. Network and a Multivariate Linear ...

Development and Optimizing of a Neural Network for Offline ... - IJRIT
IJRIT International Journal of Research in Information Technology, Volume 1, Issue ... hidden neurons layers and 5 neurons in output layers gives best results as.

Neural Network Toolbox - Share ITS
are used, in this supervised learning, to train a network. Batch training of a network proceeds by making weight and bias changes based on an entire set (batch) of input vectors. Incremental training changes the weights and biases of a network as nee

Strategies for Training Robust Neural Network Based ...
the proposed solutions. ... in modern machine learning theory is to find solutions to improve the .... and Matwin [6] propose the one-sided selection (OSS), where.

Covert Attention with a Spiking Neural Network
Neural field. ▻ Spatio-temporal ... Neural field. ▻ Spatio-temporal ... Input Map. Focus Map. Error measure. ▻ Stimulus occupied a spatial position for a time ...

offline handwritten word recognition using a hybrid neural network and ...
network (NN) and Hidden Markov models (HMM) for solving handwritten word recognition problem. The pre- processing involves generating a segmentation ...

Self-organization of a neural network with ...
Mar 13, 2009 - potential Vi a fast variable compared to the slow recovery variable Wi. i is the .... 0.9gmax blue line, and the others black line. b and c The average in-degree and ... active cells a powerful drive to the inactive ones. Hence,.