Optimal Disturbance Rejection Using Hopfield Neural Networks Mohammad Reza Rajati, Hamid Khaloozadeh, Mohammad Mehdi Korjani, Student Member, IEEE

Abstract— This paper investigates the problem of disturbance rejection and its solution by means of Hopfield neural networks. The dynamic optimization problem is transformed into a static optimization problem via linear state space analysis methods. The parameters of the Hopfield neural network are adjusted such that the network solves the static quadratic optimization problem yielding the optimal control sequence. The outputs of the neurons of the network represent the values of the optimal control signal in each time step.

I. INTRODUCTION

N

computation has witnessed its successful applications in control and estimation problems in the past years. Many successful neural controller design methods are reported in the literature [1-10]. There are also a variety of papers discussing neural estimation and identification [2], [11-13]. Recurrent neural networks, especially Hopfield-type neural networks [14] are mainly employed in optimization and pattern recognition problems. However, they are also used in control and estimation problems. Karam et. al. solved the algebraic Riccati equation for robust optimal control of a nonlinear system utilizing a recurrent neural network [15]. As a noticeable application, Hopfield neural network was used to solve the optimal control problem for homing missile guidance [16]. In another significant approach, the Linear Quadratic optimal control problem was solved for discrete-time systems by Hopfield neural network such that the network yields the optimal control sequence [17], in contrast to the aforementioned approaches which make use of the network to find the parameters of the optimal controller. Furthermore, the LQ tracking problem was solved via this approach [18]. Shen and Balakrishnan [19] proposed a class of modified Hopfield networks for optimal control of linear and nonlinear systems. Such networks solve the State Dependent Riccati Equation (SDRE) finding the optimal control gain EURAL

sequence. In this paper, we use a discretized version of the continuous Hopfield model to solve the optimal disturbance rejection problem. It is a natural extension of the recent work on obtaining the optimal control sequence by Hopfield neural network. In section 2, we review the optimal disturbance rejection problem. In section 3, the classical theory of optimal control is employed to solve the optimal disturbance rejection problem. In section 4, Hopfield neural networks are introduced briefly. Additionally, it is shown that they are capable of solving the LQ optimal disturbance rejection problem provided that their parameters are set properly. In section 5, some simulation results are presented and finally, in section 6, some conclusions are derived. II. LQ OPTIMAL DISTURBANCE REJECTION: THE REGULATORY PROBLEM Assume the following linear time variant state equation: x k +1 = Ak x k + Bk u k

(1)

If it is affected by a known disturbance it becomes: x k +1 = Ak x k + Bk u k + Dk d k

(2)

in which x k ∈ ℜ n×1 is the state vector, u k ∈ ℜ r×1 is the input vector and d k ∈ ℜ p×1 is the vector of known disturbances. In the LQ regulatory problem it is desired that the state vector eventually tends to zero minimizing the following performance index subject to the state equations: Ji =

1 T 1 x i + N Hx i + N + 2 2

k =i + N −1

∑ ( x kT Qk x + u kT Rk u k )

(3)

k =i

Here H , S k ∈ ℜ n×n and Qk ∈ ℜ n×n k = i...N are positive M. R. Rajati is with the Department of Electrical Engineering, K. N. Toosi University of Technology, Shariati St., Seyyed Khandan, Tehran, Iran (phone: +98 918 8390095; fax: +98 21 303-555-5555; e-mail: mohammadreza.rajati@ gmail.com). H. Khaloozadeh is with the Department of Electrical Engineering, K. N. Toosi University of Technology, Shariati St., Seyyed Khandan, Tehran, Iran (e-mail: [email protected]). M. M. Korjani is with the Department of Electrical Engineering, Amirkabir University of Technology, Hafez Ave., Tehran, Iran (e-mail: [email protected]).

semi-definite matrices and Rk ∈ ℜ r×r is a positive definite matrix. Henceforth, the optimization problem reduces down to finding the vector of optimal control efforts U i = [u iT , u iT+1 ,..., u iT+ N −1 ]T . This is performed analytically via solving the Riccati equation. In the present work, the neural network is made to find the optimal control sequence. In order to translate the problem to one which the neural

network is capable of solving, we solve the system (3) in terms of the initial state vector, the vector of optimal control efforts, and the vector of disturbances: x k = Φ k xi + Ψk U i + ∆ k δ i

(4)

where Φ k is the state transformation matrix and is obtained by: k −1

Φk =

∏ Aj

(5)

Hk =

1 T ( x k Qk x k + u kT Rk u k ) 2

(10)

+ λTk +1 ( Ak x k + Bk u k + Dk d k ) The state and co-state equations are at hand: x k +1 =

λk =

∂H k = Ak x k + Bk u k + Dk d k ∂λ k +1

∂H k = Qk x k + AkT λ k +1 ∂x k

(11)

j =i

From the stationarity condition we have: Furthermore, Ψk ∈ ℜ

n×rN

and is defined by:

∂H k = 0 ⇒ Rk u k + BkT λ k +1 = 0 ∂u k

Ψk = [Ψk1 | Ψk2 | ... | ΨkN ] ⎧ k −1 ⎪( A j ) Bm −1 ⎪ j =m ⎪ = ⎨B m −1 ⎪ ⎪ ⎪0 ⎩



Ψkm

m
(6)

|

∆2k

| ... |

∆Nk

⎧ A j ) Dm −1 ⎪( ⎪ j =m ⎪ ∆mk = ⎨ D m −1 ⎪ ⎪ ⎪0 ⎩

m>k

λk = S k xk + vk

= [u iT

| u iT+1

(13)

(14)

] m
(7)

| ... | u iT+ N −1 ]

(15)

Substituting (14) in (15) we have:

x k +1 = Ak x k − B k R k−1 B kT ( S k +1 x k +1 + v k +1 )

(8)

+ Dk d k ⇒ ( I + B k R k−1 B kT S k +1 ) x k +1

And the vector of all disturbances is:

δ iT = [d iT | d iT+1 | ... | d iT+ N −1 ]

Note that vk is added to calculate the effect of the disturbance on the co-states. Replacing (12) in (2) we have the mixed state and co-state equations:

x k +1 = Ak x k − Bk Rk−1 BkT λ k +1 + Dk d k

m>k

U i ∈ ℜ rN ×1 is the vector of optimal control efforts:

U iT

1 ∂ ( xiT+ N Hxi + N ) = Hxi + N ∂xi + N 2

Applying the sweep method [20], we assume that the following relation holds between the states and co-states:

k −1



The boundary condition for the co-state equations is:

λi + N =

In the same way ∆ k ∈ ℜ n× pN is: ∆ k = [ ∆1k

(12)

⇒ u k = − Rk−1 BkT λ k +1

= Ak x k − B k R k−1 B kT v k +1 + D k d k

(9)

(16)

⇒ x k +1 = ( I + B k R k−1 B kT S k +1 ) −1 × ( Ak x k − B k R k−1 B kT v k +1 + D k d k )

III. ANALYTICAL SOLUTION OF THE OPTIMAL DISTURBANCE REJECTION PROBLEM In the optimal disturbance rejection problem, the Hamiltonian is:

From the equations (14) and (11) we have: S k x k + v k = Qk x k + AkT ( S k +1 x k +1 + v k +1 ) Thereby:

(17)

S k x k + v k = Qk x k + AkT v k +1 + AkT S k +1 ( I + Bk Rk−1 BkT S k +1 ) −1 × ( Ak x k −

Bk Rk−1 BkT v k +1

(18)

+ Dk d k )

Because these equations are valid for every k, we conclude that: S k = Qk + AkT S k +1 ( I + B k Rk−1 BkT S k +1 ) −1 Ak

× (− B k R k−1 B kT v k +1

+ Dk d k )

(21)

λi + N = S i + N xi + N + vi + N

(22)

S i+ N = H

(23)

Thus:

We are also able to determine the optimal control sequence:

u k = − R k−1 B kT λ k +1 = − R k−1 B kT ( S k +1 x k +1 + v k +1 ) = − R k−1 B kT ( S k +1 ( Ak x k + B k u k + D k d k + v k +1 )) u k = ( I + R k−1 B kT S k +1 B k ) −1

τ

1 E (t ) = − a T Wa − b T a 2 n f (n) = tanh( ) ε << 1

(20)

On the other hand, by the sweep method:

vi + N = 0

(26)

The network is proved to minimize the following energy function, provided that f is selected to be a steep sigmoid function:

This is an auxiliary equation which considers the effect of the disturbances. From the boundary conditions of the co-state equations, we have:

λi + N = Hxi + N

n(t + ∆t ) − n(t ) 1 = − n(t ) + Wa(t ) + b ∆t τ ∆t n(t + ∆t ) = (1 − )n(t ) + ∆t × Wa(t ) + ∆t × b

(19)

which is a Riccati equation. Also, we have the following equation for vk: v k = AkT v k +1 + AkT S k +1 ( I + B k R k−1 B kT S k +1 ) −1

the neurons and τ is an important time-constant of the network. f(n) is a sigmoid function and acts componentwise. For simulation purposes, an Euler approximation of the derivative is used:

(24)

× (− R k−1 B kT S k +1 ( Ak x k + D k d k + v k +1 ))

(27)

ε

From the above discussion, it is clear that in order to find the optimal control sequence, the connection weight matrix and the threshold terms should be determined such that E(t) corresponds to the performance index to be minimized, and the stabilized output of the network is the optimal control sequence. As we mentioned, we should find W and b such that minimization of the energy function of the network is equal to finding the optimal control sequence. Utilizing the solution of the discrete system of (3) which is described in (4), the objective function is written:

Ji =

1 T 1 i + N −1 T x i + N Hx i + N + ( x k Q k x k + u kT R k u k ) 2 2 k =i



1 (Φ i + N x i + Ψi + N U i + ∆ i + N δ i ) T 2 × H (Φ i + N x i + Ψi + N U i + ∆ i + N δ i ) =

(28)

⎡ (Φ k x k + Ψ k U i + ∆ k δ i ) T ⎤ ⎥ 1 i + N −1 ⎢ + ⎢× Q k (Φ k x k + Ψk U i + ∆ k δ i ) ⎥ 2 k =i ⎢ ⎥ T ⎣+ u k R k u k ⎦



We define the following diagonal partitioned matrix: IV. HOPFIELD NEURAL NETWORKS Hopfield neural network is a recurrent neural network driven by the following equations:

dn(t ) 1 = − n(t ) + Wa(t ) + b dt τ a (t ) = f (n(t ))

ρ = diag[ Ri , Ri +1 ,..., Ri + N −1 ] ρ ∈ ℜ rN ×rN

(29)

This enables us to write:

(25)

where n(t) is the synaptic signal and a(t) is the output of the network. W is the weight matrix, b is the threshold vector of

i + N −1

∑ u kT Rk u k = U iT ρU i

(30)

k =i

The aforementioned terms could be expanded, but it is

clear that some terms like xiT Φ Ti + N HΦ i + N xi which are independent of U i should be neglected, because they have no impact on the optimization procedure. ~ Thus we introduce the modified version of J i as J i : ~ 1 J i = ( 2 x iT Φ Ti + N HΨi + N + 2δ iT ∆Ti + N HΨi + N )U i 2 1 + U iT ΨiT+ N HΨi + N U i 2 i + N −1 ⎡ 2Ψ T Q Φ x ⎤ 1 k k k i + U iT ∑ ⎢ ⎥ 2 ⎢+ 2ΨkT Q k ∆ k δ i ⎦⎥ k =i ⎣

(31)

V. SIMULATION RESULTS In this section, we present a simple numerical example to validate our approach in using the neural network to solve the dynamic optimization problem. Consider the following single state system with an additive step disturbance:

i + N −1 1 1 + U iT ∑ ΨkT Q k Ψk U i + U iT ρU i 2 2 k =i

From the equation (31) it is obvious that in order for the ~ Hopfield network to minimize J i , the weight matrix and the threshold vector must be as follows: W = −2( ΨiT+ N HΨi + N +

Remark3. This result is interesting from another aspect. It is promising in the complete on-line solution of the optimal disturbance rejection problem, provided that the total elapsed time for identification of the plant, measurement of the initial conditions and adjusting the parameters of the Hopfield circuit is less than the sampling period of the system. It is not an unachievable condition with the current speeds of microprocessors. However, precise mathematical analysis, based on the convergence speed of the Hopfield network might be helpful.

i + N −1

∑ ΨkT Qk Ψk + ρ )

x k +1 = a k x k + bk u k + d k z k

(34)

Here, ak=0.6, bk=0.8, dk=0.3, x0=0.8 and zk is assumed to be a unit step disturbance. A Hopfield neural network is employed to minimize the cost function:

k =i

b = −2( ΨiT+ N HΦ i + N x i + ΨiT+ N H∆ i + N δ i +

(32)

i + N −1

∑ [ΨkT Qk Φ k xi + ΨkT Qk ∆ k δ i ])

2 J 0 = 10 x10 +

9

∑ (10 x 2j + 0.05u 2j )

(35)

j =0

k =i

The weight matrix of the network is symmetric, leading us to conclude that the network is stable [21]: dE dE dn ≤0 =0⇔ =0 dt dt dt

(33)

Remark1. The previous discussion on applying the network to solve the optimization problem in the interval [-1,1] is extendable to any interval [-β,β]. This could be achieved by modifying the gain of the amplifiers of the Hopfield circuit [21]. In case of the difficulties arisen in the hardware implementation, provided that a bound on the elements of the system’s input vector is known, the problem could be translated to a normalized version by manipulating the matrix B. The aforementioned condition, however, is a mild condition. Remark2. It is frequently mentioned in the literature that optimization via Hopfield-like networks suffers from the problem of local minima [21]. It is not, however, serious for our optimization problem. To clarify this, note that optimal control is a well-defined problem due to the restrictions on the weighting matrices Q and R. On the other hand, it is known that the abovementioned selections of the network parameters (proper W, b, τ, and a high gain amplifier with ε <<1) forces the network to solve the same optimization problem as the well-defined optimal control problem.

The state trajectories of the system are shown in Fig.1. It is easily seen that the state trajectories of the optimal and neural controllers are very similar. VI. CONCLUSION AND FUTURE WORK In this paper, we presented a Hopfield neural network for solving the optimal disturbance rejection problem. The network is able to solve the optimal control problem and might be suitable for online implementation. In the future, deviation of the control inputs calculated by the neural network from the optimal control signal should be calculated and the robustness of the system to such errors should be investigated. Furthermore, based on the convergence analysis of the Hopfield network, and the technological restrictions of implementation of the related circuitry, a theoretical bound on the calculation time of the control signal may be found. The results will be helpful in online implementation of such a controller.

[17] R. Xiaogang, “Linear quadratic dynamic optimization with Hopfield network for discrete-time systems,” Proc. of The 2nd World Congress on Intelligent Control and Automation, Xian, China, pp. 1880-1883, 1997. [18] L. Mingai, R. Xiaogang,, “Dynamic tracking optimization by continuous Hopfield neural network,” Proc. of the 5th World Congress on Intelligent Control and Automation, Hangzhou, China, pp. 25982602, June 2004. [19] J. Shen, S. N. Balakrishnan, “A class of modified Hopfield Networks for control of linear and nonlinear systems,” Proc. of the American Control Conference, Philadelphia, Penn, USA, pp. 964-969, June 1998. [20] F. L. Lewis, V. L. Syrmos, Optimal Control, Wiley, USA, 1995. [21] M. T. Hagan, H. B. Demuth, M. Beale., Neural Network Design, PWS Publishing Company, Boston, Mass., USA, 1996.

Fig.1. Optimal State Trajectories of The System

REFERENCES [1]

[2]

[3]

[4]

[5] [6]

[7]

[8]

[9]

[10]

[11] [12]

[13]

[14] [15]

[16]

N. F. Hunt, D. Sbarbaro, R. Zbikowski, P. J. Gawthrop, “Neural networks for control system-a survey,” Automatica, Vol. 28, No. 6, pp. 1083-1112, 1992. K. S. Narendra, K. Parthasarathy, “Identification and control of dynamical systems using neural networks,” IEEE Trans. Neural Networks, Vol. 1, No. 1, pp. 4-27, 1990. A. G. Barto, “Connectionist learning for control: an overview,” In Neural Networks for Control, W. Thomas Miller III, R. S. Sutton and P. J. Werbos Eds., MIT Press, Cambridge-Mass., pp. 5-58, 1990. J. S. Albus, “A new approach to manipulator control: the Cerebellar Model Articullation Controller (CMAC). Trans. ASME. J. Dyn. Sys., Mea. & Control, No. 97, pp. 220-233, 1975. S.S. Kumar, A. Guez, “ART based adaptive pole placement for neurocontrollers,” Neural Networks, vol. 4, pp. 319-335, 1991. J. M. Mendel, R. W. McLaren, “Reinforcement learning control and pattern recognition systems,” In Adaptive, Learning, and Pattern Recognition Systems: Theory and Applications, J. M. Mendel and K. S. Fu Eds., Academic Press, New York, pp. 287-318, 1970. A. G. Barto, R. S. Sutton, C. W. Anderson, “Neuronlike adaptive elements that can solve difficult learning control problems,” IEEE Trans. Syst. Man and Cybern. Vol. 13, No. 5, pp. 834-846, 1983. C. –T. Lin, C. S. G. Lee, “Neural- network-based fuzzy logic control and decision system,” IEEE Trans. Computers, Vol. 40, No. 12, pp. 1320-1336, 1991. Q. H. Wu, B. W. Hogg, G. W. Irwin, “A neural network regulator for turbogenerators,” IEEE Trans. Neural Networks, Vol. 3, No.1, pp. 95100, 1992. B. Widrow, “Adaptive inverse control,” Proc. IFAC Conference on Adaptive Systems in Control and Signal Processing, Lund, Sweden, pp. 1-5, 1986. R. S. Sutton, “Learning to predict by methods of temporal differences,” Mach. Learn, Vol. 3, pp. 9-44, 1988. S.-Z Qin, H.-T Su, T.J. McAvoy “Comparison of four neural net learning methods for dynamic system identification,” IEEE Trans. Neural Networks, Vol. 3, No.1, pp. 122-130, 1992. D. T. Pham, X. Liu, “State space identification of dynamic systems using neural networks,” Engineering Applications of Artificial Intelligence, Vol. 3, pp. 198-203, 1990. J. J. Hopfield, D. W. Tank, “Computing with neural circuits: a model,” Science, Vol. 233, pp. 625-633, 1986. M. Karam, A. Mohamed Zohdy and S. S. Ferinwata, “Robust Optimal Control using Recurrent Dynamic Neural Network,” Proceedings of the 2001 IEEE International Symposium on Intelligent Control, Mexico City, Mexico, pp. 331-336, 2001. E. S. Jamesand, S. N. Balakrishnan, , “Use of Hopfield neural networks in optimal guidance,” IEEE Trans. Aerospace and Electronic Syst., Vol. 30, No. 1, pp. 287-293, 1994.

Optimal Disturbance Rejection Using Hopfield Neural ...

(phone: +98 918 8390095; fax: +98 21 303-555-5555; e-mail: ..... on Intelligent Control and Automation, Hangzhou, China, pp. 2598-. 2602, June 2004.

110KB Sizes 1 Downloads 182 Views

Recommend Documents

Optimal Disturbance Rejection Using Hopfield Neural ...
Abstract— This paper investigates the problem of disturbance rejection and its solution by means of Hopfield neural networks. The dynamic optimization problem is transformed into a static optimization problem via linear state space analysis methods

Set-valued Observers And Optimal Disturbance Rejection
the NSF under Grant ECS–92258005, EPRI under Grant #8030–23, and Ford. Motor Co. The authors are with the Department of Aerospace Engineering and. Engineering Mechanics, The .... storage of all measurements. Reference [33] goes on to provide an a

Active Disturbance Rejection Control of a 2DOF ...
considered 2DOF manipulator lead to an increase of the kinetic energy during the ... acting disturbances, the ADRC method is an alternative that significantly ..... of path following, trajectory tracking, and energy efficiency (in means of control.

Output Covariance Tracking as a Disturbance Rejection ...
output transfer function from the state space equations: cov. 1 cov cov. ) (. )( B. AIz. CzG. U. Y. -. -. ×. = = (6). 46th IEEE CDC, New Orleans, USA, Dec. 12-14 ...

A Disturbance Rejection Supervisor in Multiple-Model Based Control
and Shahrokhi, 2000; Karimi and Landau, 2000; Boling et. al., 2007). In most of articles, MMST is assumed robust if each model- controller pair is individually ...

Asymptotic Disturbance Rejection for the TORA System ...
Jul 13, 2009 - Department of Computer and Systems Science Antonio Ruberti. Sapienza University of Rome. Italy. IASTED Conference on Control and ...

Sensitivity of optimal control for diffusion Hopfield ...
a Mechanical and Automation Engineering, The Chinese University of Hong Kong, .... put capacitance of the amplifier ith and its associated input lead. di > 0 are ...

Understanding the neural response to social rejection in adolescents ...
Achieving social acceptance and avoiding rejection by peers are ... ASD, using an experimental social rejection manipulation ..... being ignored over the Internet.

Understanding the neural response to social rejection in adolescents ...
of Autism and Developmental Disorders 40, 63–73. Williams, K.D., Cheung, C.K., Choi, W., 2000. Cyberostracism: effects of being ignored over the Internet.

Genetically Evolving Optimal Neural Networks - Semantic Scholar
Nov 20, 2005 - URL citeseer.ist.psu.edu/curran02applying.html. [4] B.-T. Zhang, H. Mühlenbein, Evolving optimal neural networks using genetic algorithms.

Genetically Evolving Optimal Neural Networks - Semantic Scholar
Nov 20, 2005 - Genetic algorithms seem a natural fit for this type .... For approaches that address network architecture, one termination criteria that has been.

Optimal Placement Optimal Placement of BTS Using ABC ... - IJRIT
Wireless Communication, since the beginning of this century has observed enormous ... phone users, thus cellular telephony become the most important form of ...

Confidence Scoring and Rejection using Multi ... - Vincent Vanhoucke
using Multi-Pass Speech Recognition. Vincent Vanhoucke. Nuance Communications, Menlo Park, CA, USA [email protected]. Abstract. This paper presents a computationally efficient method for us- ing multiple speech recognizers in a multi-pass framework

Neural Graph Learning: Training Neural Networks Using Graphs
many problems in computer vision, natural language processing or social networks, in which getting labeled ... inputs and on many different neural network architectures (see section 4). The paper is organized as .... Depending on the type of the grap

Optimal Placement Optimal Placement of BTS Using ABC ... - IJRIT
IJRIT International Journal of Research in Information Technology, Volume 2, Issue 4, April .... the control functions and physical links between MSC and BTS.

Confidence Scoring and Rejection using Multi ... - Vincent Vanhoucke
the system, however, the optimization of a multi-pass system has to obey a different set of .... ances which are not well modeled in the search space, it is often observed that an ... The first pass recognition engine used is a context-dependent ...

Disturbance regime and disturbance interactions in a
western North America, quantitative data describing ... located in the White River National Forest in north- ... aerial photographs and plotted on enlarged USGS.

Electromagnetic field identification using artificial neural ... - CiteSeerX
resistive load was used, as the IEC defines. This resistive load (Pellegrini target MD 101) was designed to measure discharge currents by ESD events on the ...

Evolving Nash-optimal poker strategies using evolutionary ...
Evolving Nash-optimal poker strategies using evolutionary computation.pdf. Evolving Nash-optimal poker strategies using evolutionary computation.pdf. Open.

Credit Card Fraud Detection Using Neural Network
some of the techniques used for creating false and counterfeit cards. ..... The illustration merges ... Neural network is a latest technique that is being used in.

Scalable Object Detection using Deep Neural Networks
neural network model for detection, which predicts a set of class-agnostic ... way, can be scored using top-down feedback [17, 2, 4]. Us- ing the same .... We call the usage of priors for matching ..... In Proceedings of the IEEE Conference on.

OPTIMAL FRAME STRUCTURE DESIGN USING ...
design coding structures to optimally trade off storage size of the frame structure with ..... [2] “Stanford Light Field Archive,” http://lightfield.stanford.edu/lfs.html.

Publication rejection among ecologists
It is widely recognized that anyone pursuing a career in the arts needs a thick skin to cope with the frequent rejection that they face on the path to career success.

Rejection and valuations
Mar 30, 2009 - We can define validity more snappily, if we help ourselves to a bit of ... hyper-valuation which makes the premisses true and conclusion false, ...