Optimal risk control and investment for Markov-modulated Brownian Motion∗ Xin Zhanga a

b



Ming Zhoub

School of Mathematical Sciences, Nankai University, Tianjin 300071, P.R. China

China Institute for Actuarial Sciences, Central University of Finance and Economics, Beijing 100081, P.R. China

Abstract : In this paper, we describe a large insurance company’s surplus by a Markovmodulated Brownian motion. The insurance company controls its risk by proportional reinsurance and invests in a Black-Scholes financial market. The problem of maximizing exponential utility function from terminal wealth is well solved.

Keywords: Markov-modulated Brownian motion; stochastic control; exponential utility function; Hamilton-Jacobi-Bellman equation. JEL code: G32. Subject Category and Insurance Branch Category: IE12, IE53, IB90, IB91. ∗

Supported by the National Natural Science Foundation of China(NSFC grant No.10571092) and Na-

tional Basic Research Program of China (973 Program) 2007CB814905 †

Corresponding Author E-mail address: [email protected]

1

1

Introduction

In the past few decades there has been a resurgence of attention drawn to the utilization of stochastic control theory in economics and finance. Browne (1995) considered the optimal investment polices for maximizing exponential utility and minimizing the probability of ruin. Schmidli(2001, 2002), Promislow and Young (2005) studied the problem of minimizing the probability of ruin, while Jeanblanc-Picqu´e and Shiryaev (1995), Asmussen and Taksar (1997), Taksar (2000), Højgaard and Taksar (2004) considered the strategy of maximizing the expected present value of the dividends paid out up to the time of bankruptcy. Most of above papers modeled the financial reserve of an insurance company as a Brownian motion with positive drift and constant diffusion coefficient. However, in reality, the financial reserve of an insurance company often depends on the dynamical state of economy. In order to take into account this, Asmussen (1989) proposed the Markov-modulated risk model and studied the ruin problem of this model. The model is also called Markovian regime switching model in the finance and actuarial science literature. Such a model can incorporate the impact states of the economy have on the surplus process. Therefore it is more reasonable to model the financial reserve of an insurance company by the Markovian regime switching model. In order to get an closed form of our problems, we assume that the financial reserve of an insurance company is modeled by Markov-modulated Brownian motion. The idea of modeling by Markov-modulated stochastic process is not new in finance and actuarial science. See Reinhard (1984), Asmussen (1989), Miyazawa (2004), Rabeha2

saina and Sericola (2004), Jacobsen (2005), Lu and Li (2005) et al for an application of Markov-modulated model in actuarial science. In financial literature, most works mainly employed the Markovian regime switching model to deal with options, see Naik (1993), Di Masi, Kabanov and Runggaldier (1994), Yao, Zhang, and Zhou (2001), Buffington and Elliott (2002), Graziano and Rogers (2005). Besides dealing with option, the portfolio optimization problems in Markovian regime switching market were also studied by many authors. For instance, Cajueiro and Yoneyama (2002) studied the combined optimization of portfolio and risk exposure of an insurance company whose potential profit depends on the dynamic state of the economy. B¨auerle and Rieder (2004) investigated portfolio optimization problems with Markov-modulated stock prices and interest rates. Rieder and B¨auerle (2005) studied the problem of portfolio optimization with unobservable Markovmodulated drift process. This paper is organized as follows. In section 2, we introduce the model of Markovmodulated Brownian motion. Section 3 is devoted to formulate the control problem of maximizing utility function from terminal wealth by Hamilton-Jacobi-Bellman (HJB for abbreviation) equation. We solve the HJB equation and find the optimal strategy for maximizing the exponential utility function in section 4. In section 5 we consider the control problem of maximizing exponential utility function when the positive interest rate is taken into account.

3

2

The Model

To give a rigorous mathematical formulation of the optimization problem, we start with a probability space (Ω, F, P ) with a filtration {Ft }t≥0 containing all objects defined at the following. The filtration Ft represents the information available at time t and any control is made based upon this information. We assume in the case of no risk or investment control, the reserve of the company Rt evolves according to the Markov-modulated Brownian motion described as follows: Z Rt = x +

0

Z

t

µJs ds +

t

0

σJs dBs ,

(2.1)

where Bt is a standard Brownian motion and {Jt , t ≥ 0} is an irreducible continuous time Markov chain with a finite states space S. Let Λ = (λij )i,j∈S denote the intensity matrix for Jt and write for brevity λi = −λii for the rate parameter of exponential holding time in state i. Since Λ is irreducible, the stationary limiting distribution π of Jt exists and can be computed as the positive solution of πΛ = 0, πe = 1. Indeed the process {Rt , t ≥ 0} is a special continuous time Markov additive process introduced by Asmussen (2000). Without loss of generality we assume that there is only one risky stock available for investment, whose price process {St } is described by dSt = αSt dt + βSt dBtS , where BtS is another standard Brownian motion which is independent of Bt . In addition we also consider that the company controls its risk by a proportional reinsurance. A strategy π is described by a two dimensional stochastic process (aπ (t), bπ (t)), where 4

0 ≤ aπ (t) ≤ 1 represents the proportional reinsurance level and bπ (t) represents the amount invested in the risky asset at time t. Usually we will put no constrains on the control bπ (t). If bπ (t) < 0 then this means that the insurance company is shorting the stock, while bπ (t) > Rtπ corresponds to borrow money for investing long in stock. Denote by Rtπ the resulting reserve process when applying strategy π, thus the dynamics for Rtπ is then given by, dRtπ = aπ (t)[µ(Jt )dt + σ(Jt )dBt ] + bπ (t)(αdt + βdBtS ) = [aπ (t)µ(Jt ) + αbπ (t)]dt + aπ (t)σ(Jt )dBt + bπ (t)βdBtS ,

(2.2)

where we assume that R0π = x0 is the initial reserve. We denote by Π[t, T ] := {π(s) = (aπ (s), bπ (s))t≤s≤T : π(s) ∈ [0, 1] × R, π is Fs -adapted Z T and b2π (s)ds < ∞ almost surely}. t

Suppose now that the insurer is interested in maximizing the utility function from his terminal wealth, say at time T. The utility function is u(x), typically increasing and concave (u00 (x) < 0). For a strategy π , we define the utility attained by the insurer from the state x at time t as Vπi (t, x) = E[u(RTπ )|Rtπ = x, Jt = i]. Then our objective is to find the optimal value function V i (t, x) =

sup E[u(RTπ )|Rtπ = x, Jt = i]

(2.3)

π∈Π[t,T ]

and the optimal strategy π ∗ such that V i (t, x) = Vπi∗ (t, x) =

5

sup Vπi (t, x).

π∈Π[t,T ]

(2.4)

3

HJB equation

We use the dynamic programming approach described in Fleming and Soner (1993) to solve above problems. By the standard arguments of Fleming and Soner (1993), we have the following theorem. Theorem 3.1. Assume that V i (t, x) is twice continuous differentiable. Then V i (t, x) satisfies the following HJB equation,     X 1 i sup Vti + [aµi + αb]Vxi + (a2 σi2 + β 2 b2 )Vxx + λij V j = 0  2 π∈Π[t,T ] 

(3.5)

j∈S

with the terminal data V i (T, x) = u(x).

(3.6)

In what follows we will give two lemmas which will be used in the analysis of the solution to the HJB equation. For any function V i (t, x) denote aV (t, x, i) = −

µi Vxi (t, x) . i (t, x) σi2 Vxx

(3.7)

i (t, x) < 0 is a solution Lemma 3.1. Suppose V i (t, x) with property Vxi (t, x) > 0 and Vxx

to Vti (t, x)

· ¸ 1 µ2i α2 Vxi2 (t, x) X − + + λij V j (t, x) = 0 i (t, x) 2 σi2 β 2 Vxx

(3.8)

j∈S

with terminal boundary V i (T, x) = u(x). Let O = {(t, x, i) : aV (t, x, i) < 1}. 6

(3.9)

Then V i (t, x) satisfies the HJB equation (3.5) on O i (t, x) < 0 and the left hand of (3.5) is a second degree Proof: Since Vxi (t, x) > 0, Vxx

polynomial in a and b, differentiating with respect to a and b in (3.5) gives the maximizer a∗ = −

µi Vxi (t, x) , i (t, x) σi2 Vxx

b∗ = −

α Vxi (t, x) . i (t, x) β 2 Vxx

aV (t, x, i) = −

µi Vxi (t, x) < 1. i (t, x) σi2 Vxx

(3.10)

Since on O

Substitute a∗ and b∗ instead of a and b in (3.5) respectively and after some calculations yield the following nonlinear Cauchy problem: Vti (t, x)

· ¸ 1 µ2i α2 Vxi2 (t, x) X − + + λij V j (t, x) = 0, i (t, x) 2 σi2 β 2 Vxx

for t < T,

j∈S

V i (T, x) = u(x). Thus the desired result is obtained.

¤

i (t, x) < 0 is a solution Lemma 3.2. Suppose V i (t, x) with property Vxi (t, x) > 0 and Vxx

to Vti (t, x) + µi Vxi (t, x) −

X 1 α2 Vxi2 (t, x) σi2 i + V (t, x) + λij V j (t, x) = 0 xx i (t, x) 2 β 2 Vxx 2

(3.11)

j∈S

with terminal boundary V i (T, x) = u(x). Let O1 = {(t, x, i) : aV (t, x, i) ≥ 1}. Then V i (t, x) satisfies the HJB equation (3.5) on O1 7

(3.12)

Proof: Since the equation (3.5) is a second degree polynomial in a and b and aV (t, x, i) = −

µi Vxi (t, x) ≥ 1, i (t, x) σi2 Vxx i

. Substitute a∗ = 1 and b∗ instead the maximizer is attained at a∗ = 1 and b∗ = − βα2 VVix (t,x) (t,x) xx

of a and b in (3.5) respectively and after some simplification we obtained (3.11).

4

¤

The solution of HJB equation for exponential utility function

In this section we will give the solution of HJB equation and find the optimal strategy for exponential utility function. Suppose now that the insurer has an exponential utility function u(x) = λ −

γ −θx e , θ

(4.13)

where γ > 0 and θ > 0. Such utility functions play a prominent role in insurance mathematics and actuarial practice, since they are the only utility functions under which the principle of “zero utility” gives a fair premium that is independent of the level of reserve of an insurance company. From now on, we will denote ei the n-dimensional row vector with all components equal to 0 but the i-th equal to 1 and denote e the n-dimensional column vector with all components equal to 1. Theorem 4.1. Let hi =

· ¸ · ¸ 1 µ2i α2 θ2 σi2 α2 µi + θµi − + 1 1 µi + 2 σi2 β 2 θ> σi2 2 2β 2 θ≤ σi2 8

(4.14)

and denote A = (h1 , h2 , · · · , hn )diag − Λ.

(4.15)

γ −θx e f (t, i) θ

(4.16)

Then the function W i (t, x) = λ −

with f (t, i) = ei e−A(T −t) e is a solution of HJB equation (3.5). Proof: By Lemma 3.1, Lemma 3.2, we know that the HJB equation is equivalent to equation (3.8), (3.11) under different cases. To solve equation (3.8), (3.11), we try to fit a solution of the form W i (t, x) = λ −

γ −θx e f (t, i) θ

(4.17)

where f (t, i) is a suitable function. Note that for this trial solution we have γ Wti (t, x)(t, x) = − e−θx f 0 (t, i), θ Wxi (t, x) = γe−θx f (t, i),

(4.18)

i Wxx (t, x) = −γθe−θx f (t, i).

Thus aW (t, x, i) =

µi θσi2

(4.19)

By comparing aW (t, x, i) and 1, we have two cases to consider. If 0 < aW (t, x, i) < 1, i.e. θ >

µi , σi2

then from Lemma 3.1 we know that the HJB equation

(3.5) is equivalent to partial differential equation (3.8). Inserting (4.18) into (3.8) and canceling like terms shows that we require that f (t, i) satisfies · ¸ X α2 1 µ2i + 2 f (t, i) − λij f (t, j). f (t, i) = 2 2 σi β 0

j∈S

9

(4.20)

In addition the boundary condition W i (t, x) = λ− γθ eθx implies that we must have f (T, i) = 1. If aW (t, x, i) ≥ 1, i.e. θ ≤

µi , σi2

then from Lemma 3.2, we know that the HJB equation is

equivalent to partial differential equation (3.11). Inserting (4.18) into (3.11) and canceling like terms shows that f (t, i) must satisfy · ¸ X α2 θ2 σi2 f 0 (t, i) = θµi − + 2 f (t, i) − λij f (t, j). 2 2β

(4.21)

j∈S

Also by the boundary condition W i (t, x) = λ − γθ eθx , we must have f (T, i) = 1. Let f (t) = (f (t, 1), f (t, 2), · · · , f (t, n))0 . From the definition of A, we can rewrite equations (4.20), (4.21) in the matrix form as f 0 (t) = Af (t)

(4.22)

with boundary condition f (T ) = e. Thus from the theory of differential equation, we know that f (t) = eAt x0 where x0 is determined by eAT x0 = e. That is f (t, i) = ei e−A(T −t) e.

(4.23)

In order to apply Lemma 3.1, 3.2, we have to verify that Wxi (t, x) > 0,

i Wxx (t, x) < 0.

From equation (4.18), we only need to prove f (t, i) > 0. In fact, from the definition of matrix A, we can find a positive constant k such that all components of matrix kI − A are positive. Therefore f (t, i) = ei e−A(T −t) e = e−k(T −t) ei e(kI−A)(T −t) e > 0. 10

Thus we complete the proof of the theorem.

¤

The next theorem provides the verification that W i (t, x) as given in Theorem 4.1 is indeed the value function of our problem. Theorem 4.2. Suppose that W i (t, x) is as given in Theorem 4.1. Then 1. V i (t, x) = W i (t, x), 2. The optimal strategy π ∗ = (π ∗ (s)) that maximizes exponential utility at a terminal time T is given in feedback form as ∗



π ∗ (s) = (aπ∗ (s, Rsπ , Js ), bπ∗ (s, Rsπ , Js )),

(4.24)

where the functions aπ∗ and bπ∗ are given by aπ∗ (s, x, i) =

µi µi + I{θ≤ µi } , I θσi2 {θ> σi2 } σ2 i

bπ∗ (s, x, i) =

α . θβ 2

Proof: Let ΠM be the set of strategies π = (aπ (t), bπ (t)) such that bπ (t) ∈ [−M, M ] for i (t, x). Let π ∈ Π M ∈ R+ large and denote the corresponding value function by VM M be an

arbitrary strategy and Rtπ the corresponding reserve process. Since W i (t, x) is sufficiently smooth, we can apply Itˆo’s formula to obtain Z v AW Js (s, Rsπ , aπ (s), bπ (s))ds W Jv (v, Rvπ ) = W Jt (t, Rtπ ) + t Z v WxJs (s, Rsπ )[aπ (t)σ(Jt )dBt + bπ (t)βdBtS ], +

(4.25)

t

where X 1 i (t, x) + λij W j (t, x). AW i (t, x) = Wti (t, x) + [aµi + αb]Wxi (t, x) + (a2 σi2 + β 2 b2 )Wxx 2 j∈S

(4.26) 11

Let us then set Z Mv =

v

t

WxJs (s, Rsπ )[aπ (s)σ(Js )dBs + bπ (s)βdBsS ].

(4.27)

Since Bt and BtS are two independent standard Brownian Motion, we can rewrite Mv as Z Mv =

t

v

p WxJs (s, Rsπ ) a2π (s)σ 2 (Js ) + b2π (s)β 2 dWs ,

(4.28)

where Ws is another standard Brownian motion. In general, {Mv , t ≤ v ≤ T } is a local martingale. In the present case, we will show that Mv is actually a real martingale. By equation (2.2), we have Z Rsπ = x +

Z

s

[aπ (r)µ(Jr ) + αbπ (r)]dt +

t

sp

a2π (r)σ 2 (Jr ) + b2π (r)β 2 dWr .

t

(4.29)

Since Wxi (t, x) = γe−θx f (t, i), we have Z Mv =

v

t

p π γe−θRs f (s, Js ) a2π (s)σ 2 (Js ) + b2π (s)β 2 dWs .

(4.30)

Since f, µ, σ, aπ and bπ are bounded on[t, T ], in order to verify that Mv is a martingale we need to show Z

½

T

E

Z

exp −2θ t

sp

t

¾ a2π (r)σ 2 (Jr )

+

b2π (r)β 2 dWr

ds < ∞.

(4.31)

From the facts that f, µ, σ, aπ and bπ are bounded on[t, T ], it is easy to verify the Novikov condition(see page 198 of Karatzas and Shreve (1999)). ½ Z s ¾ 1 2 2 2 2 2 E exp (2θ) [aπ (r)σ (Jr ) + bπ (r)β ]dr < ∞ 2 t

(4.32)

and therefore {Zs }s≥t defined by ½ Zs = exp −2θ

Z t

sp

a2π (r)σ 2 (Jr )

+

b2π (r)β 2 dWr

1 − 2

Z

s

¾ 2

(2θ) t

[a2π (r)σ 2 (Jr )

+

b2π (r)β 2 ]dr (4.33)

12

is a martingale with EZs = 1. Thus from (4.32) and (4.33), we can verify condition (4.31) is satisfied and therefore Mv is a martingale. Taking condition expectation given Rtπ = x and Jt = i and let v = T in equation (4.25), we obtain ·Z E[W

JT

(T, RTπ )|Rtπ

i

= x, Jt = i] = W (t, x)+E

T

AW t

Js

(s, Rsπ , aπ (s), bπ (s))ds|Rtπ

¸ = x, Jt = i . (4.34)

From Theorem 4.1, we know that W is a solution of HJB equation (3.5) and W i (T, RTπ ) = u(RTπ ). Therefore, E[u(RTπ )|Rtπ = x, Jt = i] ≤ W i (t, x).

(4.35)

i (t, x) ≤ W i (t, x). Now suppose that π ∗ is given by Since π is arbitrary, we obtain VM

(4.24), it is easy to verify that π ∗ maximizes the HJB equation from Theorem 4.1. Thus i (t, x) = W i (t, x). we get equality in (4.35) under π ∗ . Therefore VM i (t, x) are not depend on M . Note that the optimal strategy and the value function VM

Thus taking the limit M → ∞ yields the results of this theorem.

5

¤

Positive interest rate

In this section we consider the case where there is a positive interest rate r > 0. That is there is a bond beside the risky stock. Suppose that the price process of bond St0 evolves as dSt0 = rSt0 dt. In this case, any wealth not invested in the stock, Rtπ − bπ (t), will be held

13

in the bond. For any policy π, the wealth process evolves as dRtπ = aπ (t)[µ(Jt )dt + σ(Jt )dBt ] + [Rtπ − bπ (t)]rdt + bπ (t)(αdt + βdBtS ) = [aµ(Jt ) + r(Rtπ − bπ (t)) + αbπ (t)]dt + aσ(Jt )dBt + bπ (t)βdBtS .

(5.36)

The generator of the wealth process in this case is Ar g i (t, x) = gti (t, x) + [aµi + rx − (r − α)b]gxi (t, x) X 1 i + (a2 σi2 + β 2 b2 )gxx (t, x) + λij g j (t, x). 2

(5.37)

j∈S

We extend the results of Section 3 regarding maximizing exponential utility to this case. The corresponding HJB equation for maximizing utility from terminal wealth in this case is sup Ar V i (t, x) = 0

(5.38)

V i (T, x) = u(x).

(5.39)

π

and the terminal boundary is

Here again we have V i (t, x) = supπ∈Π E[u(RTπ )|Rtπ = x, Jt = i]. To solve the problem, we must solve the nonlinear equation (5.38) and find the value (aπ , bπ ) which maximizes the function X 1 i Vti (t, x) + [aµi + rx − (r − α)b]Vxi (t, x) + (a2 σi2 + β 2 b2 )Vxx (t, x) + λij V j (t, x). (5.40) 2 j∈S

Similar to the case of no positive interest rate, the HJB equation of (5.38) also have two equivalent forms. Assuming that the HJB equation of (5.38) has a classical solution V which satisfies Vx > 0 and Vxx < 0, therefore we can differentiate with respect to a and b

14

in (5.40) to find the maximizer a∗ = −

µi Vxi (t, x) , i (t, x) σi2 Vxx

b∗ = −

α − r Vxi (t, x) . i (t, x) β 2 Vxx

(5.41)

If a∗ < 1, then substituting a∗ and b∗ instead of a and b in (5.38) respectively we obtained the equivalent form of HJB equation of (5.38), Vti (t, x)

+

rxVxi (t, x)

· ¸ 1 µ2i (α − r)2 Vxi2 (t, x) X − + λij V j (t, x) = 0. + i (t, x) 2 σi2 β2 Vxx

(5.42)

j∈S

If a∗ ≥ 1, then substituting 1 and b∗ instead of a and b in (5.38) respectively we obtained another equivalent form of HJB equation of (5.38), 1 1 (α − r)2 Vxi2 (t, x) X i Vti (t, x)+(µi +rx)Vxi (t, x)+ σi2 Vxx (t, x)− + λij V j (t, x) = 0. (5.43) i (t, x) 2 2 β2 Vxx j∈S

Both of above two equations have the same terminal boundary condition V i (T, x) = u(x). Theorem 5.1. Let f (t) = (f (t, 1), f (t, 2), · · · , f (t, n))0 , · ¸ 1 µ2i 1 2 2 2r(T −t) r(T −t) µ I h(t, i) = + µi θe − σi θ e I{θ≤ µi } (5.44) i 2 σi2 {θ> er(T −t) σi2 } 2 er(T −t) σ 2 i and denote A(t) = (h(t, 1), h(t, 2), · · · , h(t, n))diag − Λ.

(5.45)

Then 1. The function given by ¾ ½ γ (α − r)2 r(T −t) W (t, x) = λ − exp −θxe (T − t) f (t, i) − θ 2β 2 i

15

(5.46)

with f (t) determined by f 0 (t) = A(t)f (t) and f (T ) = e is a solution of HJB equation (5.38). 2. The Feynman-Kac formula yields the following stochastic representation of f

· Z f (t, i) = E exp{− t

T

¸ h(s, Js )ds}|Jt = i ,

(5.47)

where h(t, i) is given by (5.44). Proof: To solve the HJB equation (5.38), we try to fit a solution of the form ½ ¾ (α − r)2 γ r(T −t) − (T − t) f (t, i), W (t, x) = λ − exp −θxe θ 2β 2 i

(5.48)

where f (t, i) is a suitable function. The boundary condition (5.39) implies that f (T, i) = 1. Noted that for the trial solution we have Wxi (t, x) = θer(T −t) g(t, x)f (t, i), i Wxx (t, x) = −θ2 e2r(T −t) g(t, x)f (t, i), · ¸ (α − r)2 i r(T −t) Wt (t, x) = − rθxe + g(t, x)f (t, i) − g(t, x)f 0 (t, i), 2β 2

where g(t, x) =

γ θ

n exp −θxer(T −t) −

(α−r)2 (T 2β 2

(5.49)

o − t) .

i (t, x) into (5.41), we obtain that the maximizer is Thus substituting Wxi (t, x) and Wxx

given by µi 1 , θer(T −t) σi2 1 α−r b∗ (t, x, i) = r(T −t) . β2 θe a∗ (t, x, i) =

16

(5.50) (5.51)

Therefore, if a∗ < 1, i.e. θ >

µi , er(T −t) σi2

inserting (5.49) into (5.42) shows that f (t, i) must

satisfy f 0 (t, i) =

X 1 µ2i λij f (t, j). f (t, i) − 2 2 σi

(5.52)

j∈S

If a∗ ≥ 1, i.e. θ ≤

µi

er(T −t) σi2

, inserting (5.49) into (5.43) shows that f (t, i) must satisfy

· ¸ X 1 2 2 2r(T −t) r(T −t) f (t, i) = µi θe − σi θ e f (t, i) − λij f (t, j). 2 0

(5.53)

j∈S

From the notation of A(t) and f (t), we can rewrite (5.52) and (5.53) in the matrix form as f 0 (t) = A(t)f (t).

(5.54)

f (T ) = e.

(5.55)

Moreover the boundary condition is

Now we prove the Feynman-Kac representation of f . Let Z K(v) = exp{− t

v

h(s, Js )ds}.

Then the Dynkin formula implies that the process {M (v), v ≥ t} defined by   Z v X f 0 (s, Js ) + M (v) = f (v, Jv ) − f (t, Jt ) − λJs ,j f (s, j) ds t

j∈S

is a martingale. Thus d(K(v)f (v, Jv )) = K(v)(df (v, Jv ) − h(v, Jv )f (v, Jv )dv)   X = K(v) dM (v) + [f 0 (v, Jv ) + λJv ,j f (v, j)]dv − h(v, Jv )f (v, Jv )dv  . j∈S

17

(5.56)

(5.57)

This yields Z K(v)f (v, Jv )−f (t, Jt ) =

t

v

Z v X K(s)dM (s)+ [f 0 (v, Jv )+ λJv ,j f (v, j)−h(v, Jv )f (v, Jv )]dv. t

j∈S

(5.58) Let v = T and from equation (5.53), (5.55) we obtain Z K(T ) − f (t, Jt ) =

T

K(s)dM (s).

(5.59)

t

Taking expectation with respect to Jt = i yields the desired result. From (5.47), it is easy to see f (t, i) > 0 and therefore W i (t, x) given by (5.46) is indeed the solution of HJB equation (5.38)

¤

Remark 5.1. The differential equation system f 0 (t) = A(t)f (t)

(5.60)

with boundary condition f (T ) = e indeed has a unique continuous solution( see Theorem 1 on page 303 of Bronson (1991)). In fact, if we let Φ(t, t0 ) be the transition matrix of differential equation system (5.60), i.e. Φ(t, t0 ) is n × n matrix having the properties that d Φ(t, t0 ) = A(t)Φ(t, t0 ), dt

Φ(t, t0 ) = E,

where E is the identity matrix, then f (t) = Φ(t, T )e is the unique solution of (5.60). For the existence of the transitions matrix, see page 339 of Bronson(1991). The next theorem provides the verification that W i (t, x) as given in Theorem 5.1 is indeed the value function of our problem. Theorem 5.2. Suppose that W i (t, x) is as given in Theorem 5.1. Then 18

1. V i (t, x) = W i (t, x), 2. The optimal strategy π ∗ = (π ∗ (s)) that maximizes exponential utility at a terminal time T is given in feedback form as ∗



π ∗ (s) = (aπ∗ (s, Rsπ , Js ), bπ∗ (s, Rsπ , Js )),

(5.61)

where the functions aπ∗ and bπ∗ are given by µi µi I + I{θ≤ µi } , σi2 {θ> er(T −t) σi2 } er(T −t) σ 2 i 1 α−r bπ∗ (s, x, i) = r(T −t) . β2 θe aπ∗ (s, x, i) =

1

θer(T −t)

Proof: The method of proving this theorem is similar to Theorem 4.2. The only difference is the expression of Rtπ . In this case, from the differential equation (5.36), the explicit expression of Rtπ is given by Rtπ

½ ¾ Z t Z t p −rs −rs 2 2 2 2 =e x+ e [aµ(Jt ) + (α − r)bπ (t)]dt + e aπ (s)σ (Js ) + bπ (s)β dWs , rt

0

0

(5.62) where Wt is another standard Brownian motion. Following steps of proving Theorem 4.2, we can also prove this theorem.

¤

References Asmussen, S., 1989. Risk theory in a Markovian environment. Scandinavian Actuarial Journal, 69-100. Asmussen, S., 2000. Ruin probabilities. Singapore: World Scientific. Asmussen, S., Taksar M., 1997. Controlled diffusion models for optimal dividend pay-out. 19

Insurance Mathematics and Economics 20, 1-15. Bronson, R., 1991. Matrix methods: an introduction. Boston: Academic Press. Browne, S., 1995. Optimal investment policies for a firm with a random risk process: Exponential utility and minimizing the probability of ruin. Mathematics of Operations Research 20, 937-958. Cajueiro, D.O., Yoneyama, T., 2002. Combined optimization of portfolio and risk exposure of an insurance company. Proceedings of the International Symposium on Mathematical Theory of Networks and Systems (MTNS 2002), 1-11. Fleming, W.H., Soner, H.M., 1993. Controlled Markov processes and viscosity solutions. Berlin Heidelberg New York: Springer. Ghosh, M. K., Arapostathis, A., Marcus, S. I., 1993. Optimal control of switching diffusions with application to flexible manufacturing systems. SIAM Journal on Control and Optimization 31(5), 1183–1204. Højgaard, B. Taksar, M., 2004. Optimal dynamic portfolio selection for a corporation with controllable risk and dividend distribution policy. Quantitative Finance 4, 315-327. Jacobsen, M., 2005. The time to ruin for a class of Markov additive risk process with two-sided jumps. Advances in Applied Probability 37(4), 963-992. Jeanblanc-Picqu´e, M., Shiryaev, A.N., 1995. Optimization of the flow of dividends. Russian Mathematical Surveys 50, 257-277. Karatzas, I., Shreve, S.E., 1999 Brownian motion and stochastic calculus. New York: Springer-Verlag. 20

Lu, Y., Li, S., 2005. On the probability of ruin in a Markov-modulated risk model. Insurance: Mathematics and Economics 37(3), 522-532. Miyazawa, M., 2004. Hitting probabilities in a Markov additive process with linear movements and upward jumps: application to risk and queueing processes. The Annals of Applied Probability 14(2), 1029-1054. Pham, H., 2005. On the smooth-fit property for one-dimensional optimal switching problem. to appear in S´ eminaire de Probabilit´ es Vol. XL. Promislow, S.D., Young V.R., 2005. Minimizing the probability of ruin when claims follow Brownian motion with drift. North American Actuarial Journal 9(3), 109-128. Rabehasaina, L., Sericola, B., 2004. A second-order Markov-modulated fluid queue with linear service rate. Journal of Applied Probability 41(3), 758-777. Reinhard, J.M., 1984. On a class of semi-Markov risk models obtained as classical risk models in a Markovian environment. STIN Bulletin International Actuarial Association 14, 23-44. Schmidli, H., 2001. Optimal proportional reinsurance polices in a dynamic setting. Scandinavian Actuarial Journal 101(1), 55-68. Schmidli, H., 2002. On minimizing the ruin probability by investment and reinsurance. The Annals of Applied Probability 12, 890-907. Taksar, M., 2000. Optimal risk and dividend distribution control models for an insurance company. Mathematical Methods of Operations Research 51(1), 1-42.

21

Optimal risk control and investment for Markov ...

Xin Zhanga †. Ming Zhoub. aSchool of ... Di Masi, Kabanov and Runggaldier (1994), Yao, Zhang, and Zhou (2001), Buffington and. Elliott (2002), Graziano and ...

178KB Sizes 1 Downloads 283 Views

Recommend Documents

Speculative Markov Blanket Discovery for Optimal ...
It is often the case that an engineer or researcher is inter- ested in one particular ... ily “read off” the network structure: The Markov blanket of an attribute is the set ...

Speculative Markov Blanket Discovery for Optimal Feature Selection
the remaining attributes in the domain. Koller and Sahami. [4] first showed that the Markov blanket of a given target at- tribute is the theoretically optimal set of ...

Convex Synthesis of Optimal Policies for Markov ...
[11], automatic control [12], flight control [13], economics. [14], revenue ... Emails: [email protected] and [email protected] through a sequence of repeated experiments. ...... can send the vehicle to “left” with some probability). For

Optimal control and vanishing viscosity for the Burgers ...
Apr 13, 2009 - We focus on the 1−d Burgers equation although most of our results extend to more general .... viewing it as an approximation of the inviscid one (1.1) as ν → 0 and ... present some numerical experiments that show the efficiency of

Investment Guarantees: Modeling and Risk ...
company in the United States. .... Phelim Boyle, for his unstinting encouragement, support, and patience; ..... where U.K. insurance companies were influential.

Sales Planning and Control Using Absorbing Markov ...
A stochastic model that generates data for sales planning and control is described. An example is .... the salesman opportunities to call on his most valuable cus- tomers. 63 ..... The Institute of Business and Economic. Research, 1966, 45-56.

Robust Optimal Investment in Discrete Time for ...
An element of Ωt will be denoted by ωt = (ω1 ... Assumption 2.1 For all 0 ≤ t ≤ T − 1, we assume that Qt+1 is a non-empty and convex valued random set such ...

OPTIMAL CONTROL SYSTEM.pdf
How optimal control problems are classified ? Give the practical examples for each classification. 10. b) Find the extremal for the following functional dt. 2t. x (t) J.

OPTIMAL CONTROL SYSTEMS.pdf
OPTIMAL CONTROL SYSTEMS.pdf. OPTIMAL CONTROL SYSTEMS.pdf. Open. Extract. Open with. Sign In. Main menu. Displaying OPTIMAL CONTROL ...

OPTIMAL CONTROL SYSTEMS.pdf
... time and fixed end state. problem. Indicate the different cases of Euler-Lagrange equation. 10. 2. a) Find the extremal of a functional. J(x) [ ] x (t) x (t) x (t) x (t) dt.

On Optimal Investment for a Behavioural Investor in ...
Mar 5, 2013 - The issue of well-posedness is a recurrent theme in related papers (see ..... We add a comment on the case α+ = α− assuming, in addition, that ...

Optimal Investment under Credit Constraint
We characterize the optimal investment decision, the financing and ...... [6] Dixit, A. and R. Pindyck, 1994, Investment under Uncertainty, Princeton University.

Optimal Debt Maturity and Firm Investment - Barcelona GSE Events
Data on default rates is from Giesecke, Longstaff, ... To the best of our knowledge, firms' optimal maturity choice over the business ... In Gomes and Schmid (2016), firms are required to buy ...... moment is the debt recovery rate in bankruptcy.

Optimal investment taxes and efficient market provision ...
Feb 15, 2014 - interest rate R in the private retrade market and the level of income I that agents ... The illiquid, long-term asset pays nothing at t = 1 and ˆR > 1.

Non-concave optimal investment and no-arbitrage: a ...
Jul 20, 2016 - the existence of an optimal portfolio in a (generically incomplete) discrete-time financial market model with finite time horizon. Key words: no-arbitrage condition ; non-concave utility functions; optimal investment. AMS 2000 subject

Optimal Debt Maturity and Firm Investment - Barcelona GSE Events
recovery rate of 69% for Chapter 11 re-organizations and 27% for Chapter 7 ...... Billett, M. T., T.-H. D. King, and D. C. Mauer (2007): “Growth opportunities.

Iterative Learning Control for Optimal Multiple-Point Tracking
on the system dynamics. Here, the improved accuracy in trajectory tracking results has led to the development of various control schemes, such as proportional ...

Variational optimal control technique for the tracking of ... - Irisa
IRISA/INRIA Campus de Beaulieu 35042 Rennes Cedex, France npapadak ... a term related to the discrepancy between the state vari- ables evolution law and ...

Evolution of Optimal ANNs for Non-Linear Control ...
recognition, data processing, filtering, clustering, blind signal separation, compression, system identification and control, pattern recognition, medical diagnosis, financial applications, data mining, visualisation and e-mail spam filtering [5], [4

Sensitivity of optimal control for diffusion Hopfield ...
a Mechanical and Automation Engineering, The Chinese University of Hong Kong, .... put capacitance of the amplifier ith and its associated input lead. di > 0 are ...

Optimal control for rough differential equations
To obtain solutions to (1) requires in general regularity on the coefficients ..... W. B. Saunders Co., Philadelphia-London-Toronto, Ont., 331 pages, 1969. 13.

Optimal Hybrid Control For Structures ∗ 1 ... - Semantic Scholar
Phone: (937)455-6458. Robert E. ... Phone: (619)-822-1054. ...... Those are active controller configurations to save active control energy consumption and at the.

Optimal trajectory control for parallel single ... - Research at Google
Let xk,yk be the solution of the above optimization. Then ... the following analytic formula. [xk yk. ] .... operator to obtain the analytic formula of iL(t),vC(t), and io(t).