Local Volatility Function Approximation Using Reconstructed Radial Basis Function Networks Bo-Hyun Kim, Daewon Lee, and Jaewook Lee Department of Industrial and Management Engineering, Pohang University of Science and Technology, Pohang, Kyungbuk 790-784, Korea [email protected]

Abstract. Modelling volatility smile is very important in financial practice for pricing and hedging derivatives. In this paper, a novel learning method to approximate a local volatility function from a finite market data set is proposed. The proposed method trains a RBF network with fewer volatility data and finds an optimized network through option pricing error minimization. Numerical experiments are conducted on S&P 500 call option market data to illustrate a local volatility surface estimated by the method.

1

Introduction

Volatility is known as one of the most important market variable in financial practice as well as in financial theory, although it is not directly observable in the market. The celebrated Black-Scholes (BS) model that assumes constant volatility has been widely used to estimate volatility (often called implied volatility since it calculates the volatility by inverting the Black-Scholes formula with option price data given by the market) [3, 14]. If the assumption of BS model is reasonable, the implied volatility should be the same for all option market prices. In reality, however, it has been observed that the implied volatility shows strong dependence on strike price and time to maturity. This dependence, called the volatility smile, cannot be captured in BS model, and results in failure to produce an appropriate volatility for the corresponding option [17, 18]. One practical solution for the volatility smile is the constant implied volatility approach which is to simply use different volatilities for options with different strikes and maturities. In other words, if we had options with the whole range of strikes and maturities, we could simply calculate the implied volatility for each pair of strike and maturity by inverting the BS formula for each option. Although it works well for pricing simple European options, it cannot provide appropriate implied volatilities for pricing more complicated options such as exotic options or American options. Moreover, this approach can produce incorrect hedge factors like Gamma, Vega, Delta, etc. even for simple options [8, 12]. During the last decade, a number of researches based on different type of models have been conducted to model the volatility smile [15, 8, 1, 2]. One of the J. Wang et al. (Eds.): ISNN 2006, LNCS 3973, pp. 524–530, 2006. c Springer-Verlag Berlin Heidelberg 2006 

Local Volatility Function Approximation

525

successful type of model is to use a 1-factor continuous diffusion model. This model, describing volatility as a function of strikes (or stocks) and maturities (called a local volatility function), is turned out to be a complete model by excluding non-traded source of risks and allows for arbitrage pricing and hedging [5]. When the underlying asset follows this model, one important task is then to accurately approximate the local volatility function for accurately pricing exotic options and computing correct hedge factors [4]. Estimating a local volatility function is, however, a nontrivial task since it is generally an ill-posed problem due to insufficient market option price data. In this paper, we propose a novel learning method to approximate a local volatility function from a finite market data set in pricing and hedging derivatives. The proposed method consists of two phases. In the first phase, we train a preparatory radial basis function (RBF) network from an available volatility data to estimate the local volatility surface and then obtain the estimated volatilities for the whole range of strikes and maturities. In the second phase, we reconstruct an optimized RBF network by controlling the volatility values of a preparatory RBF network to minimize the error between estimated option prices and real market option prices. The organization of this paper is as follows. In section 2, we propose a learning method to model a volatility smile. Computational simulation applied to the S& P 500 index options are conducted in Section 3. Section 4 concludes the result.

2

The Proposed Method

The 1-factor continuous diffusion model assumes the underlying asset follows the following process with the initial value Sinit : dSt = μ(St , t)dt + σ(St , t)dWt , St

t ∈ [0, τ ],

τ >0

(1)

where τ is a fixed time horizon, Wt is a standard Brownian motion and μ(s, t), σ(s, t) : R+ × [0, τ ] → R are deterministic functions sufficiently well behaved to guarantee that (1) has a unique solution [9]. σ(s, t) is the local volatility function. When we estimate a local volatility function from a finite data set, we can avoid over-fitting problem by regularizing with some kind of smoothness of the local volatility function. A RBF network is a well-known method that is capable of solving ill-posed problems with regularization [7]. In this section, we propose a method to approximate the local volatility function using RBF networks when the underlying asset follows a 1-factor model (1). The local volatility function σ(s, t) will be explicitly represented by a reconstructed RBF network. 2.1

Phase I: Initial Local Volatility Function Approximation Using RBF Networks

In the first phase, we initially approximate a local volatility function by using RBF networks. A RBF network is constructed with a given training volatility

526

B.-H. Kim, D. Lee, and J. Lee

data set. Each training input has two attributes; strike price and time to maturity, (Kjtr , Tjtr ), j = 1, ..., l and training output has its corresponding local volatility value, σj . A RBF network involves searching for a suboptimal solution in a lower-dimensional space that approximates the interpolation solution where the approximated solution σ ˆRBF (w) can be expressed as follows: σ ˆRBF (w; K, T ) =

l  j=1

 wj φ

(K, T ) − (Kjtr , Tjtr ) η

 (2)

where each φ = φ((K, T ), (Kjtr , Tjtr )) is a radial basis function centered at (Kjtr , Tjtr ) and η is an user pre-specified scale parameter. The training procedure of the RBF network is to estimate the weights that connect the hidden and the output layers and these weights will be directly estimated by using the least squares algorithm [7, 6, 11]. To determine the optimal network weights that connect the hidden and the output layers from a training data set, {Kjtr , Tjtr , σj }lj=1 , we fit an initial local volatility surface by minimizing the following criterion function with a weight decay regularization term J(w) =

l 

σj − σ ˆRBF (w; Kjtr , Tjtr )2 + λw2 .

(3)

j=1

where λ is a regularization parameter introduced to avoid over-fitting. The network weights can then be explicitly given by w = (Φ + λI)−1 σ

(4)

where Φ = [φ((Kitr , Titr ), (Kjtr , Tjtr ))]i,j=1,...,l and σ = (σ1 , ..., σl )T . Substituting Eq. (4) into Eq. (2) makes us to rewrite σ ˆRBF (w; K, T ) as σ ˆRBF (σ; K, T ). Initially, we do not use the training volatility information σ at this stage for the estimated option prices to match the market option prices as closely as possible. Instead, we randomly generate the initial volatility vector, σ (0) ∈ l in Eq. (4), as nonnegative values. For this reason, we will call the RBF network obtained in the first phase a preparatory RBF network. 2.2

Reconstructing RBF Networks Via Pricing Error Minimization

To obtain a better local volatility function that minimizes the option pricing error, in the second phase, we reconstruct the RBF network by iteratively updating the current volatility vector σ (0) to a better one. Let pi be the i-th option market price with (Ki , Ti ) as its strike price and maturity for i = 1, ..., m. Note that the set {(Ki , Ti )}m i=1 has its corresponding option price pi and is a different set from the training input data set {(Kjtr , Tjtr )}lj=1 . Also let Pi be the i-th option pricing formula (highly nonlinear function with respect to its input) where its strike price, maturity, and volatility are given ˆ RBF , respectively. Utilizing this information, Phase II tries to find at Ki , Ti , σ

Local Volatility Function Approximation

527

an optimal volatility, σ ∗ = (σ1∗ , ..., σl∗ ), to reconstruct a final RBF network by solving the following nonlinear optimization: min E(σ) = σ

m 

ˆ RBF (σ; Ki , Ti )]} {pi − Pi [Ki , Ti , σ

2

(5)

i=1



l 

σjtr − σ ˆRBF (σ; Kjtr , Tjtr )2

j=1

where σjtr is a volatility output at a training input (Kjtr , Tjtr ) and α is an usercontrollable compensation parameter. In this paper, we will call the RBF corresponding to this optimal volatility σ ∗ as a reconstructed RBF. To get the optimal solution that minimizes Eq. (5) efficiently, we employ a trust region algorithm described as follows. For a volatility vector σ(n) at iteration n, the quadratic approximation Eˆ is defined by the first two terms of the Taylor approximation to E at σ(n); 1 ˆ E(s) = E(σ(n)) + g(n)T s + sT H(n)s 2

(6)

where g(n) is the local gradient vector and H(n) is the local Hessian matrix. A trial step s(n) is then computed by minimizing (or approximately minimizing) the trust region subproblem stated by ˆ subject to min E(s) s

s2 ≤ Δn

(7)

where Δn > 0 is a trust-region parameter. According to the agreement between predicted and actual reduction in the function E as measured by the ratio ρn =

E(σ(n)) − E(σ(n) + s(n)) , ˆ ˆ E(0) − E(s(n))

Δn is adjusted between iterations as follows: ⎧ ⎨ s(n)2 /4 if ρn < 0.25 if ρn > 0.75 and Δn = s(n)2 Δn+1 = 2Δn ⎩ Δn otherwise The decision to accept the step is then given by  σ(n) + s(n) if ρn ≥ 0 σ(n + 1) = σ(n) otherwise

(8)

(9)

(10)

which means that the current weight vector is updated to be σ(n) + s(n) if E(σ(n) + s(n)) < E(σ(n)); Otherwise, it remains unchanged and the trust region parameter Δn is shrunk and the trial step computation is repeated. One nice property of the trust region method is that this method retains the rapid rate of convergence of Newton’s method, but is also generally applicable and globally convergent [10, 13, 16].

528

3

B.-H. Kim, D. Lee, and J. Lee

Computational Examples

We use the S&P 500 Index European call option data of October 1995, which is also used in [1, 4]. The market option price data is given in Table 1. Only the options with no more than two years maturity are used for accuracy. The initial index, interest rate, and dividend rate are set as follows: Sinit = $590, r = 0.06, q = 0.0262 In Phase I, we used the following 16 training volatility input data to train a preparatory RBF network: (Kjtr , Tjtr ) = (K(p), T (q)), j = 4(p − 1) + q, p, q = 1, ..., 4 K = [0.8000Sinit , 0.9320Sinit, 1.0640Sinit , 1.3940Sinit ] T = [0, 0.66, 1.32, 1.98] In Phase II, we used 70 S&P 500 Index option price data given in Table 1. Table 1. S&P 500 Index call option price Maturity .175 .425 .695 .94 1 1.5 2

85% 9.13e1 9.63e1 1.02e2 1.07e2 1.08e2 1.17e2 1.26e2

90% 6.28e1 6.91e1 7.61e1 8.22e1 8.36e1 9.44e1 1.04e1

95% 3.52e1 4.40e1 5.26e1 5.99e1 6.16e1 7.31e1 8.36e1

Strike(% of spot price) 100% 105% 110% 115% 1.29e1 2.11 1.21e−1 3.73e−2 2.33e1 8.54 2.26 4.21e−1 1 1 3.26e 1.64e 5.95 1.90 3.99e1 2.38e1 1.13e1 4.71 4.16e1 2.54e1 1.28e1 5.50 5.40e1 3.73e1 2.37e1 1.43e1 6.49e1 4.82e1 3.42e1 2.36e1

120% 1.62e−2 1.94e−1 6.04e−1 1.78 2.13 7.65 1.47e1

130% 1.65e−3 2.77e−2 7.25e−2 1.82e−1 2.27e−1 1.85 5.65

140% 5.14e−4 8.72e−3 2.66e−2 4.48e−2 5.44e−2 3.10e−1 1.78

The proposed method is compared with a popularly used natural cubic spline method (cf. [4]). The spline approach normally uses the same amount of training data (with a different set of strike prices and maturities) for its knot points as that of option market data to get a reasonable performance. The spline function is then reconstructed from the 70 option market data by minimizing option pricing error with respect to 70 volatility input variables. The criteria for comparison are pricing error (MSE; mean squared error) and hedging error between predicted volatility and true implied volatility. Table 2 demonstrates the feature of our proposed method: that is, the proposed method achieved a better or comparable performance compared to spline approach in pricing and hedging derivatives with fewer input variables (16 in our method and 70 in spline approach). Fig. 1 shows the estimated volatility surfaces generated by the spline and the proposed. The proposed method showed more skewed volatility smile effects as observed in practice.

Local Volatility Function Approximation

529

Table 2. Accuracy of Pricing and Hedging

Pricing Hedging

(a)

Measure Reconstructed Spline Reconstructed RBF Price 0.0317 0.0284 Vega 7.1153 5.8509 Delta 1.1535e−5 1.0255e−5 −8 Gamma 1.3316e 1.0839e−8 Rho 1.5961 1.6449 Theta 0.2935 0.3129

(b)

Fig. 1. Estimated volatility surfaces. (a) spline approach and (b) proposed method.

4

Conclusion

In this paper, we’ve proposed a novel learning method to approximate the local volatility function. The proposed method first trains a preparatory RBF network with a training volatility data set to estimate volatility surface and then reconstructs an optimized RBF network by minimizing the errors between estimated prices and real market prices. A simulation has been conducted with a S&P500 option data example. The experimental results demonstrated a performance improvement of the proposed method compared to other approach in terms of pricing and hedging error. Acknowledgement. This work was supported by the Korea Research Foundation under grant number KRF-2004-041-D00785.

References 1. Andersen, L.B.G., Brotherton-Ratcliffe, R.: The Equity Option Volatility Smile: An Implicit Finite Difference Approach. Journal of Computational Finance 1(2) (1997) 2. Avellaneda, M., Friedman, M., Holmes, C., Samperi, D.: Calibrating Volatility Surfaces via Relative Entropy Minimazation. Applied Mathematical Finances 4 (1997) 37-64

530

B.-H. Kim, D. Lee, and J. Lee

3. Black, F., Scholes, M.: The Pricing of Options and Corporate Liabilities. Journal of Political Economy 81 (1973) 637-659 4. Coleman, T.F., Li, Y., Verma, A.: Reconstructing the Unknown Local Volatility Function. The Journal of Computational Finance 2(3) (1999) 5. Dupire, B.: Pricing with a Smile. Risk 7 (1994) 18-20 6. Han, G.S., Lee, D., Lee J. : Estimating The Yield Curve Using Calibrated Radial Basis Function Networks. Lecture Notes in Computer Science 3497 (2005) 885-890 7. Haykin, S.: Neural Networks: a Comprehensive Foundation. Prentice-Hall New York (1999) 8. Hull, J., White, A.: The Pricing of Options on Assets with Stochastic Volailities. Journal of Finance 3 (1987) 281-300 9. Lamberton, D., Lapeyre, B.: Introduction to Stochastic Calculus Applied to Finance. Chapman & Hall (1996) 10. Lee, J.: Attractor-Based Trust-Region Algorithm for Efficient Training of Multilayer Perceptrons. Electronics Letters 39 (2003) 71-72 11. Lee, D., Lee, J.: A Novel Three-Phase Algorithm for RBF Neural Network Center Selection. Lecture Notes in Computer Science 3173 (2005) 350-355 12. Lee, H.S., Lee, J., Yoon, Y.G., Kim, S.: Coherent Risk Meausure Using Feedfoward Neural Networks. Lecture Notes in Computer Science 3497 (2005) 904-909 13. Lee, J., Chiang, H.D.: A Dynamical Trajectory-Based Methodology for Systematically Computing Multiple Optimal Solutions of General Nonlinear Programming Problems. IEEE Trans. on Automatic Control 49(6) (2004) 888 - 899 14. Merton, R.: The Theory of Rational Option Pricing. Bell Journal of Economics and Managemenet Science 4 (1973) 141-183 15. Merton, R.: Option Pricing when Underlying Stock Returns Are Discontinuous. Journal of Financial Economics 3 (1976) 124-144 16. Nocedal, J., Wright, S.J.: Numerical Optimization. Springer New York (1999) 17. Rubinstein, M.: Implied Binomial Trees. The Journal of Finance 49 (1994) 771-818 18. Shimko,D.: Bounds of Probability. Risk (1993) 33-37

LNCS 3973 - Local Volatility Function Approximation ... - Springer Link

S&P 500 call option market data to illustrate a local volatility surface estimated ... One practical solution for the volatility smile is the constant implied volatility approach .... Eq. (4) into Eq. (2) makes us to rewrite ˆσRBF (w; K, T) as ˆσRBF (σ; K, T).

397KB Sizes 0 Downloads 258 Views

Recommend Documents

LNCS 6622 - NILS: A Neutrality-Based Iterated Local ... - Springer Link
a new configuration that yields the best possible fitness value. Given that the .... The neutral degree of a given solution is the number of neutral solutions in its ...

LNCS 6361 - Automatic Segmentation and ... - Springer Link
School of Eng. and Computer Science, Hebrew University of Jerusalem, Israel. 2 ... OPG boundary surface distance error of 0.73mm and mean volume over- ... components classification methods are based on learning the grey-level range.

LNCS 6621 - GP-Based Electricity Price Forecasting - Springer Link
learning set used in [14,9] at the beginning of the simulation and then we leave ..... 1 hour on a single core notebook (2 GHz), with 2GB RAM; the variable ...

LNCS 6683 - On the Neutrality of Flowshop Scheduling ... - Springer Link
Scheduling problems form one of the most important class of combinatorial op- .... Illustration of the insertion neighborhood operator for the FSP. The job located.

LNCS 4261 - Image Annotations Based on Semi ... - Springer Link
MOE-Microsoft Key Laboratory of Multimedia Computing and Communication ..... of possible research include the use of captions in the World Wide Web. ... the Seventeenth International Conference on Machine Learning, 2000, 1103~1110.

LNCS 7335 - Usage Pattern-Based Prefetching: Quick ... - Springer Link
Oct 8, 2010 - The proposed scheme is implemented on both Android 2.2 and Linux kernel 2.6.29. ... solution for reducing page faults [1, 2, 3, 10]. The number ...

LNCS 4191 - Registration of Microscopic Iris Image ... - Springer Link
Casey Eye Institute, Oregon Health and Science University, USA. {xubosong .... sity variance in element m, and I is the identity matrix. This is equivalent to.

LNCS 6719 - Multiple People Activity Recognition ... - Springer Link
Keywords: Multiple Hypothesis Tracking, Dynamic Bayesian Network, .... shared space and doing some basic activities such as answering phone, using.

LNCS 3174 - Multi-stage Neural Networks for Channel ... - Springer Link
H.-S. Lee, D.-W. Lee, and J. Lee. In this paper, we propose a novel multi-stage algorithm to find a conflict-free frequency assignment with the minimum number of total frequencies. In the first stage, a good initial assignment is found by using a so-

LNCS 3352 - On Session Identifiers in Provably Secure ... - Springer Link
a responding server process in network service protocol architecture [23]. A ... 3. proposal of an improved 3PKD protocol with a proof of security using a.

LNCS 4731 - On the Power of Impersonation Attacks - Springer Link
security or cryptography, in particular for peep-to-peer and sensor networks [4,5]. ... entity capable of injecting messages with arbitrary content into the network.

LNCS 4325 - An Integrated Self-deployment and ... - Springer Link
The VFSD is run only by NR-nodes at the beginning of the iteration. Through the VFSD ..... This mutual effect leads to Ni's unpredictable migration itinerary. Node Ni stops moving ... An illustration of how the ZONER works. The execution of the ...

LNCS 650 - Emergence of Complexity in Financial ... - Springer Link
We start with the network of boards and directors, a complex network in finance which is also a social ... from the board of Chase Manhattan Bank. Boards of ...

LNCS 4233 - Fast Learning for Statistical Face Detection - Springer Link
Department of Computer Science and Engineering, Shanghai Jiao Tong University,. 1954 Hua Shan Road, Shanghai ... SNoW (sparse network of winnows) face detection system by Yang et al. [20] is a sparse network of linear ..... International Journal of C

LNCS 6621 - GP-Based Electricity Price Forecasting - Springer Link
real-world dataset by means of a number of different methods, each cal- .... one, that we call GP-baseline, in which the terminal set consists of the same variables ...

LNCS 4261 - Image Annotations Based on Semi ... - Springer Link
Keywords: image annotation, semi-supervised clustering, soft constraints, semantic distance. 1 Introduction ..... Toronto, Canada: ACM Press, 2003. 119~126P ...

LNCS 7601 - Optimal Medial Surface Generation for ... - Springer Link
parenchyma of organs, and their internal vascular system, powerful sources of ... but the ridges of the distance map have show superior power to identify medial.

LNCS 4258 - Privacy for Public Transportation - Springer Link
Public transportation ticketing systems must be able to handle large volumes ... achieved in which systems may be designed to permit gathering of useful business ... higher powered embedded computing devices (HPDs), such as cell phones or ... embedde

LNCS 2747 - A Completeness Property of Wilke's Tree ... - Springer Link
Turku Center for Computer Science. Lemminkäisenkatu 14 ... The syntactic tree algebra congruence relation of a tree language is defined in a natural way (see ...

LNCS 7575 - Multi-component Models for Object ... - Springer Link
visual clusters from the data that are tight in appearance and configura- tion spaces .... Finally, a non-maximum suppression is applied to generate final detection ...

LNCS 6942 - On the Configuration-LP for Scheduling ... - Springer Link
insights on two key weaknesses of the configuration-LP. For the objective of maximizing the minimum machine load in the unrelated graph balancing setting ...... length. European Journal of Operational Research 156, 261–266 (2004). 19. Scheithauer,