Swaption Pricing under Libor Market Model Using Monte-Carlo Method with Simulated Annealing Optimization

Abstract

The thesis seeks to use simulated annealing optimization to minimize the difference between the value of the libor model volatility and the ones quoted in the market for congruent pricing of a Swaption contract. The simulated annealing optimization technique, being a global minimisation method, would provide accurate parameters that will simulate libor rates that are harmonious with the observed yield curve. This latter feature employed in a Monte-Carlo pricing method would price the Swaption contract fundamentally closer to its market value than other local optimization methods. The SA method starts from an initial point, often random, then searches the neighbourhood of the current solution for the next point. The neighbourhood function search is in accordance with the set probabilistic distribution that will determine the distance between the two solutions. Each solution has a cost value associated with it. The cost function determines the eligibility of the solution by measuring its discrepancy with the set limit. If the discrepancy is larger than the set limit, a new solution is sought. If the discrepancy is still large, the old and new cost value is compared, and the latter is accepted if it’s less than the former or otherwise rejected with a certain probability that is largely dependent on the control mechanism. The method terminates if the cost value attained is equal to the set tolerance level. Different from other heuristic methods that solely base their solution on the iterative improvement of the solution’s cost value, simulated annealing accepts some inferior solutions so as to have a wider search in the design space. The main advantage of the method is the ability to escape local minimum entrapment through the aforementioned acceptance/rejection criteria. The results indicate that the advantageous aspects of the Simulated annealing enable it to outperform the least square non-linear optimization method commonly used in the simulation.

Share and Cite:

Ondieki, K. (2022) Swaption Pricing under Libor Market Model Using Monte-Carlo Method with Simulated Annealing Optimization. Journal of Mathematical Finance, 12, 435-462. doi: 10.4236/jmf.2022.122024.

1. Background of the Study

Swaptions are options that give the holder the right but not the obligation to enter into a swap contract on a future date at a pre-agreed strike rate. A swap contract is an agreement between two parties to exchange a floating and fixed interest rate payment for a tenor. The interest rate payments are based on the same notional amount which is not exchanged. The exchange involves one party paying the other the net value of the two payments on scheduled dates. In the option, the holder who agrees to pay the fixed rate in the swap and receive the floating rate is the payer swaption, whereas the holder who agrees to pay the floating rate and receives the fixed is the receiver swaption. A payer swaption would exercise the option if the swap rate is higher than the option strike and the receiver swaption would exercise if the latter is higher.

There are various reasons to enter into a swaption contract, with the most common one being to transform the nature of an asset’s interest rate payments. By buying a receiver swap, the holder can transform the interest rate applicable to an asset from a floating to a fixed rate, hence hedging against interest rate risk. Since the contract is an option, it is expected to be exercised only if the payoff is positive. The other prominent reason would be to speculate against the interest rate cycle. If a firm was to expect a fall in the interest rate in the near future, they could enter into a receiver swaption to take advantage of the high fixed rate as the floating rate plumber.

There has been a significant increase in the trading of interest rate swaps over the past half-decade, which has led to the expansion of its option derivative.

Recent surveys indicate that the market turnover for swap options as of 2016 was $163.021 trillion, and the figure went up to ([wooldridge2019fx:01]). In March 2021 particularly, [1] reports that USD swaptions volume hit an all-time high i.e. 6346, which was on the back of a record month in February 2021, when 5973 trades were reported. The volume was driven by a large sell-off in fixed-income markets. This substantiates the importance of proper hedging and pricing formulae in order to value the contracts accurately. Good pricing performance is a necessary condition for a useful model; good hedging performance, however, is sufficient [2] .

Markets always quote swaptions volatilities in a periodic series of either 3, 6, or 12 months, hence a more direct pricing strategy would be to model applicable libor rates with same time discretization as opposed to modelling the instantaneous rate and then prorating the forward rates applicable over the periods. This makes the market models like the libor process more empirically suitable to model the forward rates applicable in swaptions.

Market models have become popular in pricing basic interest rate derivatives namely caps and swaptions because of their agreement with well-established market formulas i.e. Libor-forward rates (LFR) prices caps with Black’s formula whereas libor swap rates (LSM) prices swaptions with same the formula. However, LFR and LSM are not compatible with each other in theory, but empirically their distribution is not far off each other. Nevertheless, we shall adopt the LFR model to price swaption because they are more natural representative coordinates of the yield curve than the swap rates. Additively, it is natural to express the LSM in terms of a suitable preselect family of LFR rather than doing the converse [3] .

The pricing of swaptions hinges on the LFR process with parameters that are optimized in accordance with the market volatilities of existing interest rate products. Through the transformation of Black’s formula, traders extract implied volatilities from tradable assets, which are then used to derive LFR’s dynamics for simulation.

Optimization provides an excellent method to select the best element in terms of some system performance criteria from some set of available alternatives. On the other hand, a simulation is a tool that allows us to build a representation of a complex system in order to better understand the uncertainty in the system’s performance. Often the emphasis is put on simulation, leaving optimization techniques at the discretion of the trader. Sometimes the optimization technique used leads to large errors that significantly affect the simulation and hence the model performance. When considered separately, each method is important, but limited in scope. By giving the two equal weights, we can develop a powerful framework that takes advantage of each method’s strengths, so that we have at our disposal a technique that allows us to select the best element from a set of alternatives and simultaneously take account of the uncertainty in the system.

Swaption pricing has often adopted the least-square non-linear (lsqnonlin) method to minimize the error between the model parameters and the market measures. The method is computationally easy to use but it can get trapped in the local minimum, hence greatly overestimating the parameters. The method’s major drawback is that it doesn’t have any known mechanism to get out of local minimum entrapment. Local minimum parameters would have much higher errors than global minimums if the two are not one and the same.

Simulated annealing (SA) optimization searches the global minimum parameters by utilizing a probabilistic transition rule that determines the criteria for moving from one feasible solution to another in the design space. Sometimes the rule accepts inferior solutions in order to escape from local minimum entrapment. This enables it to search higher dimensions for other minimum points, and in the process converge to the global minimum. The method also doesn’t require any mathematical model making it suitable for problems that do not have an exact distribution [4] .

2. Simulated Annealing Overview

Simulated annealing optimization imitates the annealing process used in metallurgy whereby a substance is heated to its melting point and then it is slowly cooled in a controlled manner until it solidifies. The process largely depends on the cooling schedule to determine the structural integrity of the resultant substance. If the cooling is too quick, the substance forms an irregular crystalline lattice hence making it weak and brittle. If the cooling is slow the substance formed is strong since the crystal lattice is regular.

SA establishes the link between the thermal cooling behaviour and searches for the global minimum. The resultant substance with a regular crystalline lattice represents a codified solution to the problem statement and the cooling schedule represents how and when a new solution is to be generated and incorporated. The technique has basically three steps; if the old solution doesn’t meet the set requirements, perturb it to a new one, then evaluate the new solution given the old one, and finally accept or reject the new solution. The analogy of the cooling schedule could be represented using the picture below (Figure 1) [4] .

The objective is to get the ball to the lowest point of the valley. At the beginning of the the process the temperature is high and strong perturbations are exerted on the box and hence the ball can jump through the high peaks in search for the bottom of the valley. As the time goes by and the temperature reduces, the perturbations become weak hence the ball can only jump small peaks. At the time the temperature reduces to very low levels, there is a high probability that the ball will be in the lowest depression of the valley. In the algorithm, the temperature forms part of the controlling mechanism for accepting a new solution i.e. When the the temperature is high, the SA optimization is searching for the global solution in a broad region, but as the temperature reduces the search

Figure 1. SA Cooling analogy.

radius reduces hence refining the feasible solution attained at high temperatures.

The algorithm uses the solution error to evaluate the eligibility of the new solution. The solution error is calculated from the objective function to be optimized. Usually each solution X i is associated with its error E i . The error is used to evaluate the new solution by determining its acceptance or rejection. If the error in the new solution is less than the old, it is automatically accepted, on the other hand if it is more it accepted but under a set probability ( P n ) i.e.

P n = ( exp k Δ E T Δ E 0 1 Δ E 0,

where Δ E is the change in the solution error after it has been perturbed, T is the current temperature and K is a suitable constant. Through the acceptance/rejection criteria above, the algorithm can accept inferior solutions but the probability reduces as the temperature reduces or as the Δ E increases. Consequently, at high temperatures the algorithm searches a broad area hence accepts bad solutions, as the temperature reduces, the algorithm is more selective and only accepts solutions when the Δ E is very small. The algorithm termination criteria is user defined and mainly include error tolerance level, maximum number of iteration, or attained temperature level.

The algorithm components include; a neighborhood function that conducts the perturbations, a cooling function that dictates how the temperature reduces and an acceptance function for evaluation of solutions.

SA versatility can be attributed to the feedback generated by its users which has in turn been used to add various options that go beyond the basic algorithm. The power & flexibility accorded by the feedback mechanism has enabled it to be applied across many platforms. In cases where the default options are not applicable, it accepts customized functions. The introduction of re-annealing also permits adaptation to changing sensitivities in the multidimensional parameter-space [5] .

3. Literature Review

Before the libor market model was derived, the difficulty in valuation of caps and swaptions was attributed to presence of unobservable financial quantities like instantaneous forward rate or short rate, in the pricing formulae. This holds major disadvantage in that a transformation via black box in the model is needed in order to map the dynamics of this unobservable quantities to observable ones [6] .

The developments that led the derivation of the Libor process was reported in [7] where it was shown how to choose the volatility functions (and a change of measure), so that libor rates follow lognormal processes. Although it is an extension of Heath,Jarrow & Morton model (HJM), the libor process differs in that the process has observable market quantities and produces well behaved forward rates than HJM’s which can go to infinity in finite time [8] . The hallmark of the model was anchored upon having an arbitrage free process that can accommodate correlated forward rates.

The libor also holds an advantage over the short rate process in that; it can achieve decorrelation of forward rates, the estimated parameters can be interpreted, and it models the forward rate as primary process as opposed to a secondary one. Decorrelation is achieved by finding the most effective way to redistribute the volatility of the libor rates as time lapses [3] .

[9] considered the various implementation methodologies of pricing caps and swaptions in the libor framework. The paper highlights that Monte-Carlo method with parametric correlation matrix has a more stable evolution of volatility and correlation, which is a desired feature in pricing exotic options and also hedging. It also states that non-parametric approach poses minimization problems as the number of free parameters becomes large, hence impossible to estimate since the number of forward rates alive may not be sufficient.

In the libor framework, once time dependent instantaneous volatility and correlation of the forward rates have been specified, their stochastic evolution is completely determined. This leaves the matching of Black’s volatility (path independent) to the integral of instantaneous volatility (path-dependent) the only approximation needed. [10] indicates that by using a self financing strategy between the swaption and the libor rates and assuming that both have lognormal distribution and deterministic volatilities, the Black’s volatility can be approximated by a linear function of swap rates weights, correlation coefficient and instantaneous volatility.

In solving the need for the correlation matrix to have a rank less than the number of factors considered, [6] proposed a discerning parameterization method that uses a hyper-sphere decomposition before dimensionality problem is addressed. However the correlation matrix will depend on the bounds set for the angles. Use of spherical coordinates in the specification of instantaneous volatility allows for a more robust optimization scheme [11] . This methodology will be applied in this thesis.

The basic SA algorithm that was originally published by [12] had two main implementation strategies for the acceptance probability. The primary one was Boltzmann annealing which was credited to Monte Carlo importance-sampling technique for handling large-dimensional path integrals arising in statistical physics problems. The method was generalised so as to fit non-convex cost functions arising in various fields. The acceptance probability is based on the chance of obtaining a new solution error relative to the previous one. The secondary one was fast annealing, which adapted the Cauchy distribution for its acceptance probability as opposed to Boltzmann distribution. In comparison with the Gaussian Boltzmann form, the fast Cauchy distribution has fatter tails hence permits an easier access to test local solutions while searching for the global one.

In the [13] paper, “Theory and practice of Simulated annealing”, they enlist topography and size of the design space as the key variables that need to be considered while choosing the sampling function. A sampling function that imposes smooth topography in cases where the local minima are shallow is preferred to a bumpy one. On the issue of size, the main consideration is the ability of the function to reach other feasible solution in finite number of iteration i.e necessity of reachability. The paper also offer other conjures by hinting that sampling a small size of solution space is preferred as compared to large one because the latter will have the algorithm sampling large portions hence unable to concentrate on specific areas of the design space. Contrary to the suggestion above, a larger annealing sample guarantees a better SA performance than a small one. They also suggest a method of reducing the search space by isolating the strongly persistent variables1 during SA execution. Ultimately, the function will highly depend on the problem at hand.

Cooling schedules commonly used in SA was originally proposed by [14] . It includes an initial temperature T 0 ; which must be too high so that the new solution sought is accepted with probability close to 1, a temperature decreasing function; generally an exponential function with a parameter α , and the number of iterations k for each temperature level before it reduces. [15] describes three temperature reduction function commonly used in empirical application as.

1) Multiplicative Monotonic Cooling—Temperature is reduced by multiplying it with a parameter α . α generally varies from 0.8 - 0.9.

2) Non-Monotonic Adaptive Cooling—Temperature is reduced my multiplying it with an adaptive factor that is based on the difference between the current solution objective.

3) Additive Monotonic Cooling—Two additional parameters are included namely; the number n of cooling cycles, and the final temperature T n of the system. In this type of cooling, the system temperature T at cycle k is computed adding to the final temperature T n , a term that decreases with respect to cycle k.

Each has four variants namely quadratic, linear, exponential and logarithmic. Non-Monotonic Adaptive Cooling has also a trigonometric version.

[16] documents the importance of choosing a good initial temperature for the cooling schedule as it determines the ability of the algorithm to find good solutions. In the paper, he gives four accounts of calculating the initial temperature. The first is equating it to the largest cost value difference between any two solutions in the design space. The second equates it to a function of a constant and the volatility of the cost value differences i.e. k σ 2 . Temperature is taken to be at infinity and K is between 5 and 10. The third is coined from the relationship between temperature parameter and the acceptance ratio i.e. starting from a given acceptance ratio χ o , a large enough temperature is first tested and the ratio is calculated. If the ratio is less than the χ o , the temperature is multiplied by two, if larger, it is divided by three. He gives the latter method a thumbs up as it is able to avoid cycles and a good estimation of the initial temperature is easily found. The fourth borrows slightly from the former in that the initial temperature is the quotient of the average cost value changes and the logarithm of the set initial acceptance ratio χ o . The method starts by first generating some random transitions so as to be able to compute the average change of the cost value.

[17] studied the application of SA in stochastic & deterministic optimization whilst using constant and varying sample size. The constant/variable sample is incorporated in the neighbourhood function in that the next solution is evaluated as an average of chosen samples size. He motivates the importance of choosing a good cooling schedule for the temperature parameter in evaluating the neighbourhood solution so as save on computational time and increase efficiency. In his conclusion remarks, the varying sample method performs better for deterministic problems, whereas constant sample performs better for complex stochastic problems.

[18] whilst strived to maximize the expected logarithm utility function, investigated the use of simulated annealing in optimization of wealth allocation of stocks for capital growth problem. The paper adapted a Cauchy function to move from one feasible solution to another. In order to have a positive capital growth, the wealth allocations of the stocks have to be a below a critical point. Since the number of stocks available is numerous, the strategy heavily relies on the weight put on each stock. Any local optimization method will overestimate the parameters consequently affecting the outcome greatly. They showed that by using Cauchy sampling function and decreasing cooling schedule the parameters could escape the local maxima and achieve global maximum parameters with significant accuracy to warrant a positive capital growth.

[19] examined the problem of selecting optimal sparse mean reverting parameters based on observed and generated time series. They adapted SA method and compared it with the greedy method. The paper reports that the SA methods outperform the greedy method in 10% of the cases where the asset’s cardinality is small. The percentage increases to 25 when the assets and cardinality restriction are doubled, indicating that the SA method becomes more attractive for larger assets whilst maintaining asymptotic runtime of simpler heuristics.

[20] compares the various optimization techniques applied in minimization of volatility function in pricing of caps & swaptions in Cheyette model. In his paper he acknowledges the superiority of non-derivative algorithms like downhill simplex or Genetic algorithm in implementation of the model. In his execution, he embedded the downhill simplex method in the SA algorithm. Beyna results indicate lower free parameters than calculated by Downhill simplex method. The paper also tested several cooling annealing schedules i.e. linear, exponential cooling schemes and some adaptive ones. It turned out, that the linear cooling scheme delivers the best results, if the cooling factor is chosen considerably small.

4. Methodology

The libor2 process models forward rates which are in turn used to evaluate the expected payoff of swaptions under some measures. The process assumes that the rates follow log-normal distribution so as to be able to recover the [21] formula. In coming up with the formula, the thesis employs Martingale representation method.

4.1. Martingale Representation Method

Martingale is a process that has the property that at each point in time, the conditional distribution for the future point in time is centred around the mean. This constraint on the process stops it from becoming too wild [8] . Martingale representation theorem allows the payoff of interest rate derivatives to be evaluated using an expectation of a stochastic martingale process [22] . Taking the price of an option evaluated with respect to numeraire N or any traded asset with its corresponding probability measure , to be C t at timet and the payoff at time T to be X T , its martingale representation can evaluated as:

C t N t = E [ X T N T | F t ] , (1)

where:

X T = max ( R T K , 0 ) .

4.2. Change of Numeraire

Many computational applications of derivatives pricing models such as determination of derivative prices by simulation or the estimation of derivative pricing models can be significantly simplified by a change of numeraire [8] . The numeraire is chosen so as to simplify the equationi.e. when N T is stochastic it can’t be factored out, but a change of numeraire can have it change to unit hence easily simplifying the equation3. Radon-Nikodym derivative allows one to make the switch from expectation under one measure to another measure e.g. from T-forward measure to forward swap measure.

4.3. Swaps

Interest rate Swap is an agreement to exchange payments using two different indexes i.e floating and fixed. The party that agrees to pay the fixed and receive the floating index is the Payer swap and the converse is the receiver swap. The fixed index pays a fixed amount K at every instance T j while the floating side pays L j ( T j 1 ) . Different from the fixed rate, the floating have a reset date T j that stipulates the forward rate applicable at the time of payments T j + 1 , for the tenor T o T β , hence the notation L j ( T j 1 ) .

Thus, the payoff at time t < T 0 of such a payer swap contract denoted as IRS ( t 0 , T 0 , T n , K ) can be written as4:

IRS ( t 0 , T 0 , T n , K ) = E j = 0 β 1 D ( t , T j ) τ j ( L j ( T j 1 ) K ) , (2)

where τ j is the year fraction between T j T j + 1 , D ( t , T j ) is the discount factor, 0 is the commencement date of the contract and β is the expiry.

The contract can be evaluated (at time t) under the T j + 1 forward measure by using the price of the bond maturing at time T j + 1 as the numeraire, i.e.

IRS ( t 0 , T 0 , T n , K ) = τ j E T J + 1 j = 0 β 1 P ( t , T j + 1 ) P ( T j + 1 , T j + 1 ) ( L j ( T j 1 ) K ) . (3)

By assuming that L j ( t ) is a martingale under the T j + 1 forward measure, we can express the value of the payer IRS as:

IRS ( t 0 , T 0 , T n , K ) = τ j j = 0 β 1 P ( t , T j ) τ j ( L j ( t ) K ) .

The fixed rate that makes the IRS a fair contract at time t, denoted as R 0, β ( t ) , should be calculated with the condition IRS ( t 0 , T 0 , T n , K ) = 0 . By setting the expression of the payer IRS equal to zero, the forward swap rate R 0, β ( t ) may be expressed as:

R 0 , β = j = 0 β 1 W j ( t ) L j ( t ) , (4)

where

W j ( t ) = P ( t , T j ) j = 0 β 1 P ( t , T j ) . (5)

Equation (4) means that the the forward swap rate is a weighted average of the of the libor rates over the tenor since W j is bounded as 0 < W j < 1 . Through the libor-bond relation; (derivation shown in Equation (33) in the appendix),

L j ( t ) = 1 τ j ( P ( t , T j ) P ( t , T j + 1 ) 1 ) , (6)

the payer IRS can be simplified to;

IRS ( t 0 , T 0 , T n , K ) = j = 0 β 1 ( P ( t , T j ) P ( t , T j + 1 ) ) + P ( t , T β ) τ j K j = 0 β 1 P ( t , T j + 1 ) P ( t , T β ) (7)

= P ( t , T 0 ) τ j K j = 0 β 1 P ( t , T j + 1 ) P ( t , T β ) , (8)

which depicts the well known feature of IRS pricing that the value doesn’t depend on the volatility or the correlations of the underlying forward rates. Applying the fair contract condition on Equation (8), R 0, β becomes,

R 0 , β = P ( t , T 0 ) P ( t , T β ) τ j j = 0 β 1 P ( t , T j + 1 ) , (9)

rearranging Equation (9) yields,

P ( t , T 0 ) P ( t , T β ) = τ j R 0 , β j = 0 β 1 P ( t , T j + 1 ) . (10)

Using Equation (10) in Equation (8), the payoff of the IRS becomes;

IRS ( t 0 , T 0 , T n , K ) = τ j j = 0 β 1 P ( t , T j ) ( R 0 , β K ) . (11)

By using simple algebraic manipulation and dividing Equation (9) by P ( t , T 0 ) , the forward swap rate can be expressed in terms of libor rates as ( [8] );

R 0 , β = 1 j = 0 β 1 1 1 + τ j L m ( t ) m = 0 β 1 τ j j = 0 j 1 1 + τ j L j ( t ) .

4.4. Pricing Swaption under Black’s Framework

A swaption contract gives the holder the right but not the obligation to enter into a swap contract at a future date, which is the swaption maturity. Usually the first reset date of the swap is often the maturity date of the swaption. Since the contract is valued fairly, we express its payoff as a call option on the forward swap rate i.e.

S i r s ( t 0 , T 0 , T n , K ) = τ j j = 0 β 1 P ( t , T j ) τ j ( R 0 , β ( t ) K ) + . (12)

Simply put, Equation (12) implies the payoff of the swaption can be deemed as the product of an option on the forward swap rate and an annuity C t 0 , β :

τ i = o + 1 β P ( t , T i ) = C t o , β . (13)

To recover the Black’s formula for swaptions, the expectation of the payoff at time t has to be taken under forward swap measure. Taking the annuity as the new numeraire, the value of the option becomes;

S i r s ( t ) C t o , β ( t ) = E s ( S i r s ( T 0 ) C t , β ( T 0 ) ) = E s ( ( R 0 , β ( T 0 ) K ) + C t , β ( T 0 ) ) . (14)

The new measure helps to simplify the equation since the the numerator under the expectation operator is a unit metric, hence changing Equation (14) to;

S i r s = C t o , β ( t ) E s ( R 0 , β K ) + .

Using the relationships Equation (4), Equation (45), and applying Ito’s lemma, the risk neutral dynamics of R 0, β becomes [8];

d R 0 , β = j = 0 β 1 d R 0 , β ( t ) d L j ( t ) γ j ( t ) L j ( t ) ( d W ^ ( t ) j = 0 β 1 W m ( t ) ψ m + 1 ( t ) d t ) . (15)

Defining the new wiener process under forward measure as;

d W s ( t ) = d W ^ ( t ) j = 0 β 1 W m ( t ) ψ m + 1 ( t ) d t ,

and an application of Radon-Nikodym derivative [8];

d p s d p ^ = C α , β ( t ) / C α , β ( 0 ) A ( t ) / A ( 0 ) ,

expression Equation (15) becomes;

d R 0 , β = j = o + 1 β d R 0 , β d L j ( t ) γ j ( t ) L j ( t ) d W s ( t ) .

Different from other SDEs, the swap rate doesn’t automatically form log-normal dynamics when under the forward swap measure, therefore an approximation method is needed to arrive at it.

Assuming for all ν t ;

d R 0 , β = R 0 , β ( ν ) j = o + 1 β d R 0 , β ( ν ) d L j d L j ( ν ) d R 0 , β ( ν ) γ j ( ν ) d W s ( ν ) , R 0 , β ( ν ) j = o + 1 β d R 0 , β ( t ) d L j d L j ( t ) d R 0 , β ( t ) γ j ( ν ) d W s ( ν ) = R 0 , β ( ν ) j = o + 1 β ϖ j γ j ( ν ) d W s ( ν ) ,

where ϖ j expressed as,

ϖ j = d R 0 , β ( t ) d L j d L j ( t ) d R 0 , β ( t ) .

This approximate process was obtained by use of the frozen coefficient technique that allows the relaxation the dependence of state variables to the model coefficient.

Applying the approximated log dynamics of R 0, β in (14), the Black’s formula for swaptions becomes;

C t , β ( t ) [ R 0, β ( d 1 ) K ( d 2 ) ] , (16)

where,

d 1 ( t , T ) = ln ( R 0 , β K ) + 1 2 σ n 2 ( T 0 t ) σ n T 0 t ,

d 2 = d 1 ( t , T ) σ n T 0 t ,

and σ n is the variances of the forward swap rate R 0, β computed as:

σ n 2 = 1 T 0 t t T 0 ϖ k ϖ j γ k ( s ) γ j ( s ) d s .

Calibration

Swaptions are quoted as implied volatilities as opposed to dollar amount. The implied volatilities are a descriptive of the level of prices of the underlyingi.e. the swap. In the Black’s equation, σ (implied volatility) is an input. Plugging in the implied volatility in the equation will give the level of price that is acceptable in the market for each pair of forward swap rate and strike.

4.5. Numerical Pricing of Swaptions Using the Libor-Forward Rates

In contrast to analytical solutions that price swaptions only under log-normal forward swap rates, numerical methods are more flexible to accommodate other distributions and rates like the forward libor rates. This helps to take advantages of the latter congruence to yield curve.

In pricing swaptions under the LFR, unlike in caps, the payoff is not additively separable with respect to different rates. As a result, the expectation of such a payoff includes the joint distribution of the spanning forward rates in the calculation. This means that the correlation between the rates has an impact on the value of the contract. The solution to the above issue is to assign a different Brownian motion to each forward rate and then assume the Brownian motion to be instantaneously correlated. Manipulating the instantaneous correlation leads to manipulation of correlation of simple rates i.e terminal correlation. However, terminal correlation is not only determined by instantaneous correlation but also by the way the average volatility is distributed among instantaneous volatilities [3] .

Libor Rate Dynamics under Spot Libor Measure

The libor-forward rates dynamics under the Spot Libor measure5 is given by;

d L j ( t ) L j = μ j d t + σ j ( t ) d W j , (17)

where;

μ j = σ j ( t ) j = m ( t ) i τ j ρ i , j σ j ( t ) L j ( t ) 1 + τ j L j ( t ) ,

and m(t) is the quantity defined by the relation T m ( t ) 1 < t < T m ( t ) , while τ j is the time fraction associated with Jth libor rate. W is a N-dimensional geometric Brownian motion with correlation between the rates defined as

d W j ( t ) d W i ( t ) = ρ i , j .

The spot numeraire is defined as;

B ( t ) = P ( t , T m ( t ) ) n = 0 m ( t ) 1 ( 1 + τ n L n ( t ) ) .

The task with the LFR is how to model volatility and correlation and how to estimate the parameters of these models for volatility and correlation. Two straightforward parameterizations are employed.

· Volatility

σ j ( t ) = ϕ i ( a ( T j t ) + b ) exp c ( T j t ) + d .

The main advantages functional form above is that allows for a humped volatility feature and its parameters or their combinations lends themselves to an easy interpretation i.e. a + d is the value of the instantaneous volatility of any forward rate as its expiry approaches zero, d- is the value of the instantaneous volatility for very long maturities, and the maximum of the hump is given by τ = 1 c a b . Furthermore, when coupled with a simple correlation function, the volatility above describes well and in parsimonious manner the whole swaption curve [24] .

· Correlation

ρ i , j = exp η | i j | .

The function above ensures that the correlation matrix is admissible as long η is positive.

Once the functional forms have been specified, the parameters must be estimated using market data. One useful approximation, initially developed by [10] , relates the Black volatility for a European swaption, given a set of volatility functions and a correlation matrix as;

( σ α , β L f r ( t ) ) 2 = i , j = 0 β w i ( 0 ) w j ( 0 ) L i ( 0 ) L j ( 0 ) ρ i , j R o , β 0 T α σ i ( t ) σ j ( t ) d t .

w’s described as;

w i ( t ) = τ i P ( t , T j ) j = α β 1 τ j P ( t , T j ) .

The aim is to minimize the objective function;

min ( i = 0 n | σ i m k t ( t ) σ i L f r ( t ) | ) 2 . (18)

4.6. Simulated Annealing Minimization

Letting Ω be the discretized solution space and f : Ω , the objective function defined in the space. The algorithm searches the global minimum solution ω g , such that f ( ω g ) < f ( ω ) for all ω in Ω . The objective function is bounded for ω g to exist. The objective function is used to calculate the cost value as;

( | σ i m k t ( t ) σ i L f r ( t ) | ) 2 .

4.6.1. Neighbourhood Function

In implementation, I will utilise the Boltz neighbourhood option to navigate the solution space. Boltz option uses a step length equal to the square root of temperature parameter to change from one solution to another with direction uniformly at random.

4.6.2. Cooling Schedule

There is no an academic consensus in regard to the choice of initial temperature. Most choices are made with regard to problem at hand. In implementation I use a default initial temperature of 100.

Afterwards the temperature is reduced by dividing the initial temperature with the logarithm of the rank of the iteration i.e.

T k = T o log k ,

k being the rank of the iteration.

4.6.3. Acceptance or Rejection Probability

Once a new solution has been sought, it’s compared to the previous solution, and accepted if better or the acceptance probability is between 0 and 0.5.

The acceptance function is:

P n = 1 1 + exp Δ T k ,

where:

· Δ is the difference between the current and previous cost value.

· T is the temperature parameter applicable at iteration k.

Stopping rule—The search stops when the set function tolerance is met.

4.7. Algorithm

4.8. Nonlinear Least-Squares (Lsqnonlin) Minimization

Least square is the oldest and most widely used minimisation method. Its popularity is due to the fact that it can be directly applied to a deterministic model without any cognizance being taken from the probability of the observations [25] . It is a constrained method that minimizes problems of the form;

min x f ( x ) 2 2 = min x ( f 1 ( x ) 2 + f 2 ( x ) 2 + + f n ( x ) 2 ) ,

where the objective function f ( x ) 2 2 is defined in terms of auxiliary functions f i with optional lower and upper bounds on the components of x. The objective function corresponds to the residuals in a data fitting problem i.e.

f ( x ) = y ^ i y i ,

where y ^ i is the model estimate and y i is the market observations.

Through an iteration procedure, the method minimises the sum of square residual value up to a set tolerance level. However, as against the Ordinary Least Squares (OLS) estimation, there is no closed-form solution for this system of equations formed, hence we make small adjustments to the predictor values at each iteration. Rather than computing the sum of least squares, the method requires a user-defined function to compute the vector-valued function f 1 ( x ) , f 2 ( x ) , , f n ( x ) of the residuals.

4.9. Monte-Carlo Simulation

Monte-Carlo method is a numerical formula where random numbers are used for scientific experiment. Monte Carlo simulation is perchance the most common technique for propagating the incertitude of the various aspects of a system to the predicted performance. In Monte Carlo simulation, the entire system is simulated a large number of times i.e. a set of suitable sample paths is produced on [ t 0 , T ] . Each simulation is equally likely and it’s referred to as a realization of the system. For each realization, all of the uncertain parameters are sampled. For each sample, we produce a sample path solution to the SDE on [ t 0 , T ] . The realization is generally obtained from the stochastic Ito-Taylor expansion. From the Ito-Taylor expansion, we can construct numerical schemes for the interval [ t i , t i + 1 ] .

The dynamics of the libor rates under the Spot measure is:

d L k ( t ) = σ k ( t ) L k ( t ) j = 0 + 1 β p k , j τ j σ j ( t ) L j ( t ) 1 + τ j L j ( t ) d t + σ k ( t ) L k ( t ) d Z k . (19)

Since the dynamics above doesn’t yield a distributionally known results, its proper to discretize it to finer time frame that would enable the process reduce the random inputs from distributionally known Gaussian shocks. Use of logarithm of the rates can help achieve the above objective. Taking logs and applying Ito’s lemma, dynamics of (19) become;

d ln L k ( t ) = σ k ( t ) j = o + 1 β p k , j τ j σ j ( t ) L j ( t ) 1 + τ j L j ( t ) d t σ k ( t ) 2 2 d t + σ k ( t ) d Z k . (20)

Equation (20) has a diffusion process that is deterministic, as a consequence, the naive Euler scheme coincides with the more sophisticated Milstein scheme.

So the discretization becomes;

ln L k Δ t ( t + Δ t ) = ln L k Δ t ( t ) + σ k ( t ) j = o + 1 β p k , j τ j σ j ( t ) L j Δ t ( t ) 1 + τ j L j Δ t ( t ) Δ t σ k ( t ) 2 2 Δ t + σ k ( t ) ( Z k ( t + Δ t ) Z k ( t ) ) . (21)

This discretization leads to an approximation of the true process such that there exists a γ 0 with

E α ( | ln L k Δ t ( t + Δ t ) ln L k Δ t ( t ) | ) C ( T o ) Δ t ,

for Δ t γ 0 , and C ( T o ) Δ t is a positive constant. This gives strong convergence of order 1, from the exponent of Δ t on the right-hand side [3] .

For k = o + 1 , o + 2 , , β , we generate M of such realizations. After simulating the libor rates, they are subsequently used to calculate the swap rates through the relation indicated by section (4.3). Henceforth evaluating the payoff as:

i = α + 1 β P ( T o , T i ) ( R 0 , β K ) + , (22)

for each realization and the average becomes the Swaption price.

Antithetic variates method for variance reduction was used to reduce the effects of discretization and simulation errors, so as to make sure that the price difference depicted is as a result of the different optimization techniques used.

4.10. Data

The data used in this study was obtained from [26] website. The data was used to price a swaption whilst using lsqnonlin method hence making it a good series to compare with simulated annealing performance.

5. Results & Discussion

In simulated annealing options, I used the Boltz method for my neighborhood and temperature function. For both SA and lsqnonlin, the feasible set of solutions had lower bound of [0 0 0.5 0 0.01] and upper bound of [1 1 2 0.3 1] with the initial point being [1.2.05 1.05.2]. The stopping criteria were based on the objective function achieving a tolerance of 1e−5.

5.1. Data Evaluation

The stylized features of the data can be ascribed from; the zero curves, forward curve and the evolution of the Libor market model rates. (Shown in Figure 2 below).

A zero curve maps the prices of zero coupon bonds to different maturities against time. This curve helps in pricing of fixed incomes securities and derivatives since it gives a fair value of capital gains accepted in the market. The slope of the curve indicates it is normal i.e. there is more compensation for each increment of risk taken. In derivatives market, longer time increases the chance of

Figure 2. Zero & Libor curves.

negative events occurring hence higher risk.

Libor-forward rates can be transcribed from zero coupon bonds as shown in Equation (33) (in the appendix). They give insight to the price of a forward contract relative to its time to maturity. It is an experienced fact that the curve flattens out as time lapses. This because of increased correlation between the rates as time elapses.

The evolution of the market model is an indicator of the how the well the optimization techniques mimick the real world market. The two methods have managed to capture the sporadic movements of the libor rates, phenomena that become commonly observed after the financial crisis of 2007.

5.2. Volatility Function

The plots (Figure 3) below are the values of the stipulated volatility function whose parameters was estimated as [0.3744 0.0385 1.9454 0.1542] by SA method

Figure 3. Volatility Plots. (a) Simulated Annealing plot; (b) Lsqnonlin plot.

and [0.0932 0.1524 0.5745 0.0707] by lsqnonlin method.

Market implied volatility curve is humped as a result of the control from the monetary authorities. The parameterization function adopted is capable of capturing the humped term structure of volatility in case the market volatility is originally humped. However, lsqnonlin looses the stylistic feature whereas SA is able to maintain it.

5.3. Correlation Matrices

As mentioned earlier, the terminal correlation of libor-forward rates do play a role in pricing swaption. The correlation is then used to calculate the value of the price. Below (Table 1) represent the correlation matrix that was calculated using the coefficient η (0.0996) estimated by SA.

Lsqnonlin on the other hand, estimated the η to be 0.0100, and corresponding correlation matrix is (Table 2).

The methods show typical qualities of a correlation matrix i.e. positive correlation with 1’s in the leading diagonal. Also moving away from the leading

Table 1. SA correlation matrix.

Table 2. lsqnonlin correlation matrix.

diagonal, the measures are decreasing clearly showing that the joint movements of far away rates are less correlated than movements of rates with close maturities. The sub-diagonals are increasing as one approaches the leading diagonal, an indication of a larger correlation of adjacent rates.

Albeit the similarities, lsqnonlin predicts much higher correlations than SA. Such a flaw is undesired since the correlation is much lower for rates that are far off each other. SA doesn’t reflect the drawback.

5.4. Price Comparison

Table 3 below represents the prices of swaptions implied by the lsqnonlin method and the simulated annealing optimization, with their respective deviations from the Black’s prices. The swaption had a maturity of 5 years and a tenor ranging from 1 Year to 5 years. The instrument strike was 0.045.

The errors are plotted in the combined plot (Figure 4).

We set out to price a swaption using both lsqnonlin and simulated annealing optimization techniques, and did so by using the methods to minimize the difference between model volatility against market quotes. We have demonstrated that the simulated annealing produces lower errors than lsqnonlin as evident from the graph. Both methods value the swaption with a tenor of one as having an almost zero value hence same errors at onset. However as the swaption gains value, SA systematically predicts more precise prices than the contra.

Table 3. Price and errors for SA & lsqnonlin.

Figure 4. Error plots for SA & lsqnonlin (Data 1 represent the lsqnonlin errors, data 2-SA errors).

6. Conclusions & Future Work

The goal of the thesis was to conduct a simulation test as to whether the SA optimization method outperforms the lsqnonlin method in finding a better solution to the volatility & correlation parametization functions when set under the same conditions. The methods were benchmarked against desirable market features.

The term structure of implied volatility is humped in nature indicating high uncertainty of intermediate forward rates. This phenomenon is a result of influence from the monetary authorities in that; at the short end of the maturity spectrum, the authorities determine the short deposit rates which then influence the short maturity forward rates, while at the long maturity spectrum, the authorities control the forward rates so as to achieve a set inflation target. This leaves the intermediate period as a time when loose or tight regimes can be reversed or continued beyond what was originally anticipated. This state of affairs gives rise to maximum market uncertainty in the intermediate-maturity region, hence the volatility of the long-dated or of the very-short-dated forward rates will not be as pronounced as that of the intermediate-maturity forward rates [27] . Albeit not identical to the market curve, volatility functions should be able to reflect the quantitative shape for proper derivative pricing. The results indicate that lsqnonlin misses portraying the hump of the intermediate rates instead showing only monotonic decaying volatility whereas SA features both the hump and the monotonic decay.

In a time-homogeneous world, decorrelation between two forward rates depends on how distant the rates are i.e. rates that are further apart are more decorrelated than the ones in close proximity. This is as a result of the shock affecting the first rate gradually dying out hence having little or no effect on the later rates. Although both techniques show this feature, in lsqnonlin method, it is less pronounced indicating that the rates are more correlated despite the dying off of the shock. On the other hand, SA adequately depicts this effect by having a more pronounced decorrelation among non-proximal rates.

The difference in the two methods to portray market features is easily depicted in the pricing of swaptions as evident from the value of errors and the graph. The more robust SA optimization has fewer errors propagated as compared to lsqnonlin, hence proving that indeed the lsqnonlin method does get trapped in the local minimum, consequently overstating the free parameters which in turn influence the price level.

The thesis thus proposes the adoption of simulated annealing optimization as the standard methodology for minimizing the difference between the model and market volatilities for greater price accuracy.

SA has many variant components including the use of linear decreasing temperature in lieu of lognormal, fast acceptance criterion instead of Boltz, and the maximum iteration stopping criteria as opposed to error tolerance. The thesis used the default options embedded in SA i.e. Boltz neighborhood function and decreasing lognormal temperature function, leaving the variants and custom functions for future study. The variants can be investigated to verify if they have less computational time than the ones used in the thesis.

The main drawback of the correlation function adopted is that it predicts the decorrelation of any two equidistant rates to be almost the same irrespective of whether the first forward rate expires in two months or one year i.e. the first and second rates will decorrelate just as much as the 9th and the 10th. Normally, long-dated rates are less decor-related than the short ones. This financially undesirable feature is a result of the absence of explicit time dependence in the function. A more desiderate specification would be the modified exponential form. Incorporation of the latter form and SA optimization could be investigated to discern if there is a further reduction in the disparity between numerical and analytical swaption prices.

Appendix

BGM Framework

BGM derivation begins from the Brace-Musiela (BM) (1994) parameterisation of the Heath-Jarrow-Morton (HJM) model. The HJM stochastic integration equation for under the the risk neutral measure is;

f ( t , T ) = f ( 0 , T ) + 0 t a ( ν , T ) + 0 t σ ( ν , T ) d W ^ ( ν ) , (23)

where t is the time the rate is quoted and T is the time it applicable. Notably,T is fixed whereas t is a variable. For the model to be arbitrage free, the drift adopts the form,

a ( ν , T ) = σ ( t , T ) t T σ ( t , s ) .

BM considers a fixed period ahead rate

ϒ ( t , x ) f ( t , t + x ) , (24)

where ϒ ( t , x ) is the rate that is quoted at time t for instantaneous borrowing a time t + x, x being fixed time aheadi.e. 3 months. Hence Equation (23) becomes

f ( t , t + x ) = f ( 0, t + x ) + 0 t a ( ν , t + x ) + 0 t σ ( ν , t + x ) d W ^ ( ν ) . (25)

BM further redefines σ ( t , t + x ) as τ ( t , x ) . With the above notation for drift and diffusion, the variables can written as

σ ( υ , t + x ) = σ ( υ , υ + ( t + x υ ) ) = τ ( υ , ( t + x υ ) ) ,

and

a ( υ , x ) = a ( υ , υ + ( t + x υ ) ) = a ( υ , ( t + x υ ) ) .

The notation will be of great importance when adopting the Heath-Jarrow-Morton drift restriction.

Adopting the BM notation, Equation (25) becomes;

ϒ ( t , x ) = ϒ ( 0 , t + x ) + 0 t a ( υ , ( t + x υ ) ) d υ + 0 t τ ( υ , ( t + x υ ) ) d W ^ ( ν ) . (26)

In line with arbitrage free pricing, we adopt the notation of drift restriction as;

a ( ν , t + x ) = σ ( t , t + x ) t t + x σ ( t , s ) d s = σ ( t , t + x ) 0 x σ ( t , t + y ) d s = τ ( t , x ) 0 x τ ( t , y ) = τ ( t , x ) ψ ( t , x ) ,

defining the integrated volatility as ψ ( t , x ) . The notation permits further manipulation of the drift as;

a ( ν , t + x ) = d d x ( 1 2 ψ 2 ( t , x ) ) ,

and the diffusion as;

τ ( t , t + x ) = d d x ψ ( t , x ) .

Not having t appearing in the second argument will be important in carrying out volatility transformations that will enable us have a log-normal libor rates.

Equation (26) is in stochastic integration form hence needs to transform to SDE by firstly differentiating with respect to x;

d d x ϒ ( t , x ) = ϒ 2 ( t , x ) = ϒ 2 ( 0 , t + x ) + 0 t a 2 ( υ , ( t + x υ ) ) d υ + 0 t τ 2 ( υ , ( t + x υ ) ) d W ^ ( ν ) . (27)

And then forming the SDE as:

d ϒ ( t , x ) = [ ϒ 2 ( 0 , t + x ) + a ( t , x ) + 0 t a 2 ( υ , ( t + x υ ) ) d υ + 0 t τ 2 ( υ , ( t + x υ ) ) d W ^ ( ν ) ] + τ ( t , x ) d W ^ ( ν ) (28)

Adopting Equation (27)’s drift, and diffusion notation, the BGM Stochastic differential equation (SDE) for the instantaneous forward rate becomes

d ϒ ( t , x ) = d d x [ ( ϒ ( t , x ) + 1 2 ψ 2 ( t , x ) ) d t + ψ ( t , x ) d W ^ ( t ) ] .

The above representation allows bond prices to be valued in terms of time to maturity i.e. the bond matures at a fixed period ahead in lieu of fixed date i.e.

P ( t , T + x ) = exp [ t T r ( t , s t ) d s ] , (29)

where P (t, T + x) is the price of zero coupon bond at time t that will mature in x time period, x being some accrual period like 3 months.

By changing the variable u = s t , equation (29) becomes;

P ( t , T ) = exp [ 0 T t r ( t , u ) d u ] . (30)

Libor

BGM instatenous forward rate relates to the libor process, L ( t , t + x + δ ) through the equation:

1 + δ L ( t , t + x + δ ) = exp [ x x + δ r ( t , x ) ] . (31)

The libor is defined as a simple compounded rate that an investor can contract at time t for borrowing/lending over time t + x to t + x + δ , δ being a discrete time tenor of maybe 3,6, or 12 months. Equation (31) also implies that the simple compounded rate must be in line with continuously compounded rate over the period.

In accordance with Equation (30), the rate can also relate to the bond prices as;

1 + δ L ( t , t + x + δ ) = exp [ x x + δ r ( t , x ) ] = exp [ 0 x + δ r ( t , x ) 0 x r ( t , x ) ] (32)

= P ( t , T + x ) P ( t , T + x + δ ) . (33)

To determine the SDE for Libor rate, we start by evaluate the quantity [ x x + δ r ( t , x ) ] by equating to variable as;

V ( t , x ) = x x + δ r ( t , x ) , (34)

= x x + δ ϒ ( 0 , t + x ) d x + x x + δ 0 t a ( υ , ( t + x υ ) ) d υ d x + x x + δ 0 t τ ( υ , ( t + x υ ) ) d W ^ ( ν ) d x . (35)

For a proper SDE, we need to change the order of integration using Fubini theorem, i.e.

= x x + δ ϒ ( 0 , t + x ) d x + 0 t x x + δ a ( υ , ( t + x υ ) ) d x d υ + 0 t x x + δ τ ( υ , ( t + x υ ) ) d x d W ^ ( ν ) . (36)

Differentiating Equation (36) with respect to t;

d V = [ x x + δ ϒ 2 ( 0 , t + x ) + 0 t x x + δ a 2 ( υ , ( t + x υ ) ) d x d υ + x x + δ a ( t , υ ) d υ (37)

+ 0 t x x + δ τ 2 ( υ , ( t + x υ ) ) d x d W ^ ( ν ) ] d t + x x + δ τ ( t , υ ) d υ W ^ ( ν ) . (38)

From the above equation, the differentials can be simplified to;

x x + δ ϒ 2 ( t , x ) = x x + δ ϒ 2 ( 0 , t + x ) + 0 t x x + δ a 2 ( υ , ( t + x υ ) ) d x d υ + 0 t x x + δ τ 2 ( υ , ( t + x υ ) ) d x d W ^ ( ν ) ,

ϒ ( t , x + δ ) ϒ ( t , x ) = x x + δ ϒ 2 ( 0 , t + x ) + 0 t x x + δ a 2 ( υ , ( t + x υ ) ) d x d υ + 0 t x x + δ τ 2 ( υ , ( t + x υ ) ) d x d W ^ ( ν ) ,

hence Equation (38) can be rewritten as;

d V = [ ϒ ( t , x + δ ) ϒ ( t , x ) + x x + δ a ( t , υ ) d υ ] d t + x x + δ τ ( t , υ ) d υ W ^ ( ν ) . (39)

Adopting the notations;

x t + δ a ( t , υ ) d u = 1 2 ψ 2 ( t , x + δ ) ψ 2 ( t , x ) ,

x t + δ τ ( t , υ ) = ψ ( t , x + δ ) ψ ( t , x ) ,

Equation (39) can be simplified further as;

d V = μ v d t + σ v d w ^ t , (40)

where:

μ v = f ( t , x + δ ) f ( t , x ) + 1 2 ψ 2 ( t , x + δ ) ψ 2 ( t , x ) , (41)

σ v = ψ ( t , x + δ ) ψ ( t , x ) . (42)

The quantity dV and application of Ito’s lemme can be used to derive the SDE of L ( t , t + x + δ ) (henceforth L ( t , δ ) ) as;

L ( t , δ ) = exp V ( t , x ) 1 δ , (43)

d L = δ 1 exp V ( μ v + 1 2 σ v 2 ) + δ 1 exp V σ v d W ^ t . (44)

By observing that;

d d x L ( t , δ ) = δ 1 exp ( x x + δ ϒ ( t , t + x ) ) ϒ ( t , t + x + δ ) ϒ ( t , t + x ) , = δ 1 ( 1 + δ L ( t , δ ) ) ϒ ( t , t + x + δ ) ϒ ( t , t + x ) ,

the SDE for L ( t , δ ) can be expressed as;

d L ( t , δ ) = d d x L ( t , δ ) + δ 1 ( 1 + δ L ( t , δ ) ) ψ ( t , x + δ ) ( ψ ( t , x + δ ) ψ ( t , x ) ) d t + δ 1 ( 1 + δ + L ( t , δ ) ) ( ϒ ( t , t + x + δ ) ϒ ( t , t + x ) ) . (45)

By further adopting the BGM volatility function

δ 1 ( 1 + δ L ( t , δ ) ) ψ ( t , x + δ ) ψ ( t , x ) = γ ( t , x ) L ( t , δ ) ,

where γ ( t , x ) is a function of time and maturity, Equation (45) can be rewritten as;

d L ( t , δ ) = [ d d x L ( t , δ ) + γ ( t , x ) L ( t , δ ) ψ ( t , x ) + δ γ 2 ( t , x ) L 2 ( t , δ ) 1 + δ L ( t , δ ) ] d t + γ ( t , x ) L ( t , δ ) d W ^ ( υ ) . (46)

Formulation (46) is the log-normal dynamics of the libor process that helps to recover the Black’s formula for Swaption prices, albeit it has a complicated drift term. BGM solution to the drift problem is considering another process K ( t , T ) = L ( t , T t ) . By taking differential with respect to the second argument,

d K ( t , T t ) = L ( t , T t ) L 2 ( t , T t ) d t , (47)

in deriving the equation above, BGM used the fact that that L 2 ( t , T t ) is smooth function of L ( t , δ ) . A combinatorial of the BGM volatility function, Equation (45), and (47), the SDE for Libor process becomes

d L j = L j ( t ) γ j ( t ) [ ψ j + 1 ( t ) d t + d W ^ ( t ) ] (48)

Applying Girsanov theorem from risk -neutral measure to forward ( T j + 1 ) measure, changes BGM SDE to

d L j = L j ( t ) γ j ( t ) d W T j + 1 ( t ) (49)

which is drift-less,( indicating it a martingale under the measure) and also posses the desirable property of being log-normal.

NOTES

1Variables with the same value in all optimal solutions.

2Libor and BGM will be used interchangeably.

3e.g. under T-forward measure, the numeraire is a zero coupon bond whose price at maturity is a unit notional amount.

4Under risk neutral measure.

5The path-dependent derivatives can be accurately evaluated by constructing random paths of the libor process using either the forward-risk adjusted or spot Libor dynamics mainly because at each payment date T j + 1 spot Libor L j ( t ) is received and an amount equal τK is paid [23] .

Conflicts of Interest

The authors declare no conflicts of interest.

References

[1] Barnes, C. (2021) Swaption Volumes by Strike Q1 2021. Clarus Financial Technology.
[2] Fan, R., Gupta, A. and Ritchken, P. (2007) On Pricing and Hedging in the Swaption Market: How Many Factors, Really? The Journal of Derivatives, 15, 9-33.
https://doi.org/10.3905/jod.2007.694699
[3] Brigo, D. and Mercurio, F. (2007) Interest Rate Models-Theory and Practice: With Smile, Inflation and Credit. Springer Science & Business Media, Berlin.
[4] Ledesma, S., Aviña, G. and Sanchez, R. (2008) Practical Considerations for Simulated Annealing Implementation. Simulated Annealing, 20, 401-420.
https://doi.org/10.5772/5560
[5] Ingber, L. (2000) Adaptive Simulated Annealing (ASA): Lessons Learned.
[6] Rebonato, R. (1999) On the Simultaneous Calibration of Multifactor Lognormal Interest Rate Models to Black Volatilities and to the Correlation Matrix. Journal of Computational Finance, 2, 5-27. https://doi.org/10.21314/JCF.1999.031
[7] Brace, A., Gatarek, D. and Musiela, M. (1997) The Market Model of Interest Rate Dynamics. Mathematical Finance, 7, 127-155.
https://doi.org/10.1111/1467-9965.00028
[8] Chiarella, C., He, X.Z., Nikitopoulos, C.S., et al. (2016) Derivative Security Pricing. Springer, Berlin. https://doi.org/10.1007/978-3-662-45906-5
[9] Lagunzad, D.U. (2007) On the Calibration of the LIBOR Market Model. PhD Thesis, Singapore Management University, Singapore.
[10] Jackel, P. and Rebonato, R. (2003) The Link between Caplet and Swaption Volatilities in a Brace-Gatarek-Musiela/Jamshidian Framework: Approximate Solutions and Empirical Evidence. Journal of Computational Finance, 6, 41-60.
https://doi.org/10.21314/JCF.2003.100
[11] Pascucci, A. and Riga, C. (2011) The Libor Market Model: From Theory to Calibration.
[12] Ingber, L. (1989) Very Fast Simulated Re-Annealing. Mathematical and Computer Modelling, 12, 967-973. https://doi.org/10.1016/0895-7177(89)90202-1
[13] Henderson, D., Jacobson, S.H. and Johnson, A.W. (2003) The Theory and Practice of Simulated Annealing. In: Handbook of Metaheuristics, Springer, Berlin, 287-319.
https://doi.org/10.1007/0-306-48056-5_10
[14] Kirkpatrick, S., Gelatt, C.D. and Vecchi, M.P. (1983) Optimization by Simulated Annealing. Science, 220, 671-680. https://doi.org/10.1126/science.220.4598.671
[15] Martin, J.F.D. and Sierra, J.M.R. (2009) A Comparison of Cooling Schedules for Simulated Annealing (Artificial Intelligence). In: Encyclopedia of Artificial Intelligence, IGI Global, Hershey, 344-352.
https://doi.org/10.4018/978-1-59904-849-9.ch053
[16] Ben-Ameur, W. (2004) Computing the Initial Temperature of Simulated Annealing. Computational Optimization and Applications, 29, 369-385.
https://doi.org/10.1023/B:COAP.0000044187.23143.bd
[17] Gu, X. (2008) The Behavior of Simulated Annealing in Stochastic Optimization.
[18] Luo, Y., Zhu, B. and Tang, Y. (2014) Simulated Annealing Algorithm for Optimal Capital Growth. Physica A: Statistical Mechanics and Its Applications, 408, 10-18.
https://doi.org/10.1016/j.physa.2014.04.020
[19] Fogarasi, N. and Levendovszky, J. (2013) Sparse, Mean Reverting Portfolio Selection Using Simulated Annealing. Algorithmic Finance, 2, 197-211.
https://doi.org/10.3233/AF-13026
[20] Beyna, I. and Wystup, U. (2010) On the Calibration of the Cheyette Interest Rate Model. Tech. Report, CPQF Working Paper Series.
[21] Black, F. (1976) The Pricing of Commodity Contracts. Journal of Financial Economics, 3, 167-179. https://doi.org/10.1016/0304-405X(76)90024-6
[22] Burgess, N. (2014) Martingale Measures & Change of Measure Explained.
https://doi.org/10.2139/ssrn.2961006
[23] Jamshidian, F. (1997) LIBOR and Swap Market Models and Measures. Finance and Stochastics, 1, 293-330. https://doi.org/10.1007/s007800050026
[24] Rebonato, R., McKay, K. and White, R. (2009) The SABR/LIBOR Market Model: Pricing, Calibration and Hedging for Complex Interest-Rate Derivatives. John Wiley & Sons, Hoboken.
[25] Bard, Y. (1974) Nonlinear Parameter Estimation. Tech. Rep.
[26] Price Swaptions with Interest-Rate Models Using Simulation.
https://uk.mathworks.com/help/fininst/price-bermudan-swaptions-with-different-interest-rate-models.html#bts_c9j-1
[27] Rebonato, R. (2005) Volatility and Correlation: The Perfect Hedger and the Fox. John Wiley & Sons, Hoboken. https://doi.org/10.1002/9781118673539

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.