Abstract
We introduce a novel non-parametric methodology to test for the dynamical time evolution of the lag–lead structure between two arbitrary time series. The method consists of constructing a distance matrix based on the matching of all sample data pairs between the two time series. Then, the lag–lead structure is searched for as the optimal path in the distance matrix landscape that minimizes the total mismatch between the two time series, and that obeys a one-to-one causal matching condition. To make the solution robust to the presence of a large amount of noise that may lead to spurious structures in the distance matrix landscape, we generalize this optimal search by introducing a fuzzy search by sampling over all possible paths, each path being weighted according to a multinomial logit or equivalently Boltzmann factor proportional to the exponential of the global mismatch of this path. We present the efficient transfer matrix method that solves the problem and test it on simple synthetic examples to demonstrate its properties and usefulness compared with the standard running-time cross-correlation method. We then apply our ‘optimal thermal causal path’ method to the question of the lag-dependence between the US stock market and the treasury bond yields and confirm our earlier results on an arrow of the stock markets preceding the Federal Reserve Funds’ adjustments, as well as the yield rates at short maturities in the period 2000–2003. Our application of this technique to inflation, inflation change, GDP growth rate and unemployment rate unearths non-trivial lag relationships: the GDP changes lead inflation especially since the 1980s, inflation changes leads GDP only in the 1980 decade, and inflation leads unemployment rates since the 1970s. In addition, our approach seems to detect multiple competing lag structures in which one can have inflation leading GDP with a certain lag time and GDP feeding back/leading inflation with another lag time.
1. Introduction
Determining the arrow of causality between two time series X(t) and Y(t) has a long history, especially in economics, econometrics and finance, as it is often asked which economic variable might influence other economic phenomena (Chamberlain 1982, Geweke 1984). This question is raised in particular for the relationships between, respectively, inflation and GDP, inflation and growth rate, interest rate and stock market returns, exchange rate and stock prices, bond yields and stock prices, returns and volatility, advertising and consumption and so on. One simple naive measure is the lagged cross-correlation function CX,Y (τ)= ⟨X(t) Y(t+τ)⟩ , where the brackets ⟨x⟩ denotes the statistical expectation of the random variable x. Then, a maximum of CX,Y (τ) at some non-zero positive time lag τ implies that the knowledge of X at time t gives some information on the future realization of Y at the later time t+τ. However, such correlations do not necessarily imply causality in a strict sense as a correlation may be mediated by a common source influencing the two time series at different times. In addition, lagged correlations may appear between two endogenous variables when time delays are present: speaking of causality in this case would be similar to a chicken-and-egg problem. The concept of Granger causality extends the notion of lagged correlation by taking a pragmatic approach based on predictability: if the knowledge of X(t) and of its past values improves the prediction of Y(t+τ) for some τ > 0, then it is said that X Granger causes Y (Ashley et al. 1980, Geweke 1984) (see also Chen et al. (2004) for a recent extension to nonlinear time series). Neither lagged correlations nor Granger causality address the fundamental philosophical and epistemological question of the real causality links between X and Y, which would require one to distinguish between exogenous variables, which can be viewed as the primitives of a model, and endogenous variables, which are determined within the model. An endogenous variable can have no causal effect on another variable because it is determined by the model and one is not free to change it. Only exogenous variables can be causal variables. Our approach belongs to the same class as lagged-correlations and Granger causality by not addressing the issue of exogenous versus endogenous variables. It offers a new way to detect the existence and amplitude of dependences between two time series, whether exogenously or endogenously related.
However, most economic and financial time series are not strictly stationary and the lagged correlation between two time series may be changing as a function time, for instance reflecting regime switches and/or changing agent expectations. It is thus important to define tests of lagged dependence which are sufficiently reactive to such regime switches, allowing one to follow almost in real time the evolving structure of the dependence. Cross-correlation methods and Granger causality tests require a rather substantial amount of data in order to obtain reliable conclusions. In addition, cross-correlation techniques are fundamentally linear measures of dependence and may miss important nonlinear dependence properties. Granger causality tests are most often formulated using linear parametric auto-regressive models. The new technique introduced in this paper, called the ‘optimal thermal causal path’, is both non-parametric and sufficiently general so as to detect a priori arbitrary nonlinear dependence structures. Moreover, it is specifically conceived so as to adapt to the time evolution of the dependence structure. The ‘optimal thermal causal path’ can be viewed as an extension of the ‘time distance’ measure which amounts to comparing trend lines upon horizontal differences of two time series (Granger and Jeon 1997). It offers a new and different approach to the detection of dependences between two time series, which extends the standard lagged-correlation method and complements the Granger causality method.
Our method has deep roots in the field of computational statistics, in particular in the problem of recognizing related sequences going back to Needleman and Wunsch (1970). These authors introduced a general computer adaptable method for finding similarities in the amino acid sequences of two proteins, in order to determine whether significant homology exists between the proteins and to trace their possible evolutionary development. The book edited by Sankoff and Kruskal (1983) is an excellent compilation of papers explaining how to recognize related sequences, how to measure distance between sequences, and how to compute that measure effectively, with applications to molecular biology, speech recognition, analysis of bird song and error correcting in computer software. The concept of ‘time-warping,’ a technique for comparing functions by altering the time axis which has been developed for speech recognition, is conceptually similar to our approach (and is exposed in part 2 of Sankoff and Kruskal's (1983) book). Let us also mention the series of papers by Hwa and co-workers (Hwa and Lassig 1996, Drasdo et al. 1998, Hwa and Lassig 1998, Olsen et al. 1999, Drasdo et al. 2001) which studied the problem of detecting similarities between long DNA and protein sequences using concepts of statistical physics, similar to those discussed below. While the method of Hwa and co-workers makes use of the fact that there are four basic units in DNA sequences to detect similarities, our goal is to recover the lag between two time series in a dynamical way; we thus have to normalize the time series and define a distance between all possible pairs of increments of the two time series. By studying an effective attraction between two similar series, the effect of noise and the corresponding energy landscapes, Hwa and co-workers found a localization–delocalization phase transition as a function of a control parameter: the ‘fidelity’ of the alignment. For our application on two time series, this localization–delocalization transition would correspond to the fact that the optimal path (lag) is localized (the fluctuations of the lag are bounded) if the two time series match sufficiently well, while the optimal path (lag) is delocalized and controlled by noise if the two time series are not matching well enough. In the work presented below, we add the possibility of introducing a temperature so that we calculate finite-temperature-average-optimal paths. We have found that adding a non-zero temperature often improves the matching of two time series. Our interpretation is that the temperature somehow smoothes out the idiosyncratic noises and provides more robustness. We have however not explored the possibility that there could be another more subtle transition between the localized and delocalized phases, when introducing a non-zero temperature, in the spirit of Hwa and co-workers. This is an interesting question for future studies.
The organization of the paper is as follows. Section 2 defines the ‘optimal thermal causal path’ method. Section 3 applies it to simple auto-regressive models, in an initial test of its properties and limitations. Section 4 presents an application of the optimal thermal causal path method on two important economic problems: the dependence relationship between the US treasury bond yields and the stock market in the aftermath of the Internet bubble collapse and between inflation, inflation change, gross domestic product rate and unemployment rate in the United States. Section 5 concludes. The Appendix presents the mathematical algorithm underlying the construction of the optimal thermal causal path.
2. Definition of the ‘optimal thermal causal path’ method
The key ideas behind the optimal thermal causal path method can be summarized as follows.
-
A distance matrix is formed which allows one to compare systematically all values of the first time series X(t 1) along the time axis with all the values of the second time series Y(t 2), via the introduction of a distance d(X(t1),Y(t2)).
-
The dependence relationship between the two time series is searched for in the form of a one-to-one mapping t2 =φ (t1 ) between the times t1 of the first time series and the times t2 of the second time series such that the two time series are the closest in a certain sense, i.e. X(t 1) and Y(φ (t1 )) match best. We impose in addition a kind of smoothness requirement, equivalent in most cases to continuity and monotonicity of the mapping φ. But, our ‘optimal thermal causal path’ method allows us to detect situations in which the lag can jump and behave in an arbitrary way as a function of time, as in the example (12) below.
-
The optimal matching in step 2 is performed by introducing a weighted average over many potential mappings in order to remove as much as possible the influence of non-informative noises in both time series. There is an exact mapping of this problem to a well-known problem in statistical physics known as the directed polymer in a quenched random potential landscape at non-zero temperature, hence the name ‘optimal thermal causal path’.
-
The resulting mapping defines the lag between the two time series as a function of time that best synchronizes or matches them. This thus allows us to obtain the time evolution of the dependence relationship between the two time series.
We now describe in detail how to implement these ideas.
2.1. Distance matrix
To simplify, we consider time series updated in discrete time, in units of some elementary discretization step, taken as unity without loss of generality. Let us denote X(t1):t 1=0,…,N 1−1 and Y(t2):t 2=0,…, N 2−1 the two time series of which we would like to test their dependence. Note that the lengths N 1 and N 2 of the two series can in principle be different as our method generalizes straightforwardly to this case, but for the sake of pedagogy, we restrict ourselves here to the case N 1=N 2=N. These time series X(t 1) and Y(t2) can be very different in nature with largely differing units and meanings. To make them comparable, we normalize them by their respective standard deviations, so that both normalized time series have comparable typical values. From now on, the two time series X(t 1) and Y(t2) denote these normalized time series.
We introduce a distance matrix E X , Y between X to Y with elements defined as
When Y(t) is the same time series as X(t), a matrix deduced from (1), by introducing a threshold so that entries of the matrix (1) smaller (respectively larger) than the threshold are set to 0 (respectively 1), has been introduced under the name of ‘recurrence plot’ to analyse complex chaotic time series (Eckmann et al. 1987). In the physical literature, the binary matrix deduced from (1) with the use of a threshold for two different time series is called a cross-recurrence plot. This matrix and several of its statistical properties have been used to characterize the cross-correlation structure between pairs of time series (Marwan and Kurths 2002, Marwan et al. 2002, Quiroga et al. 2002, Strozzia et al. 2002).
Consider the simple example in which Y(t)=X(t−k) with k > 0 is a positive constant. Then, ε (t 1,t 2) =0 for t 2=t 1+k and is typically non-zero otherwise. The detection of this dependence relationship then amounts in this case to finding the line with zero values which is parallel to the main diagonal of the distance matrix. This line defines the affine mapping t 2=φ(t 1) = t 1+k, corresponding to a constant translation. More generally, we would like to determine a sequence of elements of this distance matrix along which the elements are the smallest, as we describe next.
2.2. Optimal path at ‘zero temperature’
When the relationship between X(t 1) and Y(t 2) is more complex than a simple constant lead-lag of the form Y(t)=X(t−k), the determination of the correspondence between the two time series is less obvious. An initial approach would correspond to an association of each entry X(t 1) of the first time series to the value Y(t 2) of the second time series which makes the distance (1) minimum over all possible t 2 for a fixed t 1. This defines the mapping t 1→t 2=φ(t 1) from the t 1-variable to the t 2-variable as
To address these two problems, we first search for a smooth mapping t 1→t 2=φ(t 1):
To implement these ideas, our first proposal is to replace the mapping (2) determined by a local minimization by a mapping obtained by the following global minimization:
This problem actually has a long history and has been extensively studied, in particular in statistical physics (see Halpin-Healy and Zhang (1995) for a review and references therein), under the name of the ‘random directed polymer at zero temperature’. Indeed, the distance matrix E X , Y given by (1) can be interpreted as an energy landscape in the plane (t 1 , t 2 ) in which the local distance ε (t 1,t 2) is the energy associated with the node (t 1 , t 2 ). The continuity constraint means that the mapping defines a path or line or ‘polymer’ of equation (t 1,t 2=φ(t 1)) with a ‘surface tension’ preventing discontinuities. The condition that φ(t 1) is non-decreasing corresponds to the fact that the polymer should be directed (it does not turn backward and there are no overhangs). The global minimization problem (4) translates into searching for the polymer configuration with minimum energy. In the case where the two time series are random, the distance matrix (and thus energy landscape) is random, and the optimal path is then called a random directed polymer at zero temperature (this last term ‘at zero temperature’ will become clear in section 2.3). Of course, we are interested in non-random time series, or at least in time series with some non-random components: this amounts to having the distance matrix and the energy landscape to have hopefully coherent structures (i.e. non-white noise) that we can detect. Intuitively, the lag–lead structure of the two time series will reveal itself through the organization and structure of the optimal path.
It is important to stress the non-local nature of the optimization problem (4), as the best path from an origin to an end point requires knowledge of the distance matrix (energy landscape) E X , Y both to the left as well as to the right of any point in the plane (t 1 , t 2 ). There is a general and powerful method to solve this problem in polynomial time using the transfer matrix method (Derrida et al. 1978, Derrida and Vannimenus 1983). Figure 1 shows a realization of the distance (or energy) landscape E X , Y given by (1) and the corresponding optimal path.
Published online:
18 February 2007Figure 1. An example of energy landscape E X , Y given by (1) for two noisy time series and the corresponding optimal path wandering at the bottom of the valley similarly to a river. This optimal path defines the mapping t 1→t 2 = φ(t 1).
Figure 1. An example of energy landscape E X , Y given by (1) for two noisy time series and the corresponding optimal path wandering at the bottom of the valley similarly to a river. This optimal path defines the mapping t 1→t 2 = φ(t 1).
The transfer matrix method can be formulated as follows. Figure 2 shows the (t 1 , t 2 ) plane and defines the notations. Note in particular that the optimal path for two identical time series is the main diagonal, so deviations from the diagonal quantify lag or lead times between the two time series. It is thus convenient to introduce a rotated frame (t,x) as shown in figure 2 such that the second coordinate x quantifies the deviation from the main diagonal, hence the lead or lag time between the two time series. In general, the optimal path is expected to wander around, above or below the main diagonal of equation x(t)=0. The correspondence between the initial frame (t 1 , t 2 ) and the rotated frame (t,x) is given in detail in the Appendix.
Published online:
18 February 2007Figure 2. Representation of the lattice (t 1 , t 2 ) and of the rotated frame ( t, x ) as defined in the text and the Appendix. We refer to the (t 1 , t 2 ) coordinate system as the □-system (square system). We refer to the ( x, t ) coordinate system as the ◃-system (triangle system). The three arrows depict the three moves that are allowed from any node in one step, in accordance with the continuity and monotonicity conditions (3).
Figure 2. Representation of the lattice (t 1 , t 2 ) and of the rotated frame ( t, x ) as defined in the text and the Appendix. We refer to the (t 1 , t 2 ) coordinate system as the □-system (square system). We refer to the ( x, t ) coordinate system as the ◃-system (triangle system). The three arrows depict the three moves that are allowed from any node in one step, in accordance with the continuity and monotonicity conditions (3).
The optimal path (and thus mapping) is constructed such that it can either go horizontally by one step from (t 1,t 2) to (t 1+1, t 2), vertically by one step from (t 1,t 2) to (t 1,t 2+1) or along the diagonal from (t 1,t 2) to (t 1+1,t 2+1). The restriction to these three possibilities embodies the continuity condition (3) and the one-to-one mapping (for vertical segments the one-to-one correspondence is ensured by the convention to map t 1 to the largest value t 2 of the segment). A given node (t 1 , t 2 ) in a two-dimensional lattice carries the ‘potential energy’ or distance ε (t 1,t 2). Let us now denote E(t 1,t 2) as the energy (cumulative distance (4)) of the optimal path starting from some origin (t 1,0,t 2,0) and ending at (t 1 , t 2 ). The transfer matrix method is based on the following fundamental relation:
The global minimization procedure is fully determined once the starting and ending points of the paths are defined. Since the lag–leads between two time series can be anything at any time, we allow the starting point to lie anywhere on the horizontal axis t 2 = 0 or on the vertical axis t 1 = 0. Similarly, we allow the ending point to lie anywhere on the horizontal axis t 2=N−1 or on the vertical axis t 1=N−1. This allows for the fact that one of the two time series may precede the other one. For each given pair of starting and ending points, we obtain a minimum path (the ‘optimal directed polymer’ with fixed end-points). The minimum energy path over all possible starting and ending points is then the solution of our global optimization problem (4) under constraint (3). This equation of this global optimal path defines the mapping t 1→t 2=φ(t 1) defining the dependence relationship between the two time series.
2.3. Optimal path at finite temperature
While appealing, the optimization program (4) under constraint (3) has an important potential drawback: it assumes that the distance matrix E X , Y between the time series X to Y defined by (1) is made only of useful information. But, in reality, the time series X(t 1) and Y(t 2) can be expected to contain a significant amount of noise or more generally irrelevant structures stemming from random realizations. Then, the distance matrix E X , Y contains a possible significant amount of noise, or in other words irrelevant patterns. Therefore, the global optimal path obtained from the procedure of section 2.2 is bound to be delicately sensitive in its conformation to the specific realizations of the noises of the two time series. Other realizations of the noises decorating the two time series would lead to different distance matrices and thus different optimal paths. In the case where the noises dominates, this question amounts to investigating the sensitivity of the optimal path with respect to changes in the distance matrix. This problem has actually be studied extensively in the statistical physics literature (see Halpin-Healy and Zhang (1995) and references therein). It has been shown that small changes in the distance matrix may lead to very large jumps in the optimal path, when the distance matrix is dominated by noise. Clearly, these statistical properties would lead to spurious interpretation of any dependence relationship between the two time series. We thus need a method which is able to distinguish between truly informative structure and spurious patterns due to noise.
In a realistic situation, we can hope for the existence of coherent patterns in addition to noise, so that the optimal path can be ‘trapped’ by these coherent structures in the energy landscape. Nevertheless, the sensitivity to specific realizations of the noise of the two time series may lead to spurious wandering of the optimal path, which does not reflect any genuine lag–lead structure. We thus propose a modification of the previous global optimization problem to address this question and make the determination of the mapping more robust and less sensitive to the existence of noise decorating the two time series. Of course, it is in general very difficult to separate the noise from the genuine signal, in the absence of a parametric model. The advantage of the method that we now propose is that it does not require any a priori knowledge of the underlying dynamics.
The idea of the ‘optimal thermal causal path’ method is the following. Building on the picture of the optimal path as being the conformation of a polymer or of a line minimizing its energy E in a frozen energy landscape determined by the distance matrix, we now propose to allow ‘thermal’ excitations or fluctuations around this path, so that path configurations with slightly larger global energies are allowed with probabilities decreasing with their energy. We specify the probability of a given path configuration with energy ΔE above the absolute minimum energy path by a multivariate logit model or equivalently by a so-called Boltzmann weight proportional to exp[−ΔE/T], where the ‘temperature’ T quantifies how much deviation from the minimum energy is allowed. For T→0, the probability for selecting a path configuration of incremental energy ΔE above the absolute minimum energy path goes to zero, so that we recover the previous optimization problem ‘at zero temperature’. Increasing T allows us to sample more and more paths around the minimum energy path. Increasing T thus allows us to wash out possible idiosyncratic dependencies of the path conformation on the specific realizations of the noises decorating the two time series. Of course, for too large temperatures, the energy landscape or distance matrix becomes irrelevant and one loses all information in the lag–lead relationship between the two time series. There is thus a compromise as usual between not extracting too much from the spurious noise (not too small T) and washing out too much of the relevant signal (too high T). Increasing T allows one to obtain an average ‘optimal thermal path’ over a larger and larger number of path conformations, leading to more robust estimates of the lag–lead structure between the two time series. The optimal thermal path for a given T is determined by a compromise between low energy (associated with paths with high Boltzmann probability weight) and large density (large number of contributing paths of similar energies as larger energies are sampled). This density of paths contributing to the definition of the optimal thermal path can be interpreted as an entropic contribution added to the pure energy contribution of the optimization problem of section 2.2. In a sense, the averaging over the thermally selected path configurations provides an effective way of averaging over the noise realizations of the two time series, without actually having to resample the two times series. This intuition is confirmed by our tests below which show that the signal-over-noise ratio is indeed increased significantly by this ‘thermal’ procedure.
Let us now describe how we implement this idea. It is convenient to use the rotated frame ( t, x ) as defined in figure 2, in which t gives the coordinate along the main diagonal of the (t 1 , t 2 ) lattice and x gives the coordinate in the transverse direction from the main diagonal. Of course, the origin (t 1=0,t 2=0) corresponds to (x=0,t=0). Note that the constraint that the path is directed allows us to interpret t as an effective time and x as the position of a path at that ‘time’ t. Then, the optimal thermal path trajectory ⟨x(t)⟩ is obtained by the following formula
emanating from (0, 0) and ending at ( x, t ) and G ◃(t) = ∑ x G ◃(x,t). In statistical physics, G ◃(x,t) is called the partition function constrained to x while G ◃(t) is the total partition function at t. Then, G ◃(x,t)/G ◃(t) is nothing but the probability for a path to be at x at ‘time’ t. Thus, expression (6) indeed defines ⟨ x⟩ as the (thermal) average position at time t. It is standard to call it ‘thermal average’ because G is made up of the Boltzmann factors that weight each path configuration. The intuition is to imagine the polymer/path as fluctuating randomly due to random ‘thermal kicks’ in the quenched random energy landscape. In the limit where the temperature T goes to zero, G ◃(x,t)/G ◃(t) becomes the Dirac function δ[x-xDP(t)] where xDP(t) is the position of the global optimal path determined previously in section 2.2. Thus, for T→0, expression (6) leads to ⟨x⟩ = xDP(t), showing that this thermal procedure generalizes the previous global optimization method. For non-vanishing T, the optimal thermal average ⟨x(t)⟩ given by (6) takes into account the set of neighbouring (in energy) paths, which allows one to average out the noise contribution to the distance matrix. The Appendix gives the recursion relation that allows us to determine G ◃(x,t). This recursion relation uses the same principle and thus has the same structure as expression (5) (Wang et al. 2000).
Similarly to expression (6), the variance of the trajectory of the optimal thermal path reads
gives a measure of the uncertainty in the determination of the thermal optimal path and thus an estimate of the error in the lag–lead structure of the two time series as seen from this method.
3. Numerical tests on simple examples
3.1. Construction of the numerical example
We consider two stationary time series X(t 1) and Y(t 2), and construct Y(t 2) from X(t 1) as follows:
The time series X(t 1) itself is generated from an AR process:
and
In our simulations, we take τ = 5, a = 0.8, b = 0.7 and σξ =1 and consider time series of duration N = 100.
For a given f, we obtain the optimal zero-temperature path by using the transfer-matrix method (5) explained in section 2.2 for 19 different starting positions around the origin and similarly 19 different ending positions around the upper-right corner at coordinate (99,99). This corresponds to solving 19×19 transfer matrix optimization problems. The absolute optimal path is then determined as the path which has the smallest energy over all these possible starting and ending points. We also determine the optimal thermal paths ⟨x(t)⟩, for different temperatures, typically from T=1/5 to 10, using relation (A3a) for the partition function and the definition (B1a) for the average transverse path trajectory (given in the Appendix).
Figure 3(a) shows that transverse trajectory x(t) as a function of the coordinate t along the main diagonal for f=1/10 and for temperatures T = 0, 1/5, 1 and 10. This graph corresponds to the case where we restrict our attention to paths with fixed imposed starting (origin) and ending (coordinates (99,99) on the main diagonal) points. This restriction is relaxed as we explain above and is applied below to prevent the boundary effects clearly visible in figure 3(a). Figure 3(b) shows the corresponding standard deviation defined by (7) of the thermal average paths.
Published online:
18 February 2007Figure 3. (a) Thermal average ⟨x(t)⟩ of the transverse fluctuations with respect to t for T = 10, 1 and 1/5 and the directed polymer. (b) The uncertainty σ x of the thermal average paths for different temperatures. All the paths are constrained to start from the diagonal (t 1=0,t 2=0) and to return to it at (t 1=99,t 2=99).
Figure 3. (a) Thermal average ⟨x(t)⟩ of the transverse fluctuations with respect to t for T = 10, 1 and 1/5 and the directed polymer. (b) The uncertainty σ x of the thermal average paths for different temperatures. All the paths are constrained to start from the diagonal (t 1=0,t 2=0) and to return to it at (t 1=99,t 2=99).
The impact of the temperature is nicely illustrated by plotting how the energy of an optimal thermal path depends on its initial starting point x(0)=x 0 (and ending point taken with the same value x(99)=x(0)). For a given x 0 and temperature T, we determine the thermal optimal path and then calculate its energy eT (x 0) by the formula
Published online:
18 February 2007Figure 4. Dependence of the thermal average energy eT (x 0) of the optimal thermal path as a function of the average ⟨x(x0)⟩ defined in turn by the coordinate of its starting point (t=|x 0|,x=x 0) for different temperatures given by T=1/50, 1/5, 1/2, 1, 2, 5 and 10 from bottom to top and for f=1/2.
Figure 4. Dependence of the thermal average energy eT (x 0) of the optimal thermal path as a function of the average ⟨x(x0)⟩ defined in turn by the coordinate of its starting point (t=|x 0|,x=x 0) for different temperatures given by T=1/50, 1/5, 1/2, 1, 2, 5 and 10 from bottom to top and for f=1/2.
Figure 5 further demonstrates the role of the temperature for different amplitudes of the noise η. It shows the position as a function of T for different relative noise levels f. Recall that ⟨x(t)⟩ is the optimal thermal position of the path for a fixed coordinate t along the main diagonal, as defined in (6). The symbol
expresses an additional average of ⟨x⟩ over all the possible values of the coordinate t: in other words,
is the average elevation (or translation) of the optimal thermal path above (or below) the diagonal. This average position is an average measure (along the time series) of the lag/lead time between the two time series, assuming that this lag/lead time is the same for all times. In our numerical example, we should obtain
close to or equal to τ = 5. Figure 5 shows the dependence of
as a function of T for different values of f.
Published online:
18 February 2007Figure 5. Dependence of upon noise level f and temperature T.
Figure 5. Dependence of upon noise level f and temperature T.
Obviously, with the increase of the signal-to-noise ratio of the realizations which is proportional to 1/f, the accuracy of the determination of τ improves. For a noise level f, approaches the correct value τ = 5 with increasing T. The beneficial impact of the temperature is clearer for more noisy signals (larger f). It is interesting to notice that an ‘optimal range’ of temperature appears for large noise level.
3.2. Test on the detection of jumps or change-of-regime in time lag
We now present synthetic tests of the efficiency of the optimal thermal causal path method to detect multiple changes of regime and compare the results with a standard correlation analysis performed in moving windows of different sizes. Consider the following model
Figure 6 shows the standard cross-correlation function calculated over the whole time interval 1≤i≤250 of the two time series X and Y given by (12), so as to compare with our method. Without further information, it would be difficult to conclude more than to say that the two time series are rather strongly correlated at zero time lag. It would be farfetched to associate the tiny secondary peaks of the correlation function at τ =±5 and 10 to genuine lags or lead times between the two time series. And since, the correlation function is estimated over the whole time interval, the time localization of possible shifts of lag/leads is impossible.
Published online:
18 February 2007Figure 6. Standard cross-correlation function of two time series with varying time lags τ = 0, 10, 5, −5 and 0 as defined in equation (12).
Figure 6. Standard cross-correlation function of two time series with varying time lags τ = 0, 10, 5, −5 and 0 as defined in equation (12).
Before presenting the results of our method, it is instructive to consider a natural extension of the correlation analysis, which consists of estimating the correlation function in a moving window [i+1−D,i] of length D, where i runs from D to 250. We then estimate the lag–lead time τ D (i) as the value that maximizes the correlation function in each window [i+1−D,i]. We have used D = 10, 20, 50 and 100 to investigate different compromises (D = 10 is reactive but does not give statistically robust estimates while D = 100 gives statistically more robust estimates but is less reactive to abrupt changes of lag). The local lags τ D (i) thus obtained are shown in figure 7 as a function of the running time i. For D = 10, this method identifies successfully the correct time lags in the first, third, fourth and fifth time periods, while τ D (i) in the second time period is very noisy and fails to unveil the correct value τ = 10. For D = 20, the correct time lags in the five time periods are identified with large fluctuations at the boundaries between two successive time periods. For D = 50, five successive time lags are detected but with significant delays compared to their actual inception times and with, in addition, high interspersed fluctuations. For D = 100, the delays of the detected inception times of each period reach about 50 time units, that is, comparable to the width of each period, and the method fails completely for this case.
Published online:
18 February 2007Figure 7. Local cross-correlation analysis of the two time series defined by (12) with (9) using moving windows of sizes D = 10, 20, 50 and 100. The value τ D (i) of the lag that makes the local cross-correlation function in each window [i+1−D,i] a maximum is plotted as a function of the right-end time i. The true time lags as defined in (12) are respectively τ = 0, 10, 5, −5 and 0 in five successive time periods of 50 time steps each.
Figure 7. Local cross-correlation analysis of the two time series defined by (12) with (9) using moving windows of sizes D = 10, 20, 50 and 100. The value τ D (i) of the lag that makes the local cross-correlation function in each window [i+1−D,i] a maximum is plotted as a function of the right-end time i. The true time lags as defined in (12) are respectively τ = 0, 10, 5, −5 and 0 in five successive time periods of 50 time steps each.
Let us now turn to our optimal thermal causal path method. We determine the average thermal path (transverse trajectory x(i) as a function of the coordinate i along the main diagonal) starting at the origin, for four different temperatures T = 2, 1, 1/2 and 1/5. Figure 8 plots x(i) as a function of i. The time lags in the five time periods are recovered clearly. At the joint points between the successive time periods, there are short transient crossovers from one time lag to the next. Our new method clearly outperforms the above cross-correlation analysis.
Published online:
18 February 2007Figure 8. Average thermal path (transverse trajectory x(i) as a function of the coordinate i along the main diagonal) starting at the origin, for four different temperatures (T = 2 (dotted-dash), T = 1 (dotted), T=0.5 (dashed) and 0.2 (continuous)) obtained by applying the optimal thermal causal path method to the synthetic time series (12) with (9).
Figure 8. Average thermal path (transverse trajectory x(i) as a function of the coordinate i along the main diagonal) starting at the origin, for four different temperatures (T = 2 (dotted-dash), T = 1 (dotted), T=0.5 (dashed) and 0.2 (continuous)) obtained by applying the optimal thermal causal path method to the synthetic time series (12) with (9).
The advantage of our new method compared with the moving cross-correlation method for two time series with varying time lags can be further illustrated by a test of predictability. It is convenient to use an example with unidirectional dependence lags (only positive lags) and not with bidirectional jumps as exemplified by (12). We thus consider a case in which X leads Y in general and use the following model
4. Applications to economics
4.1. Revisiting the dependence between the US treasury bond yield and the stock market antibubble since August 2000
In a recent paper (Zhou and Sornette 2004), we have found evidence for the following dependence in the time period from October 2000 to September 2003: stock market → Fed Reserve (Federal funds rate) → short-term yields → long-term yields (as well as a direct and instantaneous influence of the stock market on the long-term yields). These conclusions were based on (1) lagged cross-correlation analysis in running windows and (2) the dependence of the parameters of a ‘log-periodic power law’ calibration to the yield time series at different maturities (see Sornette and Johansen (2001), Sornette and Zhou (2002) and Sornette (2003) for a recent exposition of the method and synthesis of the main results on a variety of financial markets).
Let us now revisit this question by using the optimal thermal causal path method. The data consist of the S & P 500 index, the Federal Funds rate (FFR) and ten treasury bond yields spanning three years from 9 September 2000 to 9 September 2003. The optimal thermal paths, x(i)'s, of the distance matrix between the monthly returns of the S & P 500 index with each of the monthly relative variations of the eleven yields are determined for a given temperature T, giving the corresponding lag–lead times τ(i)=x(i)'s as a function of present time i. Figure 9 shows these τ( i )'s for T = 1, where positive values correspond to the yields lagging behind or being caused by the S & P 500 index returns. The same analysis was performed also for T = 10, 5, 2, 1, 1/2 and 1/5, yielding a very consistent picture, confirming indeed that τ is positive for short-term yields and not significantly different from zero for long-term yields, as shown in figure 9. One can also note that the lag τ( i ) seems to have increased with time from September 2000 to a peak in the last quarter of 2003.
Published online:
18 February 2007Figure 9. Instantaneous lags between the S & P 500 index and the Federal fund rate (FFR), and between the S & P 500 index and each of ten treasury bond yields, calculated using the optimal thermal causal path method at temperature T = 1 using monthly returns for the S & P 500 index and monthly relative variations for the yields. Positive lags correspond to the yields lagging behind the S & P 500 index.
Figure 9. Instantaneous lags between the S & P 500 index and the Federal fund rate (FFR), and between the S & P 500 index and each of ten treasury bond yields, calculated using the optimal thermal causal path method at temperature T = 1 using monthly returns for the S & P 500 index and monthly relative variations for the yields. Positive lags correspond to the yields lagging behind the S & P 500 index.
We also performed the same analysis with weakly and quarterly data of the returns and yield changes. The results (not shown) confirm the results obtained at the monthly time scale. This analysis seems to confirm the existence of a change of regime in the sign of the lag of the dependence between the S & P 500 index and the Federal Funds rate: it looks as if the Fed (as well as the short term yields) started to be influenced by the stock market after a delay following the crash in 2000, waiting until mid-2001 for the causality to be revealed. The positivity of the time lag suggests that the yields follow (are caused by ?) the stock index. This phenomenon is consistent with the evidence previously presented by Zhou and Sornette (2004) and thus provides further evidence on the dependence flowing from the stock market to the treasury yields. The instantaneous lag–lead functions τ( t ) actually provide much clearer signatures of the dependence structure than our previous analysis: compare for instance with the cross-correlation coefficient shown in figure 10 of Zhou and Sornette (2004). From an economic view point, we interpret this lag relationship as meaning that the FRB is causally influenced by the stock market (at least for the studied period); it should be cautioned however that we have not proved causality stricto sensu, only the existence of a clear lagged relationship between FRB and the stock market. Persuing our tentative interpretation, we take the established relationship as an indication that the stock markets are considered as proxies of the present and are conditioning the future health of the economy, according to the FRB model of the US economy. In a related study, causality tests performed by Lamdin (2004) also confirm that stock market movements precede changes in yield spread between corporate bonds and government bonds. Abdulnasser and Manuchehr (2002) have also found that Granger causality is unidirectionally running from stock prices to effective exchange rates in Sweden.
Published online:
18 February 2007Figure 10. Data used in our analysis, that is, the normalized inflation rate, its normalized quarterly change, and the normalized GDP growth rate from 1947 to 2003.
Figure 10. Data used in our analysis, that is, the normalized inflation rate, its normalized quarterly change, and the normalized GDP growth rate from 1947 to 2003.
4.2. Are there any dependence relationships between inflation and gross domestic product (GDP) and inflation and unemployment in the USA?
The relationship between inflation and real economic output quantified by GDP has been discussed many times in the last several decades. Different theories have suggested that the impact of inflation on the real economy activity could be either neutral, negative or positive. Based on Mundell's (1963) story that higher inflation would lower real interest rates, Tobin (1965) argued that higher inflation causes a shift from money to capital investment and raises output per capita. Conversely, Fischer (1974) suggested a negative effect, stating that higher inflation resulted in a shift from money to other assets and reduced the efficiency of transactions in the economy due to higher search costs and lower productivity. In the middle, Sidrauski (1967) proposed a neutral effect where exogenous time preference fixed the long-run real interest rate and capital intensity. These arguments are based on the rather restrictive assumption that the Phillips curve (inverse relationship between inflation and unemployment), taken in addition to be linear, is valid.
To evaluate which model characterizes better real economic systems, numerous empirical efforts have been performed. Fama (1982) applied the money demand theory and the rational expectation quality theory of money to the study of inflation in the USA and observed a negative relation during the post-1953 period. Barro (1995) used data for around 100 countries from 1960 to 1990 to assess the effects of inflation on economic output and found that an increase in average inflation led to a reduction of the growth rate of real per capita GDP, conditioned by the fact that the inflation was high. Fountas et al. (2002) used a bivariate GARCH model of inflation and output growth and found evidence that higher inflation and more inflation uncertainty led to lower output growth in the Japanese economy. Apergis (2004) found that inflation causally affected output growth using univariate GARCH models with a panel set for the G7 countries.
Although cross-country regressions explain that output growth often obtains a negative effect from inflation, Ericsson et al. (2001) argued that these results are not robust and demonstrated that annual time series of inflation and the log-level of output for most G7 countries are co-integrated, thus rejecting the existence of a long-run relation between output growth and inflation. A causality analysis using annual data from 1944 to 1991 in Mexico performed by Shelley and Wallace (2004) showed that it is important to separate the changes in inflation into predictable and unpredictable components whose differences respectively had a significant negative and positive effect on real GDP growth. Huh (2002) and Huh and Lee (2002) utilized a vector autoregression (VAR) model to accommodate the potentially important departure from linearity of the Phillips curve motivated by a strand of theoretical and empirical evidence in the literature suggesting nonlinearity in the output–inflation relationship. The empirical results indicated that their model captured the nonlinear features present in the data in Australia and Canada. This study implies that there might exist a nonlinear causality from inflation to economic output. It is therefore natural to use our novel method to detect possible local nonlinear lagged-dependence relationships.
Our optimal thermal causal path method is applied to the GDP quarterly growth rates paired with the inflation rate updated every quarter on the one hand and with the quarterly changes of the inflation rates on the other hand, for the period from 1947 to 2003 in the USA. The GDP growth rate, the inflation rate and and the inflation rate changes have been normalized by their respective standard deviations. The inflation and inflation changes are calculated from the monthly customer price index (CPI) obtained from the Fed II database (federal reserve bank). Eight different temperatures T=50, 20, 10, 5, 2, 1, 1/2 and 1/5 have been investigated.
Figure 10 shows the data used for the analysis, that is, the normalized inflation rate, its normalized quarterly change and the normalized GDP growth rate from 1947 to 2003.
Figure 11 shows the lag–lead times τ(t)=x(t)'s (units in year) for the pair (inflation, GDP growth) as a function of present time t for T = 2 and for 19 different starting positions (and their ending counterparts) in the (t 1 , t 2 ) plane, where positive values of τ(t)=x(t) correspond to the GDP lagging behind or being caused by inflation. This figure is representative of the information at all the investigated temperatures. Overall, we find that τ is negative in the range −2 years ≤τ≤0 year, indicating that it is more the GPD which leads inflation than the reverse. However, this broad-brush conclusion must be toned down somewhat at a finer time resolution as two time periods can be identified in figure 11.
-
From 1947 (and possibly earlier) to the early 1980s, one can observe two clusters, one with negative −2 years ≤τ = x(t)≤0 years implying that the GDP has a positive lagged effect on future inflation, and another with positive 0 years ≤τ = x(t)≤4 years implying that inflation has a lagged effect on GDP with a longer lag.
-
From the mid-1980s to the present, there is no doubt that it is GDP which has had the dominating impact on future inflation lagging behind by about 1–2 years.
Published online:
18 February 2007Figure 11. Lag–lead times τ(t)=x(t)'s (units in year) for the pair (inflation, GDP) as a function of present time t for T = 2 and for 19 different starting positions (and their ending counterparts) in the (t 1 , t 2 ) plane, where positive values of τ(t)=x(t) correspond to the GDP lagging behind or being caused by inflation. The dashed blue line is the optimal path with the minimal ‘energy’.
Figure 11. Lag–lead times τ(t)=x(t)'s (units in year) for the pair (inflation, GDP) as a function of present time t for T = 2 and for 19 different starting positions (and their ending counterparts) in the (t 1 , t 2 ) plane, where positive values of τ(t)=x(t) correspond to the GDP lagging behind or being caused by inflation. The dashed blue line is the optimal path with the minimal ‘energy’.
In summary, our analysis suggests that the interaction between GDP and inflation is more subtle than previously discussed. Perhaps past controversies in which one causes the other may be due to the fact that, to a certain degree, these different time series have entangled dynamics, with different time lags. Any measure of a dependence relationship allowing for only one lag is bound to miss such subtle interplay. It is interesting to find that GDP impacts on future inflation with a relatively small delay of about one year while inflation has in the past influenced future GDP with a longer delay of several years.
Figure 12 shows the lag–lead times τ(t)=x(t)'s (units in year) for the pair (inflation change, GDP) as a function of present time t for T = 2 and for 19 different starting positions (and their ending counterparts) in the (t 1 , t 2 ) plane, where positive values of τ(t)=x(t) correspond to the GDP lagging behind or being caused by inflation change. Due to the statistical fluctuations, we cannot conclude on the existence of a significant dependence relationship between inflation change and GDP, except in the decade of the 1980s for which there is a strong causal effect of a change of inflation on GDP. The beginning of this decade was characterized by a strong decrease of the inflation rate from a two-digit value in 1980, following a vigorous monetary policy implemented under the Fed's chairman Paul Volker. The end of the 1970s and the better half of the 1980s were characterized by an almost stagnant GDP. In the mid-1980s, the GDP started to grow again at a strong pace. It is probably this lag between the significant reduction of inflation in the first half of the 1980s and the rise of the GDP growth that we detect here. Our analysis may help in improving our understanding in the intricate relationship between different economic variables and their impact on growth and on stability and in addressing the difficult problem of model errors, which Cogley and Sargent (2005) have argued to be the cause for the lack of significant action from the Fed in the 1970s.
Published online:
18 February 2007Figure 12. Same as figure 11 for the pair (inflation change, GDP). Positive values of τ(t)=x(t) correspond to the GDP lagging behind or being caused by inflation change. The dashed blue line is the optimal path with the minimal ‘energy’.
Figure 12. Same as figure 11 for the pair (inflation change, GDP). Positive values of τ(t)=x(t) correspond to the GDP lagging behind or being caused by inflation change. The dashed blue line is the optimal path with the minimal ‘energy’.
Figure 13 shows the lag–lead times τ(t)=x(t)'s (units in year) for the pair (inflation, unemployment rate) as a function of present time t for T = 2 and for 19 different starting positions (and their ending counterparts) in the (t 1 , t 2 ) plane, where positive values of τ(t)=x(t) correspond to the unemployment rate lagging behind or being caused by inflation. We use quarterly data from 1948 to 2004 obtained from the Fed II database (federal reserve bank). This figure is representative of the information at all the investigated temperatures.
-
From 1947 (and possibly earlier) to 1970, one can observe large fluctuations with two clusters, suggesting a complex dependence relationship between the two time series, similarly to the situation discussed above for the (inflation, GDP) pair.
-
From 1970 to the present, there is no doubt that inflation has predated unemployment in the sense of the optimal thermal causal path method. It is also noteworthy that the lag between unemployment and inflation has disappeared in recent years. From a visual examination of figure 13, we surmise that what is detected is probably related to the systematic lags between inflation and employment in the four large peak pairs: (1970 for inflation; 1972 for employment), (1975 for inflation; 1976 for unemployment), (1980 for inflation; 1983 for unemployment) and (1991 for inflation; 1993 for unemployment).
Published online:
18 February 2007Figure 13. Same as figure 11 for the pair (inflation, unemployment rate). Positive values of τ(t)=x(t) correspond to the unemployment lagging behind or being caused by inflation. The dashed line is the optimal path with the minimal ‘energy’.
Figure 13. Same as figure 11 for the pair (inflation, unemployment rate). Positive values of τ(t)=x(t) correspond to the unemployment lagging behind or being caused by inflation. The dashed line is the optimal path with the minimal ‘energy’.
5. Concluding remarks
In summary, we have developed a novel method for the detection of lagged-dependence structures between two time series, based on the search for a robust optimal path in a distance matrix. Our optimal thermal causal path method determines the thermal average paths emanating from different starting lag–lead times in the distance matrix constructed from the two original time series and choose the one with minimal average mismatch (‘energy’). The main advantage of our method is that it enables us to detect lagged dependence locally and is thus particularly useful when the dependence relation is nonlinear and changes intermittently. An advantage of the method is that it is robust with respect to noise, i.e. it does not attribute a dependence relationship between two time series from patterns in the distance matrix that may arise randomly. This robustness is acquired by using the ‘thermal’ averaging procedure which provides a compromise between optimizing the matching between the two time series and maximizing the local density of optimal paths to ensure a strong relationship.
We have applied this method to the stock market and treasury bond yields and confirmed our earlier results in Zhou and Sornette (2004) on a lagged arrow of the stock markets preceding the Federal Reserve Funds’ adjustments as well as the yield rates at short maturities. Another application to the inflation and GDP growth rate and to unemployment have unearthed non-trivial lagged-dependence relationships: the GDP changes have led inflation especially since the 1980s, whereas inflation changes lead GDP only in the 1980s, but inflation has led unemployment rates since the 1970s.
Our approach seems to detect multiple competing dependence paths with intertwined arrows of dependence in which one can have inflation leading GDP with a certain lag time and GDP feeding back/leading inflation with another lag time. This suggests that the predictive skills of models with one-way lagged-dependence structure are fundamentally limited and more elaborate measurements as proposed here and models with complex feedbacks are necessary to account for the multiple lagged feedback mechanisms present in the economy.
Acknowledgments
We are grateful to X.-H. Wang for fruitful discussion and to N. Marwan for the permission of the use of his MATLAB programs (http://www.agnld.uni-potsdam.de) at the early stage of this work.
References
- Abdulnasser, H-J and Manuchehr, I. 2002. On the causality between exchange rates and stock prices: a note. Bull. Econ. Res., 54: 197–203. [Google Scholar]
- Apergis, N. 2004. Inflation, output growth, volatility and causality: evidence from panel data and the G7 countries. Econ. Lett., 83: 185–191. [Crossref], [Web of Science ®], [Google Scholar]
- Ashley, R, Granger, CWJ and Schmalensee, R. 1980. Advertising and aggregate consumption: an analysis of causality. Econometrica, 48: 1149–1167. [Crossref], [Web of Science ®], [Google Scholar]
- Barro, RJ. 1995. Inflation and economic growth. Bank of England Q. Bull., 35: 166–175. [Google Scholar]
- Bouchaud, J-P, Mézard, M and Yededia, J. 1991. Phys. Rev. Lett., 67: 3840–3843. [Google Scholar]
- Chamberlain, G. 1982. The general equivalence of Granger and Sims causality. Econometrica, 50(3): 569–582. [Google Scholar]
- Chen, Y, Rangarajan, G, Feng, J and Ding, M. 2004. Analyzing multiple nonlinear time series with extended Granger causality. Phys. Lett. A, 324: 26–35. [Crossref], [Web of Science ®], [Google Scholar]
- Cogley, T and Sargent, TJ. 2005. The conquest of US inflation: learning and robustness to model uncertainty. Rev. Econ. Dynamics, 8: 528–563. [Crossref], [Web of Science ®], [Google Scholar]
- Derrida, B, Vannimenus, J and Pomeau, Y. 1978. Simple frustrated systems: chains, strips and squares. J. Phys. C, 11: 4749–4765. [Crossref], [Web of Science ®], [Google Scholar]
- Derrida, B and Vannimenus, J. 1983. Interface energy in random systems. Phys. Lett. B, 27: 4401–4411. [Google Scholar]
- Drasdo, D, Hwa, T and Lassig, M. 1998. A statistical theory of sequence alignment with gaps. Proceedings of The Sixth International Conference on Intelligent Systems for Molecular Biology (ISMB98). 1998. Edited by: Glasnow, J, Littlejohn, I, Major, F, Lathrop, R, Sankoff, D and Sensen, C. pp.52–58. Menlo Park: AAAI Press. edited by [Google Scholar]
- Drasdo, D, Hwa, T and Lassig, M. 2001. Scaling laws and similarity detection in sequence alignment with gaps. J. Comput. Biol., 7: 115–141. [Google Scholar]
- Eckmann, J-P, Kamphorst, SO and Ruelle, D. 1987. Recurrence plots of dynamical systems. Europhys. Lett., 4: 973–977. [Crossref], [Google Scholar]
- Ericsson, NR, Irons, JS and Tryon, RW. 2001. Output and inflation in the long run. J. Appl. Econ., 16: 241–253. [Crossref], [Web of Science ®], [Google Scholar]
- Fama, EF. 1982. Inflation, output, and money. J. Bus., 55: 201–231. [Crossref], [Google Scholar]
- Fischer, S. 1974. Money and production function. Econ. Inquiry, 12: 517–533. [Google Scholar]
- Fountas, S, Karanasos, M and Kim, J. 2002. Inflation and output growth uncertainty and their relationship with inflation and output growth. Econ. Lett., 75: 293–301. [Crossref], [Web of Science ®], [Google Scholar]
- Geweke, J. 1984. “Inference and causality in economic time series models”. In Handbook of Economics, Edited by: Griliches, Z and Intriligator, MD. Vol. II, 1101–1144. Amsterdam: Elsevier Science Publisher BV. edited by, Chapter 19 [Google Scholar]
- Gillman, M and Kejak, M. 2005. Contrasting models of the effect of inflation on growth. J. Econ. Surveys, 19(1): 113–136. [Crossref], [Web of Science ®], [Google Scholar]
- Granger, CWJ and Jeon, Y. 1997. Measuring lag structure in forecasting models—the introduction of time distance. Discussion Paper 97-24. 1997. San Diego: University of California. [Google Scholar]
- Halpin-Healy, T and Zhang, Y-C. 1995. Kinetic roughening phenomena, stochastic growth directed polymers and all that. Phys. Rep., 254: 215–415. [Crossref], [Web of Science ®], [Google Scholar]
- Huh, H-S. 2002. Estimating asymmetric output cost of lowering inflation for Australia. Southern Econ. J., 68: 600–616. [Google Scholar]
- Huh, H-S and Lee, H-H. 2002. Asymmetric output cost of lowering inflation: empirical evidence for Canada. Can. J. Econ., 35: 218–238. [Google Scholar]
- Hwa, T and Lassig, M. 1996. Similarity-detection and localization. Phys. Rev. Lett., 76: 2591–2594. [Google Scholar]
- Hwa, T and Lassig, M. 1998. Optimal detection of sequence similarity by local alignment. Proceedings of the Second Annual International Conference on Computational Molecular Biology (RECOMB98). 1998. Edited by: Istrail, S, Pevzner, P and Waterman, MS. pp.109–116. New York: ACM Press. edited by [Google Scholar]
- Lamdin, DJ. 2004. Corporate bond yield spreads in recent decades: an examination of trends, changes and stock market linkages. Bus. Econ., 39: 28–35. [Google Scholar]
- Lansing, KJ. 2000. Exploring the causes of the great inflation. July 72000. Federal Reserve Bank of San Francisco. 2000-21 [Google Scholar]
- Marwan, N and Kurths, J. 2002. Nonlinear analysis of bivariate data with cross recurrence plots. Phys. Lett. A, 302: 299–307. [Crossref], [Web of Science ®], [Google Scholar]
- Marwan, N, Thiel, M and Nowaczyk, NR. 2002. Cross recurrence plot based synchronization of time series. Nonlinear Process. Geophys., 9: 325–331. [Crossref], [Web of Science ®], [Google Scholar]
- Mundell, R. 1963. Inflation and real interest. J. Politic. Econ., 71: 280–283. [Crossref], [Web of Science ®], [Google Scholar]
- Needleman, SB and Wunsch, CD. 1970. A general method applicable to the search for similarities in the amino acid sequence of two proteins. J. Molec. Biol., 48(3): 443–453. [Crossref], [PubMed], [Web of Science ®], [Google Scholar]
- Olsen, R, Hwa, T and Lassig, M. 1999. Optimizing Smith–Waterman alignments. Pacific Symposium on Biocomputing 4. 1999. pp.302–313. [Google Scholar]
- Quiroga, RQ, Kreuz, T and Grassberger, P. 2002. Event synchronization: a simple and fast method to measure synchronicity and time delay patterns. Phys. Rev. E, 66: 041904 [Crossref], [Web of Science ®], [Google Scholar]
- Sankoff, D and Kruskal, JB. 1983. Time Warps, String Edits, and Macromolecules: The Theory and Practice of Sequence Comparison, Boston: Addison-Wesley. (Eds) [Google Scholar]
- Shelley, GL and Wallace, FH. 2004. Inflation, money, and real GDP in Mexico: a causality analysis. Appl. Econ. Lett., 11: 223–225. [Taylor & Francis Online], [Web of Science ®], [Google Scholar]
- Sidrauski, M. 1967. Rational choice and patterns of growth in a monetary economy. Am. Econ. Rev., 57: 534–544. [Web of Science ®], [Google Scholar]
- Sornette, D. 2003. Why Stock Markets Crash (Critical Events in Complex Financial Systems), Princeton: Princeton University Press. [Google Scholar]
- Sornette, D and Johansen, A. 2001. Significance of log-periodic precursors to financial crashes. Quant. Finance, 1(4): 452–471. [Taylor & Francis Online], [Google Scholar]
- Sornette, D and Zhou, W-X. 2002. The US 2000–2002 market descent: how much longer and deeper?. Quant. Finance, 2: 468–481. [Taylor & Francis Online], [Google Scholar]
- Strozzia, F, Zaldvarb, J-M and Zbilutc, JP. 2002. Application of nonlinear time series analysis techniques to high-frequency currency exchange data. Physica A, 312: 520–538. [Google Scholar]
- Tobin, J. 1965. Money and economic growth. Econometrica, 33: 671–684. [Crossref], [Web of Science ®], [Google Scholar]
- Wang, XH, Havlin, S and Schwartz, M. 2000. Directed polymers at finite temperatures in 1+1 and 2+1 dimensions. J. Phys. Chem. B, 104: 3875–3880. [Google Scholar]
- Zhou, W-X and Sornette, D. 2004. Causal slaving of the US treasury bond yield antibubble by the stock market antibubble of August 2000. Physica A, 337: 586–608. [Google Scholar]
Appendix A: Recursive scheme of partition function
In order to calculate the thermal average position ⟨x(t)⟩ for t=0,1,2,… , over all possible paths in the distance matrix landscape, one needs to determine the values at all nodes of G ◃ (x,t), defined in equation (6) and subsequent paragraph. For clarity, we present firstly the recursive relation in the (t 1 , t 2 ) coordinates and then transform it into the ( x, t ) coordinates. The transformation from the coordinates (t 1 , t 2 ) to ( x, t ) is
geometrically.
If two time series are perfectly related (they are the same up to a factor), then the optimal path is the diagonal, that is, made up of the diagonal bonds of the square lattice, or alternatively the nodes on the diagonals. Since the ‘energy’ (i.e. local mismatch defined by expression (1)) is defined only on the nodes, a path has a Boltzmann weight contributed only by the nodes and there is no contribution from bonds. We should thus allow the path not only along the horizontal and vertical segments of each square of the lattice but also along the main diagonal of each square. The directedness means that a given path is not allowed to go backward on any of the three allowed moves. As illustrated in figure 2, in order to arrive at (t 1+1,t 2+1), the path can come from (t 1+1, t 2) vertically, (t 1,t 2+1) horizontally, or (t 1,t 2) diagonally. The recursive equation on the Boltzmann weight factor is thus
Using the axes transformation (A1), equation (A2a) can be rewritten in the following form
Appendix B: Relations between the two schemes
Consider a t-slide in the ◃-scheme, that is, in the x , t coordinates system. There are t + 1 nodes on the t-slide. For simplicity, we denote the t + 1 partition functions as Gi , i=1,2,…,t+1, and denote . We define two thermal averages of the transverse fluctuations for the t-slide in the □-scheme and the ◃-scheme, respectively: