Lecture Notes in Mathematics Editors: J.-M. Morel, Cachan B. Teissier, Paris
For further volumes: http://www.springer.com/series/304
2046
•
Catherine Donati-Martin Alain Rouault
Antoine Lejay
Editors
Séminaire de Probabilités XLIV
123
Editors Catherine Donati-Martin Université de Versailles-St-Quentin Versailles France
Alain Rouault Université de Versailles-St-Quentin Versailles France
Antoine Lejay Nancy-Université, INRIA Vandoeuvre-lès-Nancy France
ISBN 978-3-642-27461-9 (eBook) ISBN 978-3-642-27460-2 DOI 10.1007/978-3-642-27461-9 Springer Heidelberg New York Dordrecht London Lecture Notes in Mathematics ISSN print edition: 0075-8434 ISSN electronic edition: 1617-9692 Library of Congress Control Number: 2012933110 Mathematics Subject Classification (2010): 60-XX, 60JXX, 60J60, 60J10, 60J65, 60J55, 46L54 c Springer-Verlag Berlin Heidelberg 2012 This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and transmission or information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed. Exempted from this legal reservation are brief excerpts in connection with reviews or scholarly analysis or material supplied specifically for the purpose of being entered and executed on a computer system, for exclusive use by the purchaser of the work. Duplication of this publication or parts thereof is permitted only under the provisions of the Copyright Law of the Publisher’s location, in its current version, and permission for use must always be obtained from Springer. Permissions for use may be obtained through RightsLink at the Copyright Clearance Center. Violations are liable to prosecution under the respective Copyright Law. The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication does not imply, even in the absence of a specific statement, that such names are exempt from the relevant protective laws and regulations and therefore free for general use. While the advice and information in this book are believed to be true and accurate at the date of publication, neither the authors nor the editors nor the publisher can accept any legal responsibility for any errors or omissions that may be made. The publisher makes no warranty, express or implied, with respect to the material contained herein. Printed on acid-free paper Springer is part of Springer Science+Business Media (www.springer.com)
Preface
As usual, some of the contributions to this 44th Séminaire de Probabilités were exposed during the Journées de Probabilités held in Dijon in June 2010. The other ones come from spontaneous submissions or were solicited by the editors. The traditional and historical themes of the Séminaire are present, such as stochastic calculus, local times and excursions and martingales. Some subjects already largely present in the previous volumes are stil here: free probability, rough paths, limit theorems for general processes (here fractional Brownian motion and polymers) and large deviations. Finally, this volume explores new topics, including variable length Markov chains and peacoks. We hope that the whole volume is a good sample of the main streams of current research on probability and stochastic processes, in particular those active in France. We remind that the web site of the Séminaire is http://portail.mathdoc.fr/SemProba/ and that all the articles of the Séminaire from Volume I in 1967 to Volume XXXVI in 2002 are freely accessible from the web site http://www.numdam.org/numdam-bin/feuilleter?j=SPS We thank the Cellule Math Doc for hosting all these articles within the NUMDAM project.
Catherine Donati-Martin Antoine Lejay Alain Rouault
v
•
Contents
Context Trees, Variable Length Markov Chains and Dynamical Sources .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . Peggy Cénac, Brigitte Chauvin, Frédéric Paccaut, and Nicolas Pouyanne Martingale Property of Generalized Stochastic Exponentials . . . . . . . . . . . . . . Aleksandar Mijatovi´c, Nika Novak, and Mikhail Urusov Some Classes of Proper Integrals and Generalized Ornstein–Uhlenbeck Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . Andreas Basse-O’Connor, Svend-Erik Graversen, and Jan Pedersen Martingale Representations for Diffusion Processes and Backward Stochastic Differential Equations . . . . . . . .. . . . . . . . . . . . . . . . . . . . Zhongmin Qian and Jiangang Ying
1
41
61
75
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 105 Markus Mocha and Nicholas Westray The Derivative of the Intersection Local Time of Brownian Motion Through Wiener Chaos . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 141 Greg Markowsky On the Occupation Times of Brownian Excursions and Brownian Loops . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 149 Hao Wu Discrete Approximations to Solution Flows of Tanaka’s SDE Related to Walsh Brownian Motion .. . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 167 Hatem Hajri
vii
viii
Contents
Spectral Distribution of the Free Unitary Brownian Motion: Another Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 191 Nizar Demni and Taoufik Hmidi Another Failure in the Analogy Between Gaussian and Semicircle Laws . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 207 Nathalie Eisenbaum Global Solutions to Rough Differential Equations with Unbounded Vector Fields . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 215 Antoine Lejay Asymptotic Behavior of Oscillatory Fractional Processes .. . . . . . . . . . . . . . . . . . 247 Renaud Marty and Knut Sølna Time Inversion Property for Rotation Invariant Self-similar Diffusion Processes .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 271 Juha Vuolle-Apiala On Peacocks: A General Introduction to Two Articles . .. . . . . . . . . . . . . . . . . . . . 279 Antoine-Marie Bogso, Christophe Profeta, and Bernard Roynette Some Examples of Peacocks in a Markovian Set-Up . . . . .. . . . . . . . . . . . . . . . . . . . 281 Antoine-Marie Bogso, Christophe Profeta, and Bernard Roynette Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 317 Antoine-Marie Bogso, Christophe Profeta, and Bernard Roynette Branching Brownian Motion: Almost Sure Growth Along Scaled Paths.. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 375 Simon C. Harris and Matthew I. Roberts On the Delocalized Phase of the Random Pinning Model . . . . . . . . . . . . . . . . . . . 401 Jean-Christophe Mourrat Large Deviations for Gaussian Stationary Processes and Semi-Classical Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 409 Bernard Bercu, Jean-François Bony, and Vincent Bruneau Girsanov Theory Under a Finite Entropy Condition . . . .. . . . . . . . . . . . . . . . . . . . 429 Christian Léonard Erratum to Séminaire XXVII. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 467 Michel Émery and Marc Yor Erratum to Séminaire XXXV . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 469 Michel Émery and Walter Schachermayer
Context Trees, Variable Length Markov Chains and Dynamical Sources Peggy Cénac, Brigitte Chauvin, Frédéric Paccaut, and Nicolas Pouyanne
Abstract Infinite random sequences of letters can be viewed as stochastic chains or as strings produced by a source, in the sense of information theory. The relationship between Variable Length Markov Chains (VLMC) and probabilistic dynamical sources is studied. We establish a probabilistic frame for context trees and VLMC and we prove that any VLMC is a dynamical source for which we explicitly build the mapping. On two examples, the “comb” and the “bamboo blossom”, we find a necessary and sufficient condition for the existence and the uniqueness of a stationary probability measure for the VLMC. These two examples are detailed in order to provide the associated Dirichlet series as well as the generating functions of word occurrences.
P. Cénac () Université de Bourgogne, Institut de Mathématiques de Bourgogne IMB UMR 5584 CNRS, 9 avenue Alain Savary - BP 47870, 21078 DIJON CEDEX, Bourgogne, France e-mail:
[email protected] B. Chauvin INRIA Rocquencourt, project Algorithms, Domaine de Voluceau B.P.105, 78153 Le Chesnay CEDEX, France Laboratoire de Mathématiques de Versailles, CNRS, UMR 8100, Université de Versailles St-Quentin, 45 avenue des Etats-Unis, 78035 Versailles CEDEX, France e-mail:
[email protected] F. Paccaut LAMFA, CNRS, UMR 6140, Université de Picardie Jules Verne, 33 rue Saint-Leu, 80039 Amiens, France e-mail:
[email protected] N. Pouyanne Laboratoire de Mathématiques de Versailles, CNRS, UMR 8100, Université de Versailles St-Quentin, 45 avenue des Etats-Unis, 78035 Versailles CEDEX, France e-mail:
[email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__1, © Springer-Verlag Berlin Heidelberg 2012
1
2
P. Cénac et al.
Keywords Variable length Markov chains • Dynamical systems of the interval • Dirichlet series • Occurrences of words • Probabilistic dynamical sources AMS Classification: 60J05, 37E05
1 Introduction Our objects of interest are infinite random sequences of letters. One can imagine DNA sequences (the letters are A, C , G, T ), bits sequences (the letters are 0, 1) or any random sequence on a finite alphabet. Such a sequence can be viewed as a stochastic chain or as a string produced by a source, in the sense of information theory. We study this relation for the so-called Variable Length Markov Chains (VLMC). From now on, we are given a finite alphabet A . An infinite random sequence of letters is often considered as a chain .Xn /n2Z , i.e. an A Z -valued random variable. The Xn are the letters of the chain. Equivalently such a chain can be viewed as a random process .Un /n2N that takes values in the set L WD A N of left-infinite words1 and that grows by addition of a letter on the right at each step of discrete time. The L -valued processes we consider are Markovian ones. The evolution from Un D : : : X1 X0 X1 : : : Xn to UnC1 D Un XnC1 is described by the transition probabilities P.UnC1 D Un ˛jUn /, ˛ 2 A . In the context of chains, the point of view has mainly been a statistical one until now, going back to Harris [14] who speaks of chains of infinite order to express the fact that the production of a new letter depends on a finite but unbounded number of previous letters. Comets et al. [7] and Gallo and Garcia [11] deal with chains of infinite memory. Rissanen [23] introduces a class of models where the transition from the word Un to the word UnC1 D Un XnC1 depends on Un through a finite suffix of Un and he calls this relevant part of the past a context. Contexts can be stored as the leaves of a so-called context tree so that the model is entirely defined by a family of probability distributions indexed by the leaves of a context tree. In this paper, Rissanen develops a near optimal universal data compression algorithm for long strings generated by non independent information sources. The name VLMC is due to Bühlmann and Wyner [5]. It emphasizes the fact that the length of memory needed to predict the next letter is a not necessarily bounded function of the sequence Un . An overview on VLMC can be found in Galves and Löcherbach [12]. We give in Sect. 2 a complete probabilistic definition of VLMC. Let us present here a foretaste, relying on the particular form of the transition probabilities P.UnC1 D Un ˛jUn /. Let T be a saturated tree on A , which means that every internal node of the tree—i.e. a word on A —has exactly jA j children. With each leaf c of the tree, also called a context, is associated a probability distribution qc on A . The basic fact is that any left-infinite sequence can thus be “plugged in” a
1
In the whole text, N denotes the set of nonnegative integers.
Context Trees, Variable Length Markov Chains and Dynamical Sources
3
unique context of the tree T : any Un can be uniquely written Un D : : : c, where, for any word c D ˛1 ˛N , c denotes the reversed word c D ˛N ˛1 . In other terms, for any n, there is a unique context c in the tree T such that c is a suffix of Un ; this word is denoted by c D pref .Un /. We define the VLMC associated with these data as the L -valued homogeneous Markov process whose transition probabilities are, for any letter ˛ 2 A , P.UnC1 D Un ˛jUn / D q
pref .Un / .˛/:
When the tree is finite, the final letter process .Xn /n0 is an ordinary Markov chain whose order is the height of the tree. The case of infinite trees is more interesting, providing concrete examples of non Markov chains. In the example of Fig. 1, the context tree is finite of height 4 and, for instance, P.UnC1 D Un 0jUn D 0101110/ D q011 .0/ because pref . 0101110/ D 011 (read the word 0101110 right-to-left and stop when finding a context). In information theory, one considers that words are produced by a probabilistic source as developed in Vallée and her group papers (see Clément et al. [6] for an overview). In particular, a probabilistic dynamical source is defined by a coding function W Œ0; 1 ! A , a mapping T W Œ0; 1 ! Œ0; 1 having suitable properties and a probability measure on Œ0; 1. These data being given, the dynamical source produces the A -valued random process .Yn /n2N WD ..T n //n2N , where is a -distributed random variable on Œ0; 1. On the right side of Fig. 1, one can see the graph of some T , a subdivision of Œ0; 1 in two subintervals I0 D 1 .0/ and I1 D 1 .1/ and the first three real numbers x, T x and T 2 x, where x is a realization
q1
q00
q011
q0100
q0101
T2x x I0
Tx I1
Fig. 1 Example of probabilized context tree (on the left) and its corresponding dynamical system (on the right)
4
P. Cénac et al.
of the random variable . The right-infinite word corresponding to this example has 010 as a prefix. We prove in Theorem 1 that every stationary VLMC is a dynamical source. More precisely, given a stationary VLMC, .Un /n2N say, we construct explicitly a dynamical source .Yn /n2N such that the letter processes .Xn /n2N and .Yn /n2N are symmetrically distributed, which means that for any finite word w of length N C 1, P.X0 : : : XN D w/ D P.Y0 : : : YN D w/. In Fig. 1, the dynamical system together with Lebesgue measure on Œ0; 1 define a probabilistic source that corresponds to the stationary VLMC defined by the drawn probabilized context tree. The previous result is possible only when the VLMC is stationary. The question of existence and uniqueness of a stationary distribution arises naturally. We give a complete answer in two particular cases (Propositions 1 and 4 in Sect. 4) and we propose some tracks for the general case. Our two examples are called the “infinite comb” and the “bamboo blossom”; they can be visualized in Figs. 6 and 7, respectively pages 17 and 28. Both have an infinite branch so that the letter process of the VLMC is non Markovian. They provide quite concrete cases of infinite order chains where the study can be completely handled. We first exhibit a necessary and sufficient condition for existence and uniqueness of a stationary measure. Then the dynamical system is explicitly built and drawn. In particular, for some suitable data values, one gets in this way examples of intermittent sources. Quantifying and visualizing repetitions of patterns is another natural question arising in combinatorics on words. Tries, suffix tries and digital search trees are usual convenient tools. The analysis of such structures relies on the generating functions of the word occurrences and on the Dirichlet series attached to the sources. In both examples, these computations are performed. The paper is organized as follows. Section 2 is devoted to the precise definition of variable length Markov chains. In Sect. 3 the main result Theorem 1 is established. In Sect. 4, we complete the paper with our two detailed examples: “infinite comb” and “bamboo blossom”. The last section gathers some prospects and open problems.
2 Context Trees and Variable Length Markov Chains In this section, we first define probabilized context trees; then we associate with a probabilized context tree a so-called variable length Markov chain (VLMC).
2.1 Words and Context Trees Let A be a finite alphabet, i.e. a finite ordered set. Its cardinality is denoted by jA j. For the sake of shortness, our results in the paper are given for the alphabet A D f0; 1g but they remain true for any finite alphabet. Let
Context Trees, Variable Length Markov Chains and Dynamical Sources
W D
[
5
An
n0
be the set of all finite words over A . The concatenation of two words v D v1 : : : vM and w D w1 : : : wN is vw D v1 : : : vM w1 : : : wN . The empty word is denoted by ;. Let L D A N be the set of left-infinite sequences over A and R DAN be the set of right-infinite sequences over A . If k is a nonnegative integer and if w D ˛k ˛0 is any finite word on A , the reversed word is denoted by w D ˛0 ˛k : The cylinder based on w is defined as the set of all left-infinite sequences having w as a suffix: L w D fs 2 L ; 8j 2 fk; ; 0g; sj D ˛j g: By extension, the reversed sequence of s D ˛1 ˛0 2 L is s D ˛0 ˛1 2 R. The set L is equipped with the -algebra generated by all cylinders based on finite words. The set R is equipped with the -algebra generated by all cylinders wR D fr 2 R; w is a prefix of rg. Let T be a tree, i.e. a subset of W satisfying two conditions: • ;2T • 8u; v 2 W , uv 2 T H) u 2 T This corresponds to the definition of rooted planar trees in algorithmics. Let C F .T / be the set of finite leaves of T , i.e. the nodes of T without any descendant: C F .T / D fu 2 T ; 8j 2 A ; uj … T g: An infinite word u 2 R such that any finite prefix of u belongs to T is called an infinite leaf of T . Let us denote the set of infinite leaves of T by C I .T / D fu 2 R; 8v prefix of u; v 2 T g: Let C .T / D C F .T / [ C I .T / be the set of all leaves of T . The set T n C F .T / is constituted by the internal nodes of T . When there is no ambiguity, T is omitted and we simply write C ; C F and C I . Definition 1. A tree is saturated when each internal node w has exactly jA j children, namely the set fw˛; ˛ A g T .
6
P. Cénac et al.
Definition 2. (Context tree) A context tree is a saturated tree having a finite or countable set of leaves. The leaves are called contexts. Definition 3. (Probabilized context tree) A probabilized context tree is a pair
T ; .qc /c2C .T /
where T is a context tree over A and .qc /c2C .T / is a family of probability measures on A , indexed by the countable set C .T / of all leaves of T . Examples. See Fig. 1 for an example of finite probabilized context tree with five contexts. See Fig. 6 for an example of infinite probabilized context tree, called the infinite comb. Definition 4. A subset K of W [ R is a cutset of the complete jA j-ary tree when both following conditions hold (i) no word of K is a prefix of another word of K (ii) 8r 2 R, 9u 2 K , u prefix of r. Condition (i) entails uniqueness in (ii). Obviously a tree T is saturated if and only if the set of its leaves C is a cutset. Take a saturated tree, then 8r 2 R; either r 2 C I or 9Šu 2 W ; u 2 C F ; u prefix of r:
(1)
This can also be said on left-infinite sequences: 8s 2 L ; either s 2 C I or 9Šw 2 W ; w 2 C F ; w suffix of s: In other words:
[
L D
s2C I
fsg [
[
L w:
(2)
(3)
w2C F
This partition of L will be extensively used in the sequel. Both cutset properties (1) and (2) will be used in the paper, on R for trees, on L for chains. Both orders of reading will be needed. Definition 5. (Prefix function) Let T be a saturated tree and C its set of contexts. For any s 2 L , pref .s/ denotes the unique context ˛1 : : : ˛N such that s D : : : ˛N : : : ˛1 . The map pref W L ! C is called the prefix function. For technical reasons, this function is extended to
pref W L [ W ! T in the following way:
• if w 2 T then pref .w/ D w;
Context Trees, Variable Length Markov Chains and Dynamical Sources
7
• if w 2 W n T then pref .w/ is the unique context ˛1 : : : ˛N such that w has ˛N : : : ˛1 as a suffix. Note that the second item of the definition is also valid when w 2 C . Moreover pref .w/ is always a context except when w is an internal node.
2.2 VLMC Associated with a Context Tree Definition 6. (VLMC) Let .T ; .qc /c2C / be a probabilized context tree. The associated Variable Length Markov Chain (VLMC) is the order 1 Markov chain .Un /n0 with state space L , defined by the transition probabilities 8n 0; 8˛ 2 A ; P .UnC1 D Un ˛jUn / D q
pref .Un /
.˛/ :
(4)
Remark 1. As usually, we speak of the Markov chain defined by the transition probabilities (4), because these data together with the distribution of U0 define a unique L -valued Markov random process .Un /n0 (see for example Revuz [22]). The rightmost letter of the sequence Un 2 L will be denoted by Xn so that 8n 0; UnC1 D Un XnC1 : The final letter process .Xn /n0 is not Markov of any finite order as soon as the context tree has at least one infinite context. As already mentioned in the introduction, when the tree is finite, .Xn /n0 is a Markov chain whose order is the height of the tree, i.e. the length of its longest branch. The vocable VLMC is somehow confusing but commonly used. Definition 7. (SVLMC) Let .Un /n0 be a VLMC. When a stationary probability measure on L exists and when it is the initial distribution, we say that .Un /n0 is a Stationary Variable Length Markov Chain (SVLMC). Remark 2. In the literature, the name VLMC is usually applied to the chain .Xn /n2Z . There exists a natural bijective correspondence between A -valued chains .Xn /n2Z and L -valued processes .Un D U0 X1 : : : Xn ; n 0/. Consequently, finding a stationary probability for the chain .Xn /n2Z is equivalent to finding a stationary probability for the process .Un /n0 .
3 Stationary Variable Length Markov Chains The existence and the uniqueness of a stationary measure for two examples of VLMC will be established in Sect. 4. In the present section, we assume that a stationary measure on L exists and we consider a -distributed VLMC. In the
8
P. Cénac et al.
preliminary Sect. 3.1, we show how the stationary probability of finite words can be expressed as a function of the data and the values of at the tree nodes. In Sect. 3.2, the main theorem is proved.
3.1 General Facts on Stationary Probability Measures For the sake of shortness, when is a stationary probability for a VLMC, we write .w/ instead of .L w/, for any w 2 W : .w/ D P.U0 2 L w/ D P.X.jwj1/ : : : X0 D w/:
(5)
Extension of notation qu for internal nodes. The VLMC is defined by its context tree T together with a family .qc /c2C of probability measures on A indexed by the contexts of the tree. When u is an internal node of the context tree, we extend the notation qu by 8 .u˛/ ˆ ˆ if .u/ ¤ 0 < .u/ qu .˛/ D (6) ˆ ˆ : 0 if .u/ D 0 for any ˛ 2 A . Thus, in any case, being stationary, .u˛/ D .u/qu .˛/ as soon as u is an internal node of the context tree. With this notation, the stationary probability of any cylinder can be expressed by the following simple formula (8). Lemma 1. Consider a SVLMC defined by a probabilized context tree and let denote any stationary probability measure on L . Then, (i) for any finite word w 2 W and for any letter ˛ 2 A , .w˛/ D .w/q
pref .w/ .˛/:
(7)
(ii) For any finite word w D ˛1 : : : ˛N 2 W , .w/ D
N 1 Y
q
pref .˛1 :::˛k / .˛kC1 /
(8)
kD0
(if k D 0, ˛1 : : : ˛k denotes the empty word ;, pref .;/ D ;, q; .˛/ D .˛/ and .;/ D .L / D 1).
Proof. (i) If w is an internal node of the context tree, then pref .w/ D w and the formula comes directly from the definition of qw . If not, .w˛/ D .U1 2 L w˛/ by stationarity; because of Markov property, .w˛/ D P.U0 2 L w/P.U1 2 L w˛jU0 2 L w/ D .w/q
pref .w/ .˛/:
Context Trees, Variable Length Markov Chains and Dynamical Sources
9
Finally, (ii) follows from (i) by a straightforward induction.
t u
Remark 3. When A D f0; 1g and is any stationary probability of a SVLMC, then, for any natural number n, .10n / D .0n 1/. Indeed, on one hand, by disjoint union, .0n / D .0nC1 / C .10n /. On the other hand, by stationarity, .0n / D P.X1 : : : Xn D 0n / D P.X0 : : : Xn1 D 0n / D P.X0 : : : Xn D 0nC1 / C P.X0 : : : Xn D 0n 1/ D .0nC1 / C .0n 1/: These equalities lead to the result. Of course, symmetrically, .01n / D .1n 0/ under the same assumptions.
3.2 Dynamical System Associated with a VLMC We begin with a general presentation of a probabilistic dynamical source in Sect. 3.2.1. Then we build step by step partitions of the interval Œ0; 1 (Sect. 3.2.2) and a mapping (Sect. 3.2.3) based on the stationary measure of a given SVLMC. It appears in Sect. 3.2.4 that this particular mapping keeps Lebesgue measure invariant. All these arguments combine to provide in the last Sect. 3.2.5 the proof of Theorem 1 which allows us to see a VLMC as a dynamical source. In the whole section, I stands for the real interval Œ0; 1 and the Lebesgue measure of a Borelian J will be denoted by jJ j. 3.2.1 General Probabilistic Dynamical Sources Let us present here the classical formalism of probabilistic dynamical sources (see Clément et al. [6]). It is defined by four elements: • A topological partition of I by intervals .I˛ /˛2A • A coding function W I ! A , such that, for each letter ˛, the restriction of to I˛ is equal to ˛ • A mapping T W I ! I • A probability measure on I Such a dynamical source defines an A -valued random process .Yn /n2N as follows. Pick a random real number x according to the measure . The mapping T yields the orbit .x; T .x/; T 2 .x/; : : :/ of x. Thanks to the coding function, this defines the right-infinite sequence .x/.T .x//.T 2 .x// whose letters are Yn WD .T n .x// (see Fig. 2). For any finite word w D ˛0 : : : ˛N 2 W , let Bw D
N \ kD0
T k I˛k
10
P. Cénac et al.
Fig. 2 The graph of a mapping T , the intervals I0 and I1 that code the interval I by the alphabet A D f0; 1g and the first three points of the orbit of an x 2 I by the corresponding dynamical system
I1
I0
x I0
Tx
T2x I1
be the Borelian set of real numbers x such that the sequence .Yn /n2N has w as a prefix. Consequently, the probability that the source emits a sequence of symbols starting with the pattern w is equal to .Bw /. When the initial probability measure on I is T -invariant, the dynamical source generates a stationary A -valued random process which means that for any n 2 N, the random variable Yn is ı-distributed. The following classical examples often appear in the literature: let p 20; 1Œ, I0 D Œ0; 1 p and I1 D1 p; 1. Let T W I ! I be the only function which maps linearly and increasingly I0 and I1 onto I (see Fig. 3 when p D 0:65, left side). Then, starting from Lebesgue measure, the corresponding probabilistic dynamical source is Bernoulli: the Yn are i.i.d. and P.Y0 D 1/ D p. In the same vein, if T is the mapping drawn on the right side of Fig. 3, starting from Lebesgue measure, the f0; 1g-valued process .Yn /n2N is Markov and stationary, with transition matrix
0:4 0:6 : 0:7 0:3
The assertions on both examples are consequences of Thales theorem. These two basic examples are particular cases of Theorem 1.
3.2.2 Ordered Subdivisions and Ordered Partitions of the Interval Definition 8. A family .Iw /w2W of subintervals of I indexed by all finite words is said to be an A -adic subdivision of I whenever
Context Trees, Variable Length Markov Chains and Dynamical Sources
I0
I1
I0
11
I1
Fig. 3 Mappings generating a Bernoulli source and a stationary Markov chain of order 1. In both cases, Lebesgue measure is the initial one
(i) for any w 2 W , Iw is the disjoint union of Iw˛ , ˛ 2 A ; (ii) for any .˛; ˇ/ 2 A 2 , for any w 2 W , ˛ < ˇ H) 8x 2 Iw˛ ; 8y 2 Iwˇ ; x < y: Remark 4. For any integer p 2, the usual p-adic subdivision of I is a particular case of A -adic subdivision for which jA j D p and jIw j D p jwj for any finite word w 2 W . For a general A -adic subdivision, the intervals associated with two k-length words need not have the same length. The inclusion relations between the subintervals Iw of an A -adic subdivision are thus coded by the prefix order in the complete jA j-ary planar tree. In particular, for any w 2 W and for any cutset K of the complete jA j-ary tree, Iw D
[
Iwv
v2K
(this union is a disjoint one; see Sect. 2.1 for a definition of a cutset). We will use the following convention for A -adic subdivisions: we require the intervals Iv to be open on the left side and closed on the right side, except the ones of the form I0n that are compact. Obviously, if is any probability measure on R D A N , there exists a unique A -adic subdivision of I such that jIw j D .wR/ for any w 2 W . Given an A -adic subdivision of I , we extend the notation Iw to right-infinite words by
12
P. Cénac et al.
8r 2 R; Ir D
\
Iw :
w2W w prefix of r
Definition 9. A family .Iv /v2V of subintervals of I indexed by a totally ordered set V is said toSdefine an ordered topological partition of I when (i) I D v2V cl.Iv /, (ii) for any v; v0 2 V , v ¤ v0 H) int.Iv / \ int.Iv0 / D ;, (iii) for any v; v0 2 V , v v0 H) 8x 2 Iv ; 8x 0 2 Iv0 ; x x 0 where cl.Iv / and int.Iv / stand respectively for the closure and the interior of Iv . We will denote X I D " Iv : v2V
P P We will use the following fact: if I D v2V " Iv D v2V " Jv are two ordered topological partitions of I indexed by the same countable ordered set V , then Iv D Jv for any v 2 V as soon as jIv j D jJv j for any v 2 V . 3.2.3 Definition of the Mapping T Let .Un /n0 be a SVLMC, defined by its probabilized context tree .T ; .qc /c2C / and a stationary2 probability measure on L . We first associate with the unique A -adic subdivision .Iw /w2W of I , defined by: 8w 2 W ; jIw j D .w/; (recall that if w D ˛1 : : : ˛N , w is the reversed word ˛N : : : ˛1 and that .w/ denotes .L w/). We consider now three ordered topological partitions of I . • The coding partition: It consists in the family .I˛ /˛2A : I D
X
" I˛ D I0 C I1 :
˛2A
• The vertical partition: The countable set of finite and infinite contexts C is a cutset of the A -ary tree. The family .Ic /c2C thus defines the so-called vertical ordered topological
2
Note that this construction can be made replacing by any probability measure on L .
Context Trees, Variable Length Markov Chains and Dynamical Sources
partition I D
X
13
" Ic :
c2C
• The horizontal partition: A C is the set of leaves of the context tree A T D f˛w; ˛ 2 A ; w 2 T g. As before, the family .I˛c /˛c2A C defines the so-called horizontal ordered topological partition X I D " I˛c : ˛c2A C
Definition 10. The mapping T W I ! I is the unique left continuous function such that: (i) the restriction of T to any I˛c is affine and increasing; (ii) for any ˛c 2 A C , T .I˛c / D Ic . The function T is always increasing on I0 and on I1 . When qc .˛/ ¤ 0, the slope of T on an interval I˛c is 1=qc .˛/. Indeed, with formula (7), one has jI˛c j D .c˛/ D qc .˛/.c/ D jIc jqc .˛/: When qc .˛/ D 0 and jIc j ¤ 0, the interval I˛c is empty so that T is discontinuous at xc D .fs 2 L ; s cg/ ( denotes here the alphabetical order on R). Note that jIc j D 0 implies jI˛c j D 0. In particular, if one assumes that all the probability measures qc , c 2 C , are nontrivial (i.e. as soon as they satisfy qc .0/qc .1/ ¤ 0), then T is continuous on I0 and I1 . Furthermore, T .I0 / D cl.T .I1 // D I and for any c 2 C , T 1 Ic D I0c [ I1c (see Fig. 4). Example: the four flower bamboo. The four flower bamboo is the VLMC defined by the finite probabilized context tree of Fig. 5. There exists a unique stationary measure under conditions which are detailed later, in Example 3. We represent on Fig. 5 the mapping T built with this , together with the respective subdivisions
Ic
I 0c
I 1c
Fig. 4 Action of T on horizontal and vertical partitions. On this figure, c is any context and the alphabet is f0; 1g
14
P. Cénac et al.
of x-axis and y-axis by the four Ic and the eight I˛c . The x-axis is divided by both coding and horizontal partitions; the y-axis is divided by both coding and vertical partitions. This figure has been drawn with the following data on the four flower bamboo: q00 .0/ D 0:4, q010 .0/ D 0:6, q011 .0/ D 0:8 and q1 .0/ D 0:3. 3.2.4 Properties of the Mapping T The following key lemma explains the action of the mapping T on the intervals of the A -adic subdivision .Iw /w2W . More precisely, it extends the relation T .I˛c / D Ic , for any ˛c 2 A C , to any finite word. Lemma 2. For any finite word w 2 W and any letter ˛ 2 A , T .I˛w / D Iw . Proof. Assume first that w … T . Let then c 2 C be the unique context such that c is a prefix of w. Because of the prefix order structure of the A -adic subdivision .Iv /v , one has the first ordered topological partition Ic D
X
" Iv
(9)
v2W ; jvjDjwj c prefix of v
(the set of indices is a cutset in the tree of c descendants). On the other hand, the same topological partition applied to the finite word ˛w leads to I˛c D
X
" I˛v :
v2W ; jvjDjwj c prefix of v
q1
I1
I1
I011 I010
I0
I00
q00 q010
q011
I0 I000
I1 I01 I100
I0010 I0011
I11
I1010 I1011
Fig. 5 On the left, the four flower bamboo context tree. On the right, its mapping together with the coding, the vertical and the horizontal partitions of Œ0; 1
Context Trees, Variable Length Markov Chains and Dynamical Sources
15
Taking the image by T , one gets the second ordered topological partition X
Ic D
" T .I˛v /:
(10)
v2W ; jvjDjwj c prefix of v
Now, if c is a prefix of a finite word v, I˛v I˛c and the restriction of T to I˛c is affine. By Thales theorem, it comes jT .I˛v /j D jI˛v j:
jIc j : jI˛c j
Since is a stationary measure for the VLMC, jI˛c j D .c˛/ D qc .˛/.c/ D jIc jqc .˛/: Furthermore, one has .v˛/ D qc .˛/.v/. Finally, jT .I˛v /j D jIv j. Relations (9) and (10) are two ordered countable topological partitions, the components with the same indices being of the same length: the partitions are necessarily the same. In particular, because w belongs to the set of indices, this implies that T .I˛w / D Iw . Assume now that w 2 T . Since the set of contexts having w as a prefix is a cutset of the tree of the descendants of w, one has the disjoint union I˛w D
[
I˛c :
c2C w prefix of c
Taking the image by T leads to T .I˛w / D
[
Ic D Iw
c2C ;w prefix of c
t u
and the proof is complete.
Remark 5. The same proof shows in fact that if w is any finite word, T 1 Iw D I0w [ I1w (disjoint union). Lemma 3. For any ˛ 2 A , for any context c 2 C , for any Borelian set B Ic , jI˛ \ T 1 Bj D jBjqc .˛/: Proof. It is sufficient to show the lemma when B is an interval. The restriction of T t to I˛c is affine and T 1 Ic D I0c [I1c . The result is thus due to Thales Theorem. u Corollary 1. If T is the mapping associated with a SVLMC, Lebesgue measure is invariant by T , i.e. jT 1 Bj D jBj for any Borelian subset of I .
16
P. Cénac et al.
S Proof. Since B D c2C B \ Ic (disjoint union), it suffices to prove that jT 1 Bj D jBj for any Borelian subset of Ic where c is any context. If B Ic , because of Lemma 3, jT 1 Bj D jI0 \ T 1 Bj C jI1 \ T 1 Bj D jBj.qc .0/ C qc .1// D jBj: t u
3.2.5 SVLMC as Dynamical Source We now consider the stationary probabilistic dynamical source ..I˛ /˛2A ; ; T; j:j/ built from the SVLMC. It provides the A -valued random process .Yn /n2N defined by Yn D .T n / where is a uniformly distributed I -valued random variable and the coding function. Since Lebesgue measure is T -invariant, all random variables Yn have the same law, namely P.Yn D 0/ D jI0 j D .0/. Definition 11. Two A -valued random processes .Vn /n2N and .Wn /n2N are called symmetrically distributed whenever for any N 2 N and for any finite word w 2 A N C1 , P.W0 W1 : : : WN D w/ D P.V0 V1 : : : VN D w/. In other words, .Vn /n2N and .Wn /n2N are symmetrically distributed if and only if for any N 2 N, the random words W0 W1 : : : WN and VN VN 1 : : : V0 have the same distribution. Theorem 1. Let .Un /n2N be a SVLMC and a stationary probability measure on L . Let .Xn /n2N be the process of final letters of .Un /n2N . Let T W I ! I be the mapping defined in Sect. 3.2.3. Then, (i) Lebesgue measure is T -invariant. (ii) If is any uniformly distributed random variable on I , the processes .Xn /n2N and ..T n //n2N are symmetrically distributed. Proof. (i) has been already stated and proven in Corollary 1. T k (ii) As before, for any finite word w D ˛0 : : : ˛N 2 W , let Bw D N I˛k kD0 T be the Borelian set of real numbers x such that the right-infinite sequence ..T n x//n2N has w as a prefix. By definition, B˛ D I˛ if ˛ 2 A . More generally, we prove the following claim: for any w 2 W , Bw D Iw . Indeed, if ˛ 2 A and w 2 W , B˛w D I˛ \ T 1 Bw ; thus, by induction on the length of w, B˛w D I˛ \ T 1 Iw D I˛w , the last equality being due to Lemma 2. There is now no difficulty in finishing the proof: if w 2 W is any finite word of length N C 1, then P.X0 : : : XN D w/ D .w/ D jIw j. Thus, because of the claim, P.X0 : : : XN D w/ D jBw j D P.Y0 : : : YN D w/. This proves the result. t u
Context Trees, Variable Length Markov Chains and Dynamical Sources
q1
17
I1
q01 q001
I0
q0001 q0 ∞
q0n1
I0
I1
Fig. 6 Infinite comb probabilized context tree (on the left) and the associated dynamical system (on the right)
4 Examples 4.1 The Infinite Comb 4.1.1 Stationary Probability Measures Consider the probabilized context tree given on the left side of Fig. 6. In this case, there is one infinite leaf 01 and countably many finite leaves 0n 1, n 2 N. The data of a corresponding VLMC consists thus in probability measures on A D f0; 1g: q01 and q0n 1 ; n 2 N: Suppose that is a stationary measure on L . We first compute .w/ (notation (5)) as a function of .1/ when w is any context or any internal node. Because of formula (7), .10/ D .1/q1 .0/ and an immediate induction shows that, for any n 0, .10n / D .1/cn ; (11) where c0 D 1 and, for any n 1, cn D
n1 Y
q0k 1 .0/:
(12)
kD0
The stationary probability of a reversed context is thus necessarily given by formula (11). Now, if 0n is any internal node of the context tree, we need going
18
P. Cénac et al.
down along the branch in T to reach the contexts; using then the disjoint union .0nC1 / D .0n / .10n /, by induction, it comes for any n 0, .0n / D 1 .1/
n1 X
ck :
(13)
kD0
The stationary probability of a reversed internal node of the context tree is thus necessarily given by formula (13). It remains to compute .1/. The countable partition of the whole probability space given by all cylinders based on leaves in the context tree (Formula (3)) implies 1 .01 / D .1/ C .10/ C .100/ C : : : , i.e. 1 .01 / D
X
.1/cn :
(14)
n0
This leads to the following statement that covers all cases of existence, uniqueness and nontriviality for a stationary probability measure for the infinite comb. In the generic case (named irreducible case hereunder), we give a necessary and sufficient condition on the data for the existence of a stationary probability measure; moreover, when a stationary probability exists, it is unique. The reducible case is much more singular and gives rise to nonuniqueness. Proposition 1. (Stationary probability measures for an infinite comb) Let .Un /n0 be a VLMC defined by a probabilized infinite comb. (i) Irreducible case. Assume that q01 .0/ ¤ 1. (i.a) Existence. The Markov process .Un /n0P admits a stationary probability measure on L if and only if the numerical series cn defined by (12) converges. P (i.b) Uniqueness. Assume that the series cn converges and denote S.1/ D
X
cn :
(15)
n0
Then, the stationary probability measure on L is unique; it is characterized by .1/ D
1 S.1/
(16)
and formula (11), (13) and (8). Furthermore, is trivial if and only if q1 .0/ D 0, in which case it is defined by .11 / D 1. (ii) Reducible case. Assume that q01 .0/ D 1. P (ii.a) If the series cn diverges, then the trivial probability measure on L defined by .01 / D 1 is the unique stationary probability.
Context Trees, Variable Length Markov Chains and Dynamical Sources
19
P (ii.b) If the series cn converges, then there is a one parameter family of stationary probability measures on L . More precisely, for any a 2 Œ0; 1, there exists a unique stationary probability measure a on L such that a .01 / D a. The 1a probability a is characterized by a .1/ D S.1/ and formula (11), (13) and (8). Furthermore, a is non trivial except in the two following cases: • a D 1, in which case 1 is defined by 1 .01 / D 1; • a D 0 and q1 .0/ D 0, in which case 0 is defined by 0 .11 / D 1. Proof. (i) Assume that q01 .0/ ¤ 1 and that is a stationary probability measure. By definition of probability transitions, .01 / D .01 /q01 .0/P so that .01 / necessarily vanishes. Thus, thanks to (14), .1/ ¤ 0, the series cn converges and formula (16) is valid. Moreover, when w is any context or any internal node of the context tree, .w/ is necessarily given by formula (16), (11) and (13). This shows that for any finite word w, .w/ is determined by formula (8). Since the cylinders L w, w 2 W span the -algebra on L , there is at most one stationary probability measure. This proves the only if part of (i.a), the uniqueness and the characterization claimed in (i.b). Conversely, when the series converges, formula (16), (11), (13) and (8) define a probability measure on the semiring spanned by cylinders, which extends to a stationary probability measure on the whole -algebra on L (see Billingsley [3] for a general treatment on semirings, -algebra, definition and characterization of probability measures). This proves the if part of (i.a). Finally, the definition of cn directly implies that S.1/ D 1 if and only if q1 .0/ D 0. This proves the assertion of (i.b) on the triviality of . 1 (ii) Assume Pthat q0 .0/ D 1. Formula (14) is always valid so that the divergence of the series cn forces .1/ to vanish and, consequently, any stationary measure to be the trivial one defined by .01 / D 1. Besides, with the assumption q01 .0/ D 1, one immediately sees that this trivial probability is stationary, proving (ii.a). P To prove (ii.b), assume furthermore that the series cn converges and let a 2 Œ0; 1. As before, any stationary probability measure is completely determined 1a by .1/. Moreover, the probability measure defined by a .1/ D S.1/ , formula (11), (13) and (8) and in a way extended to the whole -algebra on L is clearly stationary. Because of formula (14), it satisfies a .01 / D 1 a .1/S.1/ D a: This proves the assertion on the one parameter family. Finally, a is trivial only if a .1/ 2 f0; 1g. If a D 1 then a .1/ D 0 thus 1 is the trivial probability that only charges 01 . If a D 0 then a .1/ D 1=S.1/ is nonzero and it equals 1 if and only if S.1/ D 1, i.e. if and only if q1 .0/ D 0, in which case 0 is the trivial probability that only charges 11 . t u Remark 6. This proposition completes previous results which give sufficient conditions for the existence of a stationary measure for an infinite comb. For instance, in
20
P. Cénac et al.
Galves and Löcherbach [12], the intervening condition is X
q0k 1 .1/ D C1;
k0
P which is equivalent with our notations to cn ! 0. Note that if cn is divergent, then the only possible stationary distribution is the trivial Dirac measure ı01 . 4.1.2 The Associated Dynamical System The vertical partition is made of the intervals I0n 1 for n 0. The horizontal partition consists in the intervals I00n 1 and I10n 1 , for n 0, together with two intervals coming from the infinite context, namely I01 and I101 . In the irreducible case, .01 / D 0 and these two last intervals become two accumulation points of the partition: 0 and .0/. The following lemma is classical and helps understand the behaviour of the mapping T at these accumulation points. Lemma 4. Let f W Œa; b ! R be continuous on Œa; b, differentiable on a; bŒnD where D is a countable set. The function f admits a right derivative at a and fr0 .a/ D
lim f 0 .x/
x!a;x>a x…D
as soon as this limit exists. Corollary 2. If .q0n 1 .0//n2N converges, then T is differentiable at 0 and .0/ (with a possibly infinite derivative) and Tr0 .0/ D lim
n!C1
1 1 ; Tr0 ..0// D lim : n!C1 q0n 1 .1/ q0n 1 .0/
When .q0n 1 .0//n2N converges to 1, Tr0 .0/ D 1. In this case, 0 is an indifferent fixed point and Tr0 ..0// D C1. The mapping T is a slight modification of the socalled Wang map (Wang [27]). The statistical properties of the Wang map are quite well understood (Lambert et al. [17]). The corresponding dynamical source is said intermittent. 4.1.3 Dirichlet Series For a stationary infinite comb, the Dirichlet series is defined on a suitable vertical open strip of C as X .s/ D .w/s : w2W
P
In the whole section we suppose that cn is convergent. Indeed, if it is divergent then the only stationary measure is the Dirac measure ı01 and .s/ is never defined.
Context Trees, Variable Length Markov Chains and Dynamical Sources
21
The computation of the Dirichlet series is tractable because of the following formula: for any finite words w; w0 2 W , .w1w0 /.1/ D .w1/.1w0 /:
(17)
This formula, which comes directly from formula (8), is true because of the very particular form of the contexts in the infinite comb. It is the expression of its renewal property. The computation of the Dirichlet series is made in two steps. Step 1. A finite word either does not contain any 1 or is of the form w10n , w 2 W , n 0. Thus, X XX .s/ D .0n /s C .w10n /s : n0
n0 w2W n
Because of formula (17) and (16), .w10 / D S.1/.w1/.10n/. Let us denote 1 .s/ D
X
.w1/s :
w2W
With this notation and formula (11) and (13), .s/ D
X 1 X s Rn C 1 .s/ cns s S.1/ n0 n0
where Rn stands for the rest
Rn D
X
ck :
(18)
kn
Step 2. It consists in the computation of 1 . A finite word having 1 as last letter either can be written 0n 1, n 0 or is of the form w10n 1, w 2 W , n 0. Thus it comes, X XX 1 .s/ D .0n 1/s C .w10n 1/s : n0
n0 w2W n
By formula (17) and (11), .w10 1/ D .w1/cn q0n 1 .1/ D .w1/.cn cnC1 /, so that X 1 X s 1 .s/ D cn C 1 .s/ .cn cnC1 /s s S.1/ n0 n0 and 1 .s/ D
P s 1 n0 cn P : s S.1/ 1 n0 .cn cnC1 /s
Putting results of both steps together, we obtain the following proposition. Proposition 2. With notations (12), (15) and (18), the Dirichlet series of a source obtained from a stationary infinite comb is
22
P. Cénac et al.
2 3 P s 2 c 1 4X s 5: P n0 n .s/ D R C S.1/s n0 n 1 n0 .cn cnC1 /s P
2 cns / . Remark 7. The analytic function 1P n0 s is always singular for s D 1 n0 .cn cnC1 / because its denominator vanishes while its numerator is a convergent series.
Examples. (1) Suppose that 0 < a < 1 and that q0n 1 .0/ D a for any n 0. Then an 1 cn D an , Rn D 1a and S.1/ D 1a . For such a source, the Dirichlet series is .s/ D
1 : 1 Œas C .1 a/s
In this case, the source is memoryless: all letters are drawn independently with the same distribution. The Dirichlet series of such sources have been extensively studied in Flajolet et al. [8] in the realm of asymptotics of average parameters of a trie. (2) Extension of Example 1: take a; b 20; 1Œ and consider the probabilized infinite comb defined by q0n 1 .0/ D
a if n is even; b if n is odd:
After computation, the Dirichlet series of the corresponding source under the stationary distribution turns out to have the explicit form 1 a C ab s 1 ab s .s/ D 1C C 1 .ab/s 1Ca 1Ca .1 C as /2 : 1 .ab/s .1 a/s as .1 b/s The configuration of poles of depends on arithmetic properties (approximation by rationals) of the logarithms of ab, 1 a and a.1 b/. The poles of such a series are the same as in the case of a memoryless source with an alphabet of three letters, see Flajolet et al. [8]. This could be extended to a family of examples. (3) Let ˛ > 2. We take data q0n 1 .0/, n 0 in such a way that c0 D 1 and, for any n 1, 1 X 1 cn D .n; ˛/ WD ; .˛/ k˛ kn
where is the Riemann function. Since cn 2 O.n1˛ / when n tends to infinity, there exists a unique stationary probability measure on L . One obtains
Context Trees, Variable Length Markov Chains and Dynamical Sources
S.1/ D 1 C
23
.˛ 1/ .˛/
and, for any n 1, Rn D
.˛ 1/ .n; ˛ 1/ .n 1/ .n; ˛/: .˛/
In particular, Rn 2 O.n2˛ / when n tends to infinity. The final formula for the Dirichlet series of this source is 2 P 3 s 2 c 1 4X s n0 n 5 R C : .s/ D S.1/s n0 n 1 .˛s/s .˛/
(4) One case of interest is when the associated dynamical system has an indifferent fixed point (see Sect. 4.1.2), for example when q0n 1 .0/ D 1
1 nC2
˛
;
with 1 < ˛ < 2. In this situation, cn D .1 C n/˛ and .s/ D
X n1
.n; ˛/s C
.˛s/2 .˛/s
1 ˛ : X 1 1 1 1 1 n˛s nC1 n1
4.1.4 Generating Function for the Exact Distribution of Word Occurrences in a Sequence Generated by a Comb The behaviour of the entrance time into cylinders is a natural question arising in dynamical systems. There exists a large literature on the asymptotic properties of entrance times into cylinders for various kind of systems, symbolic or geometric; see Abadi and Galves [1] for an extensive review on the subject. Most of the results deal with an exponential approximation of the distribution of the first entrance time into a small cylinder, sometimes with error terms. The most up-to-date result on this framework is Abadi and Saussol [2] in which the hypothesis are made only in terms of the mixing type of the source (so-called ˛-mixing). We are here interested in exact distribution results instead of asymptotic behaviours. Several studies in probabilities on words are based on generating functions. For example one may cite Régnier [20], Reinert et al. [21], Stefanov and Pakes [25]. For i.i.d. sequences, Blom and Thorburn [4] give the generating function of the first occurrence of a word, based on a recurrence relation on the probabilities. This result is extended to Markovian sequences by Robin and Daudin [24]. Nonetheless, other
24
P. Cénac et al.
approaches are considered: one of the more general techniques is the so-called Markov chain embedding method introduced by Fu [9] and further developed by Fu and Koutras [10], Koutras [16]. A martingale approach (see Gerber and Li [13], Li [18], Williams [28]) is an alternative to the Markov chain embedding method. These two approaches are compared in Pozdnyakov et al. [19]. We establish results on the exact distribution of word occurrences in a random sequence generated by a comb (or a bamboo in Sect. 4.2.4). More precisely, we make explicit the generating function of the random variable giving the r th occurrence of a k-length word, for any word w such that w is not an internal node of T . Let us consider the process X D .Xn /n0 of final letters of .Un /n0 , in the particular case of a SVLMC defined by an infinite comb. Let w D w1 : : : wk be a word of length k 1. We say that w occurs at position n k in the sequence X if the word w ends at position n: fw at ng D fXnkC1 : : : Xn D wg D fUn 2 L wg: .r/
Let us denote by Tw generating function:
.r/
the position of the r th occurrence of w in X and ˚w its X ˚w.r/ .x/ WD P Tw.r/ D n x n : n0
The following notation is used in the sequel: for any finite word u 2 W , for any finite context c 2 C and for any n 0, qc.n/ .u/ D P XnjujC1 : : : Xn D ujX.jcj1/ : : : X0 D c : These quantities may be computed in terms of the data qc . Proposition 3 generalizes results of Robin and Daudin [24]. Proposition 3. For a SVLMC defined by an infinite comb, with the above notations, for a word w such that w is non internal node, the generating function of its first occurrence is given, for jxj < 1, by ˚w.1/ .x/ D
x k .w/ .1 x/Sw .x/
and the generating function of its r th occurrence is given, for jxj < 1, by r1 1 .r/ .1/ ˚w .x/ D ˚w .x/ 1 ; Sw .x/ where Sw .x/ D Cw .x/ C
1 X j Dk
q
.j / .w/x j ; pref .w/
Context Trees, Variable Length Markov Chains and Dynamical Sources
Cw .x/ D 1 C
k1 X
1fwj C1 :::wk Dw1 :::wkj g q
j D1
25
.j / wkj C1 : : : wk x j : pref .w/
Remark 8. The term Cw .x/ is a generalization of the probabilized autocorrelation polynomial defined in Jacquet and Szpankowski [15] in the particular case when the .Xn /n0 are independent and identically distributed. For a word w D w1 : : : wk this polynomial is equal to cw .x/ D
k1 X
cj;w
j D0
1 xj ; .w1 : : : wkj /
where cj;w D 1 if the k j -length suffix of w is equal to its k j -length prefix, and is equal to zero otherwise. When the .Xn /n0 are independent and identically distributed, we have k1 X
1fwj C1:::wk Dw1 :::wkj g q
j D1
.j / wkj C1 : : : wk x j pref .w/
D
k1 X j D1
cj;w
.w/ xj w1 : : : wkj
that is Cw .x/ D .w/cw .x/: Proof. We first deal with w D 10k1 , that is the only word w of length k such that w 2 C . For the sake of shortness, we will denote by pn the probability that .1/ Tw D n. From the obvious decomposition fw at ng D fTw.1/ D ng [ fTw.1/ < n and w at ng;
(disjoint union)
it comes by stationarity of .w/ D pn C
n1 X
pz P XnkC1 : : : Xn D wjTw.1/ D z :
zDk
Due to the renewal property of the comb, the conditional probability can be rewritten
P .XnkC1 : : : Xn D wjXzkC1 : : : Xz D w/ if z n k 0 if z > n k
the second equality being due to the lack of possible auto-recovering in w. Consequently, we have .w/ D pn C
nk X zDk
.nz/
pz qw
.w/:
26
P. Cénac et al.
Hence, for x < 1, it comes C1
C1
nk
nDk
nDk
zDk
X X X x k .w/ .nz/ D pn x n C xn pz qw .w/; 1x 0 1 1 X x k .w/ .j / D ˚w.1/ .x/ @1 C x j qw .w/A ; 1x
so that
j Dk
which leads to ˚w.1/ .x/ D
x k .w/ : .1 x/Sw .x/
Note that when w D 10k1 , Cw .x/ D 1. Proceeding in the same way for the r th occurrence, from the decomposition fw at ng D fTw.1/ D ng [ fTw.2/ D ng [ : : : [ fTw.r/ D ng [ fTw.r/ < n and w at ng; .`/
and denoting p.n; `/ D P.Tw D n/, the following recursive equation holds: .w/ D pn C p.n; 2/ C : : : C p.n; r/ C
n1 X P Tw.r/ D z and w at n : zDk
Again, by splitting the last term into two terms and using the non-overlapping structure of w, one gets .w/ D pn C p.n; 2/ C : : : C p.n; r/ C
nk X
.nz/
pz qw
.w/:
zDk
From this recursive equation, proceeding exactly in the same way, one gets for the generating function, for x < 1, ˚w.r/ .x/ D ˚w.1/ .x/ 1
1 Sw .x/
r1
:
Let us now consider the case of words w such that w … T , that is the words w such that wj D 1 for at least one integer j 2 f2; : : : ; kg. We denote by i the last position of a 1 in w, that is pref .w/ D 0ki 1. Once again we have .w/ D pn C
n1 X zDk
pz P XnkC1 : : : Xn D wjTw.1/ D z :
Context Trees, Variable Length Markov Chains and Dynamical Sources
27
When z n k, due to the renewal property, the conditional probability can be rewritten as .nz/ P XnkC1 : : : Xn D wjTw.1/ D z D q pref .w/ .w/: w1
wknCz
wk n
z
w1
wnzC1
wk
When z > n k (see figure above), .nz/ P w at njTw1 D z D 1fwnzC1:::wk Dw1 :::wknCz g q pref .w/ .wknCzC1 : : : wk /; ki this equality holding if n k C i ¤ z. But when ‚…„ƒ z D n k C i , because the first occurrence of w w D 10 0 is at z, necessarily wk D 1 and hence i D k, and 6 6 z D n which contradicts z < n. Consequently for nkCi n z D n k C i we have P XnkC1 : : : Xn D wjTw1 D z D 0 D 1fwnzC1 :::wk Dw1 :::wknCz g :
Finally one gets .w/ D pn C
nk X
pz q
.nz/ .w/ pref .w/
zD1 n1 X
C
pz 1fwnzC1 :::wk Dw1 :::wknCz g q
.nz/ .wknCzC1 : : : wk /; pref .w/
zDnkC1
and hence x k .w/=.1 x/
˚w.1/ .x/ D 1C
1 X j Dk
.j / x j q pref .w/ .w/ C
k1 X j D1
x
j
.j / 1fwj C1 :::wk Dw1 :::wkj g q pref .w/ .wkj C1
: : : : wk /
Proceeding exactly in the same way by induction on r, we get the expression of Theorem 3 for the r-th occurrence. u t Remark 9. The case of internal nodes w D 0k is more intricate, due to the absence of any symbol 1 allowing a renewal argument. Nevertheless, for the forthcoming
28
P. Cénac et al.
q1 q00
I1 q011
q0100
I0
q01011 q(01)n00 q(01)∞
I0
I1
q(01)n+11
Fig. 7 Bamboo blossom probabilized context tree (on the left) and the associated dynamical system (on the right)
applications, we will not need the explicit expression of the generating function of such words occurrences.
4.2 The Bamboo Blossom 4.2.1 Stationary Probability Measures Consider the probabilized context tree given by the left side of Fig. 7. The data of a corresponding VLMC consist in probability measures on A indexed by the two families of finite contexts .q.01/n 1 /n0 and .q.01/n 00 /n0 together with a probability measure on the infinite context q.01/1 . As before, assuming that is a stationary probability measure on L , we compute the probabilities of any .w/, w being an internal node or w being a context, as functions of the data and of both .1/ and .00/. Determination of stationary probabilities of cylinders based on both contexts 1 and 00 then leads to assumptions that guarantee existence and uniqueness of such a stationary probability measure. Computation of .w/; w Context Two families of cylinders, namely L 1.10/n and L 00.10/n, correspond to contexts. For any n 0, .1.10/nC1/ D .1.10/n/q.01/n 1 .1/q1 .0/ and .00.10/nC1/ D
Context Trees, Variable Length Markov Chains and Dynamical Sources
29
.00.10/n/q.01/n 00 .1/q1 .0/. A straightforward induction implies thus that for any n 0, 8 < .1.10/n/ D .1/cn .1/ (19) : .00.10/n/ D .00/cn .00/ where c0 .1/ D c0 .00/ D 1 and 8 n1 Y ˆ ˆ n ˆ c .1/ D q .0/ q.01/k 1 .1/ n 1 ˆ ˆ ˆ < kD0 ˆ ˆ n1 ˆ Y ˆ n ˆ ˆ c .00/ D q .0/ q.01/k 00 .1/ 1 : n kD0
for any n 1. Computation of .w/; w Internal Node Two families of cylinders, L 0.10/n and L .10/n , correspond to internal nodes. By disjoint union of events, they are related by
.0.10/n / D ..10/n / .1.10/n/ ..10/nC1 / D .0.10/n / .00.10/n/
for any n 0. By induction, this leads to: 8n 0, 8 < .0.10/n / D 1 .1/Sn .1/ .00/Sn1 .00/ :
(20) ..10/n / D 1 .1/Sn1 .1/ .00/Sn1 .00/
where S1 .1/ D S1 .00/ D 0 and, for any n 0, 8 Pn < Sn .1/ D kD0 ck .1/ :
Sn .00/ D
Pn
kD0 ck .00/:
These formula give, by quotients, the conditional probabilities on internal nodes defined by (6) and appearing in formula (8). Computation of .1/ and of .00/ The context tree defines a partition of the set L of left-infinite sequences (see (3)). In the case of bamboo blossom, this partition implies 1 ..10/1 / D
X n0
.1.10/n/ C
X n0
.00.10/n/
(21)
30
P. Cénac et al.
D
X
.1/cn .1/ C
n0
We denote
X
.00/cn .00/:
(22)
n0
8 P < S.1/ D n0 cn .1/ :
S.00/ D
P
n0 cn .00/
2 Œ1; C1:
Note that the series S.1/ always converges. Indeed, the convergence is obvious if q1 .0/ ¤ 1; otherwise, q1 .0/ D 1 and q1 .1/ D 0, so that any cn .1/, n 1 vanishes and S.1/ D 1. In the same way, the series S.00/ is finite as soon as q1 .0/ ¤ 1. Proposition 4. (Stationary measure on a bamboo blossom) Let .Un /n0 be a VLMC defined by a probabilized bamboo blossom context tree. (i) Assume that q1 .0/ ¤ 1, then the Markov process .Un /n0 admits a stationary probability measure on L which is unique if and only if S.1/S.00/.1Cq1.0//¤ 0. (ii) Assume that q1 .0/ D 1. (ii.a) If S.00/ D 1, then .Un /n0 admits D 12 ı.10/1 C 12 ı.10/1 1 as unique stationary probability measure on L . (ii.b) If S.00/ < 1, then .Un /n0 admits a one parameter family of stationary probability measures on L . Proof. (i) Assume that q1 .0/ ¤ 1 and that is a stationary probability measure. Applying (7) gives ..10/1 / D q1 .0/q.01/1 .1/..10/1 /
(23)
and consequently ..10/1 / D 0. Therefore, (21) becomes S.1/.1/ C S.00/.00/ D 1. We get another linear equation on .1/ and .00/ by disjoint union of events: .0/ D 1 .1/ D .10/ C .00/ D .1/q1 .0/ C .00/. Thus .1/ and .00/ are solutions of the linear system 8 < S.1/.1/ C S.00/.00/ D 1 :
(24) Œ1 C q1 .0/ .1/ C .00/ D 1:
This system has a unique solution if and only if the determinantal assumption S.1/ S.00/ Œ1 C q1 .0/ ¤ 0 is fulfilled, which is a very light assumption (if this determinant happens to be zero, it suffices to modify one value of some qu , u context for the assumption to be satisfied). Otherwise, when the determinant vanishes, System (24) is reduced to its second equation, so that it admits a one parameter family of solutions. Indeed,
Context Trees, Variable Length Markov Chains and Dynamical Sources
1 S.1/ 1 C q1 .0/.1 q1 .0// C
X
31
q1 .0/n .1 q1 .0// D 1 C q1 .0/
n2
and S.00/ 1, so that S.1/ S.00/.1 C q1 .0// D 0 implies that S.1/ D 1 C q1 .0/ and S.00/ D 1. In any case, System (24) has at least one solution, which ensures the existence of a stationary probability measure with formula (20), (19) and (8) by a standard argumentation. Assertions on uniqueness are straightforward. (ii) Assume that q1 .0/ D 1. This implies q1 .1/ D 0 and consequently S.1/ D 1. Thus, .1/ and .00/ are solutions of 8 < .1/ C S.00/.00/ D 1 ..10/1 / :
(25) 2.1/ C .00/ D 1:
so that, since S.00/ 1, the determinantal condition S.1/ S.00/.1 C q1 .0// ¤ 0 is always fulfilled. (ii.a) When S.00/ D 1, .00/ D 0, .1/ D 12 and ..10/1 / D 12 . This defines uniquely a stationary probability measure . Because of (23), q.01/1 .1/ D 1 so that ..10/1 1/ D ..10/1 // D 12 . This shows that D 12 ı.10/1 C 12 ı.10/1 1 . (ii.b) When S.00/ < 1, if we fix the value a D ..10/1 /, System (25) has a unique solution that determines in a unique way the stationary probability measure a . t u
4.2.2 The Associated Dynamical System The vertical partition is made of the intervals I.01/n 00 and I.01/n 1 for n 0. The horizontal partition consists in the intervals I0.01/n 00 , I1.01/n 00 , I0.01/n 1 and I1.01/n 1 for n 0, together with the two intervals coming from the infinite context, namely I0.01/1 and I1.01/1 . If we make an hypothesis to ensure ..10/1 / D 0, then these two last intervals become two accumulation points of the horizontal partition, a0 and a1 . The respective positions of the intervals and the two accumulation points are given by the alphabetical order 0.01/n1 00 < 0.01/n00 < 0.01/1 < 0.01/n1 < 0.01/n1 1 1.01/n1 00 < 1.01/n00 < 1.01/1 < 1.01/n1 < 1.01/n1 1 Lemma 5. If .q.01/n 00 .0//n2N and .q.01/n 1 .0//n2N converge, then T is right and left differentiable in a0 and a1 – with possibly infinite derivatives – and T`0 .a0 / D lim
n!1
1 1 ; Tr0 .a0 / D lim n!1 q.01/n 1.0/ q.01/n 00.0/
32
P. Cénac et al.
T`0 .a1 / D lim
n!1
1 1 ; Tr0 .a1 / D lim : n!1 q.01/n 1.1/ q.01/n 00.1/ t u
Proof. We use Lemma 4. 4.2.3 Dirichlet Series
As for the infinite comb, the Dirichlet series of a source generated by a stationary bamboo blossom can be explicitly computed as a function of the SVLMC data. For simplicity, we assume that the generic Condition (i) of Proposition 4 is satisfied. An internal node is of the form .01/n or .01/n 0 while a context writes .01/n 00 or .01/n1. Therefore, by disjoint union, X
.s/ D A.s/ C
n0;w2W
where A.s/ D
X
.w00.10/n/s C
X
.w1.10/n /s
n0;w2W
..10/n /s C
n0
X
.0.10/n/s
n0
is explicitly given by formula (20) and (24). Because of the renewal property of the bamboo blossom, formula (7) leads by two straightforward inductions to .w00.10/n/ D .w00/cn .00/ and .w1.10/n/ D .w1/cn .1/ for any n 0. This implies that .s/ D A.s/ C 00 .s/
X
cns .00/ C 1 .s/
n0
where 00 .s/ D
X
X
cns .1/
n0
.w00/s and 1 .s/ D
w2W
X
.w1/s :
w2W
It remains to compute both Dirichlet series 00 and 1 , which can be done by a similar procedure. By disjoint union of finite words, 00 .s/ D A00 .s/ C
X
.w00.10/n00/s C
n0;w2W
where A00 .s/ D
X n0
X
.w1.10/n00/s
(26)
n0;w2W
..10/n 00/s C
X
.0.10/n00/s
n0
and 1 .s/ D A1 .s/ C
X n0;w2W
.w00.10/n1/s C
X n0;w2W
.w1.10/n 1/s
(27)
Context Trees, Variable Length Markov Chains and Dynamical Sources
with A1 .s/ D
X
..10/n 1/s C
n0
X
33
.0.10/n1/s :
n0
Computation of A1 and A00 By disjoint union and formula (7), ..10/nC1 00/ D .0.10/n00/ .00.10/n/q.01/n 00 .0/q00 .0/; n 0 and
.0.10/n00/ D ..10/n 00/ .1.10/n/q.01/n 1 .0/q00 .0/; n 1
where .00.10/n/ and .1.10/n/ are already computed probabilities of contexts (19). Since .000/ D .00/q00 .0/, one gets recursively ..10/n 00/ and .0.10/n00/ from these two relations as functions of the data. This computes A00 . A very similar argument leads to an explicit form of A1 . Ultimate Computation of 1 and 00 Start with (26) and (27). As above, for any n 0, by induction and with formula (7), .w00.10/n00/ D .w00/cn .00/q.01/n00 .0/q00 .0/: In the same way, but only when n 1, .w1.10/n 00/ D .w1/cn .1/q.01/n 1 .0/q00 .0/: Similar computations lead to similar formula for .w00.10/n1/ and .w1.10/n1/, for any n 0. So, (26) and (27) lead to 00 .s/ D A00 .s/ C 100 .s/ C 00 .s/B00 .s/ C 1 .s/B1 .s/
(28)
where B00 .s/ and B1 .s/ are explicit functions of the data and where 100 .s/ D
X
.w100/:
w2W
As above, after disjoint union of words, splitting by formula (7) and double induction, one gets 100 .s/ D A100 .s/ C 00 .s/C00 .s/ C 1 .s/C1 .s/ where A100 .s/, C00 .s/ and C1 .s/ are explicit series, functions of the data. Replacing 100 by this value in formula (28) leads to a first linear equation between 00 .s/ and 1 .s/. A second linear equation between them is obtained from (27) by similar arguments. Solving the system one gets with both linear equations gives an explicit form of 00 .s/ and 1 .s/ as functions of the data, completing the expected computation.
34
P. Cénac et al.
4.2.4 Generating Function for the Exact Distribution of Word Occurrences in a Sequence Generated by a Bamboo Blossom Let us consider the process X D .Xn /n0 of final letters of .Un /n0 in the particular case of a SVLMC defined by a bamboo blossom. We only deal with finite words w such that w is not an internal node, i.e. w is a finite context or w … T . One can see that such a word of length k > 1 can be written in the form 11.10/`1p or 00.10/`1p , with p 2 f0; 1g and ` 2 N, where stands for any finite word. Proposition 5. For a SVLMC defined by a bamboo blossom, with notations of Sect. 4.1.4, the generating function of the first occurrence of a finite word w D w1 : : : wk is given for jxj < 1 by ˚w.1/ .x/ D
x k .w/ .1 x/Sw .x/
and the generating function of the r th occurrence of w is given by ˚w.r/ .x/ D ˚w.1/ .x/ 1
1 Sw .x/
r1
;
where (i) if w is of the form 00.10/` or 11.01/`0, with ` 2 N, Sw .x/ is defined in Proposition 3 and (ii) if w is of the form 00.10/`1, ` 2 N, Sw .x/ D Cw .x/ C
1 X j Dk
Cw .x/ D 1 C
k1 X j D1
.j /
q1.01/` 00 .w/x j ;
.j / 1fwj C1 :::wk Dw1 :::wkj g q1.01/` 00 wj C1 : : : wk x j :
and if w is of the form 11.01/`, ` 2 N, Sw .x/ D Cw .x/ C
1 X j Dk
Cw .x/ D 1 C
k1 X j D1
.j /
q.10/` 11 .w/x j ;
.j / 1fwj C1 :::wk Dw1 :::wkj g q.10/` 11 wj C1 : : : wk x j :
Proof. (i) We first deal with the words w such that
pref .w/ D .01/` 00 or
pref .w/ D .01/` 1:
Context Trees, Variable Length Markov Chains and Dynamical Sources
35
.1/
Let us denote by pn the probability that Tw D n. Proceeding exactly in the same way as for Proposition 3, from the decomposition .w/ D pn C
n1 X
pz P XnkC1 : : : Xn D wjTw.1/ D z ;
zDk
and due to the renewal property of the bamboo, one has .w/ D pn C
nk X
ˇ
ˇ pz P Un 2 L wˇUz 2 L suff.w/
zDk
C
ˇ
ˇ pz 1fwnzC1 :::wk Dw1 :::wknCz g P Un 2 L wˇUz 2 L suff.w/
n1 X zDnkC1
where suff.w/ is the suffix of w equal to the reversed word of pref .w/. Hence, for x < 1, it comes C1 C1 nk X X X x k .w/ .nz/ D pn x n C xn pz q pref .w/ .w/ 1x nDk
C
C1 X nDk
nDk
xn
n1 X
zDk
pz 1fwnzC1 :::wk Dw1 :::wknCz g q
.nz/ .w/ pref .w/
zDnkC1 .1/
which leads to the expression of ˚w .x/ given in Proposition 3. The r th occurrence can be derived exactly in the same way from the decomposition fw at ng D fTw.1/ D ng [ fTw.2/ D ng [ : : : [ fTw.r/ D ng [ fTw.r/ < n and w at ng: (ii) In the particular case of words w D 00.10/`1, the main difference is that the context 1 is not sufficient for the renewal property. The computation relies on the equality P XnkC1 : : : Xn D wjTw.1/ D z
D P XnkC1 : : : Xn D wjXz2`2 : : : Xz D 00.10/` 1 : The sketch of the proof remains the same replacing q The case w D 11.01/` is analogous.
pref .w/ .w/
by q1.01/` 00 .w/. t u
36
P. Cénac et al.
5 Some Remarks, Extensions and Open Problems 5.1 Stationary Measure for a General VLMC Infinite comb and bamboo blossom are two instructive but very particular examples, close to renewal processes. Nevertheless, we think that an analogous of Propositions 1 or 4 can be written for a VLMC defined by a general context tree with a finite or countable number of infinite branches. In order to generalize the proofs, it is clear that formula (8) in Lemma 1 is crucial. In this formula, for a given finite word w D ˛1 : : : ˛N 2 W it is important to check whether the subwords pref .˛1 : : : ˛k /; k < N; are internal nodes of the tree or not. Consequently, the following concept of minimal context is natural. Definition 12. (Minimal context) Define the following binary relation on the set of the finite contexts as follows: 8u; v 2 C F ; u v ” 9w; w0 2 W ; v D wuw0 (in other words u is a sub-word of v). This relation is a partial order. In a context tree, a finite context is called minimal when it is minimal for this partial order on contexts. Remark 10. (Alternative definition of a minimal context) Let T be a context tree. Let c D ˛N : : : ˛1 be a finite context of T . Then c is minimal if and only if 8k 2 f1; : : : ; N 1g; pref .˛1 : : : ˛k / … C F .T /. Example 1. In the infinite comb, the only minimal context is 1. In the bamboo blossom, the minimal contexts are 1 and 00. Remark 11. There exist some trees with infinitely many infinite leaves and a finite number of minimal contexts. Take the infinite comb and at each 0k branch another infinite comb. In such a tree, the finite leaf 10 is the only minimal context. Nonetheless, a tree with a finite number of infinite contexts has necessarily a finite number of minimal contexts. As one can see for the infinite comb or for the bamboo blossom (see Sects. 4.1.1 and 4.2.1), minimal contexts play a special role in the computation of stationary probability measures. First of all, when is a stationary probability measure and w a finite word such that w … T , formula (8) implies that .w/ is a rational monomial of the data qc .˛/ and of the .u/ where u belongs to T . This shows that any stationary probability is determined by its values on the nodes of the context tree. In both examples, we compute these values as functions of the data and of the .m/, where m are minimal contexts, and we finally write a rectangular linear system satisfied by these .m/. Assuming that this system has maximal rank can be viewed as making an irreducibility condition for the Markov chain on L . We conjecture that this situation happens in any case of VLMC.
Context Trees, Variable Length Markov Chains and Dynamical Sources
37
Fig. 8 .n C 1/-teeth comb probabilized context tree
q1 q01 q001 q0n+11
q0n1
In the following example, we detail the above procedure, in order to understand how the two main principles (the partition (3) and the disjoint union) give the linear system leading to the irreducibility condition. Example 2. Let T be a probabilized context tree corresponding to Fig. 8 (finite comb with n C 1 teeth). There are two minimal contexts: 1 and 0nC1 . Assume that is a stationary probability measure on L . Like in the case of the infinite comb, the probability of a word that corresponds to a teeth is .10k / D .1/ck , 0 k n where ck is the product defined by (12). Also, the probabilities of the internal nodes and of the handle are .0k / D 1 .1/Sk1 ; 0 k n C 1; Pp where Sp WD j D0 cj . By means of these formula, is determined by .1/. In order to compute .1/, one can proceed P as follows. First, by the partition principle (3), we have 1 D .0nC1 / C .1/ nkD0 ck . Secondly, by disjoint union, .0nC1 / D .0nC2 / C .10nC1 / D .0nC1 /q0nC1 .0/ C .10n /q0n 1 .0/: This implies the linear relation between both minimal contexts probabilities: .0nC1 / C Sn .1/ D 1 q0nC1 .1/.0nC1 / q0n 1 .0/cn .1/ D 0: In particular, this leads to the irreducibility condition q0nC1 .1/Sn C q0n 1 .0/cn ¤ 0 for the VLCM to admit a stationary probability measure. One can check that this irreducibility condition is the classical one for the corresponding A -valued Markov chain of order n C 1. Example 3. Let T be a probabilized context tree corresponding to Fig. 5 (four flower bamboo). This tree provides another example of computation procedure using formula (7) and (8), the partition principle (3) and the disjoint union. This VLMC admits a unique stationary probability measure if the determinantal condition q00 .1/Œ1 C q1 .0/ C q1 .0/2 q010 .0/ C q1 .0/q1 .1/q011 .0/ ¤ 0 is satisfied; it is fulfilled if none of the qc is trivial.
38
P. Cénac et al.
5.2 Tries In a first kind of problems, n words independently produced by a source are inserted in a trie. There are results on the classical parameters of the trie (size, height, path length) for a dynamical source Clément et al. ([6]), which rely on the existence of a spectral gap for the underlying dynamical system. We would like to extend these results to cases when there is no spectral gap, as may be guessed in the infinite comb example. Another interesting application consists in producing a suffix trie from one sequence coming from a VLMC dynamical source, and analyzing its parameters. For his analysis, Szpankowski [26] puts some mixing assumptions (called strong ˛-mixing) on the source. A first direction consists in trying to find the mixing type of a VLMC dynamical source. In a second direction, we plan to use the generating function for the occurrence of words to improve these results. Acknowledgements We are very grateful to Antonio Galves, who introduced us to the challenging VLMC topics. We warmly thank Brigitte Vallée for valuable and stormy discussions.
References 1. M. Abadi, A. Galves, Inequalities for the occurrence times of rare events in mixing processes. The state of the art. Markov Proc. Relat. Field. 7(1), 97–112 (2001) 2. M. Abadi, B. Saussol, Stochastic processes and their applications. 121(2), 314–323 3. P. Billingsley, Probability and Measure, 3rd edn. Wiley Series in Probability and Mathematical Statistics (Wiley, New York, 1995) 4. G. Blom, D. Thorburn, How many random digits are required until given sequences are obtained? J. Appl. Probab. 19, 518–531 (1982) 5. P. Bühlmann, A.J. Wyner, Variable length markov chains. Ann. Statist. 27(2), 480–513 (1999) 6. J. Clément, P. Flajolet, B. Vallee, Dynamical sources in information theory: Analysis of general tries. Algorithmica 29, 307–369 (2001) 7. F. Comets, R. Fernandez, P. Ferrari, Processes with long memory: Regenerative construction and perfect simulation. Ann. Appl. Prob. 12(3), 921–943 (2002) 8. P. Flajolet, M. Roux, B. Vallée, Digital trees and memoryless sources: From arithmetics to analysis. DMTCS Proc. AM, 233–260 (2010) 9. J.C. Fu, Bounds for reliability of large consecutive-k-out-of-n W f system. IEEE Trans. Reliab. 35, 316–319 (1986) 10. J.C. Fu, M.V. Koutras, Distribution theory of runs: A markov chain approach. J. Amer. Statist. Soc. 89, 1050–1058 (1994) 11. S. Gallo, N. Garcia, Perfect simulation for stochastic chains of infinite memory: Relaxing the continuity assumptions. pp. 1–20 (2010) [arXiv:1105.5459v1] 12. A. Galves, E. Löcherbach, Stochastic chains with memory of variable length. TICSP Series 38, 117–133 (2008) 13. H. Gerber, S. Li, The occurrence of sequence patterns in repeated experiments and hitting times in a markov chain. Stoch. Process. Their Appl. 11, 101–108 (1981) 14. T.E. Harris, On chains of infinite order. Pac. J. Math. 5, 707–724 (1955) 15. P. Jacquet, W. Szpankowski, Autocorrelation on words and its applications. Analysis of suffix trees by string-ruler approach. J. Combin. Theor. A.66, 237–269 (1994)
Context Trees, Variable Length Markov Chains and Dynamical Sources
39
16. M.V. Koutras, Waiting Times and Number of Appearances of Events in a Sequence of Discrete Random Variables. Advances in Combinatorial Methods and Applications to Probability and Statistics, Stat. Ind. Technol., (Birkhäuser Boston, Boston, 1997), pp. 363–384 17. A. Lambert, S. Siboni, S. Vaienti, Statistical properties of a nonuniformly hyperbolic map of the interval. J. Stat. Phys. 72(5/6), 1305–1330 (1993) 18. S.-Y.R. Li, A martingale approach to the study of occurrence of sequence patterns in repeated experiments. Ann. Probab. 8(6): 1171–1176 (1980) 19. V. Pozdnyakov, J. Glaz, M. Kulldorff, J.M. Steele, A martingale approach to scan statistics. Ann. Inst. Statist. Math. 57(1), 21–37 (2005) 20. M. Régnier, A unified approach to word occurrence probabilities. Discrete Appl. Math. 104, 259–280 (2000) 21. G. Reinert, S. Schbath, M.S. Waterman, Probabilistic and statistical properties of words: An overview. J. Comput. Biol. 7(1/2), 1–46 (2000) 22. D. Revuz, Markov Chains. (North-Holland Mathematical Library, Amsterdam, 1984) 23. J. Rissanen, A universal data compression system. IEEE Trans. Inform. Theor. 29(5), 656–664 (1983) 24. S. Robin, J.J. Daudin, Exact distribution of word occurrences in a random sequence of letters. J. Appl. Prob. 36, 179–193 (1999) 25. V. Stefanov, A.G. Pakes, Explicit distributional results in pattern formation. Ann. Appl. Probab. 7, 666–678 (1997) 26. W. Szpankowski, A generalized suffix tree and its (un)expected asymptotic behaviors. SIAM J. Comput. 22(6), 1176–1198 (1993) 27. X-J. Wang, Statistical physics of temporal intermittency. Phys. Rev. A 40(11), 6647–6661 (1989) 28. D. Williams, Probability with Martingales. Cambridge Mathematical Textbooks (Cambridge University Press, Cambridge, 1991)
Martingale Property of Generalized Stochastic Exponentials Aleksandar Mijatovi´c , Nika Novak, and Mikhail Urusov
Abstract For a real Borel measurable function b, which satisfies certain integrability conditions, it is possible to define a stochastic integral of the process b.Y / with respect to a Brownian motion W , where Y is a diffusion driven by W . It is wellknown that the stochastic exponential of this stochastic integral is a local martingale. In this paper we consider the case of an arbitrary Borel measurable function b where it may not be possible to define the stochastic integral of b.Y / directly. However the notion of the stochastic exponential can be generalized. We define a non-negative process Z, called generalized stochastic exponential, which is not necessarily a local martingale. Our main result gives deterministic necessary and sufficient conditions for Z to be a local, true or uniformly integrable martingale. Keywords Generalized stochastic exponentials • Local martingales vs. true martingales • One-dimensional diffusions
AMS Classification: 60G44, 60G48, 60H10, 60J60
A. Mijatovi´c Department of Statistics, University of Warwick, Coventry, England, CV4 8JY, UK e-mail:
[email protected] N. Novak () University of Ljubljana, Ljubljana, Slovenia Department of Mathematics, Imperial College London, London, UK e-mail:
[email protected] M. Urusov Institute of Mathematical Finance, Ulm University, Ulm, Germany e-mail:
[email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__2, © Springer-Verlag Berlin Heidelberg 2012
41
42
A. Mijatovi´c et al.
1 Introduction A stochastic exponential of X is a process E .X / defined by 1 E .X /t D exp Xt X0 hX it 2 for some continuous local martingale X , where hX i denotes a quadratic variation of X . It is well-known that the process E .X / is also a continuous local martingale. Sufficient conditions for the martingale property of E .X / have been studied extensively in the literature because this question appears naturally in many situations. Novikov’s and Kazamaki’s sufficient conditions (see [6, 9]) for E .X / to be a martingale are particularly well-known. Novikov’s condition consists in E expf.1=2/hX it g < 1, t 2 Œ0; 1/. Kazamaki’s condition consists in that expf.1=2/X g should be a submartingale. Novikov’s criterion is of narrower scope than Kazamaki’s one but often easier to apply. In this respect let us note that none of conditions E expf.1=2 "/hX it g < 1 (with " > 0) is sufficient for E .X / to be a martingale (see Liptser and Shiryaev [7, Chap. 6]). For a further literature review see the bibliographical notes in the monographs Karatzas and Shreve [5, Chap. 3], Liptser and Shiryaev [7, Chap. 6], Protter [10, Chap. III], and Revuz and Yor [11, Chap. VIII]. In the case of one-dimensional processes, necessary and sufficient conditions for the process E .X / to be a martingale were recently studied by Engelbert and Senf in [3], Blei and Engelbert in [1] and Mijatovi´c and Urusov in [8]. In [3] X is a general continuous local martingale and the characterisation is given in terms of the Dambis–Dubins–Schwartz time-change that turns X into a Brownian motion. In [1] X is a strong Markov continuous local martingale and the condition is deterministic, expressed in terms of the speed measure of X . Rt In [8] the local martingale X is of the form Xt D 0 b .Yu / dWu for some measurable function b and a one-dimensional diffusion Y with drift and volatility driven by a Brownian motion W . In order to define the stochastic integral X , an 2 assumption that the function b 2 is locally integrable on the entire state space of the process Y is required. Under this restriction the characterization of the martingale property of E .X / is studied in [8], where the necessary and sufficient conditions are deterministic and are expressed in terms of functions ; and b only. In the present paper we consider an arbitrary Borel measurable function b. In this case the stochastic integral X can only be defined on some subset of the probability space. However, it is possible to define a non-negative possibly discontinuous process Z, known as a generalized stochastic exponential, on the entire probability space. It is a consequence of the definition that, if the function b satisfies the required local integrability condition, the process Z coincides with E .X /. We show that the process Z is not necessarily a local martingale. In fact Z is a local martingale if and only if it is continuous. We find a deterministic necessary and sufficient condition for Z to be a local martingale, which is expressed in terms of local integrability of
Martingale Property of Generalized Stochastic Exponentials
43
2
the quotient b 2 multiplied by a linear function. We also characterize the processes Z that are true martingales and/or uniformly integrable martingales. All the necessary and sufficient conditions are deterministic and are given in terms of functions ; and b. The paper is structured as follows. In Sect. 2 we define the notion of generalized stochastic exponential and study its basic properties. The main results are stated in Sect. 3, where we give a necessary and sufficient condition for the process Z defined by (8) and (12) to be a local martingale, a true martingale or a uniformly integrable martingale. Finally, in Sect. 4 we prove Theorem 3 that is central in obtaining the deterministic characterisation of the martingale property of the process Z. Appendix A contains an auxiliary fact that is used in Sect. 2.
2 Definition of Generalized Stochastic Exponential Let J D .l; r/ be our state space, where 1 l < r 1. Let us define a J -valued diffusion Y on a probability space .˝; F ; .Ft /t 2Œ0;1/ ; P/ driven by a stochastic differential equation dYt D .Yt / dt C .Yt / dWt ;
Y0 D x0 2 J;
where W is a .Ft /-Brownian motion, and , are real, Borel measurable functions defined on J that satisfy the Engelbert–Schmidt conditions .x/ ¤ 0
8x 2 J;
(1)
1 ; 2 L1loc .J /: 2 2
(2)
With L1loc .J / we denote the class of locally integrable functions, i.e. real Borel measurable functions defined on J that are integrable on every compact subset of J . Engelbert–Schmidt conditions guarantee existence of a weak solution that might exit the interval J and is unique in law (see [5, Chap. 5]). Denote by the exit time of Y . In addition, we assume that the boundary points are absorbing, i.e. the solution Y stays at the boundary point at which it exits on the set f < 1g: Let us note that we assume that .Ft / is generated neither by Y nor by W . We would like to define a process X as a stochastic integral of a process b.Y / with respect to Brownian motion W , where b W J ! R is an arbitrary Borel measurable function. Before further discussion, we should establish if the stochastic integral can be defined. Define a set A D fx 2 J I
b2 2
62 L1loc .x/g;
44
A. Mijatovi´c et al.
where L1loc .x/ denotes a space of real Borel measurable functions f such that R xC" x" jf .y/jdy < 1 for some " > 0. Then A is closed and its complement is a union of open intervals. Furthermore we can define maps ˛ and ˇ on J n A so that ˛.x/; ˇ.x/ 2 A [ fl; rg
and
x 2 .˛.x/; ˇ.x// J n A:
(3)
In other words ˛.x/ is the point in A [ fl; rg that is closest to x from the left side and ˇ.x/ is the closest point in A [ fl; rg from the right side. Let us also note that the equality \ L1loc .I / D L1loc .x/ (4) x2I
holds for any interval I by a simple compactness argument. For any x; y 2 J we define the stopping times x D infft 0I Yt D xg;
(5)
x;y D x ^ y ;
(6)
with the convention inf ; D 1, where c ^ d WD minfc; d g. Define the stopping time A D ^ A , where A D infft 0I Yt 2 Ag: (7) Then for all t 0 we have Z
t
b 2 .Yu / du < 1 P-a.s. on ft < A g:
0
This follows from Proposition 1 given in Appendix A and the fact that a continuous process Y on ft < A g reaches only values in an open interval that is a component of 2 the complement of A, where b 2 is locally integrable (note that (4) is applied here). Let us define An D fx 2 J I .x; A [ fl; rg/ n1 g; where .x; y/ D j arctan x arctan yj; x; y 2 JN , and set nA D infft 0I Yt 2 An g. Since nA < A on the set R t ^ A f A < 1g, we have 0 n b 2 .Yu /du < 1 P-a.s. Thus, we can define the stochastic R t ^ A R t ^ A R t ^ A integral 0 n b.Yu /dWu for every n. Since 0 n b.Yu /dWu and 0 nC1 b.Yu /dWu R t ^ A coincide on ft < nA g and nA % A , we can define 0 b.Yu /dWu as a P-a.s limit of integrals Z
t ^ A 0
Z b.Yu /dWu D lim
n!1 0
t ^nA
˚
b.Yu /dWu on t <
A
(Z
A
[
) b 2 .Yu /du < 1 :
0
In the case where A is not empty or Y exits the interval J , the stochastic exponential cannot be defined. However, we can define a generalized stochastic exponential Z in the following way for every t 2 Œ0; 1/:
Martingale Property of Generalized Stochastic Exponentials
45
8 Rt Rt expf 0 b.Yu /dWu 12 0 b 2 .Yu /dug if t < A ˆ ˆ < R R R Zt D expf 0 b.Yu /dWu 12 0 b 2 .Yu /dug if t A D ; 0 b 2 .Yu / du < 1 ˆ ˆ R : 0 if t A D A or t A D ; 0 b 2 .Yu / du D 1 (8) The different behaviour of Z on ft A D g and ft A D A g follows from the fact that after the exit time the process Y is stopped, while this does not happen R t 2 after A . The definition of the set A and Proposition 1 imply that the integral 0 b .Yu / du is infinite on the event ft > A g P-a.s. Therefore, we set Z D 0 on the set ft A D A g. Let us define the processes ZN t D exp
(Z
t ^ A 0
1 b .Yu / dWu 2
Z
t ^ A
) 2
b .Yu / du ;
(9)
0
R A where we set ZN t D 0 for t A on f A < 1; 0 b 2 .Yu /du D 1g. The process ZN has continuous trajectories: continuity at time A on the set R A f A < 1; 0 b 2 .Yu / d u D 1g follows from the Dambis–Dubins–Schwarz theorem on stochastic intervals; see Theorem 1.6 and Exercise 1.18n[11, Chap. V]. Noteo R A ˚ further that ZN t is strictly positive on the event t < A [ 0 b 2 .Yu /du < 1 P-a.s. and is equal to zero on its complement. Lemma 1. The process ZN D .ZN t /t 0 defined in (9) is a continuous local martingale. Remark 1. Fatou’s lemma and Lemma 1 imply that the process ZN is a continuous non-negative supermartingale. Proof. If the starting point x0 of the diffusion Y is in A, then A D 0 P-a.s and ZN is constant and hence a local martingale. If x0 2 J n A, then pick a decreasing (resp. increasing) sequence .˛n /n2N (resp. .ˇn /n2N ) in the open interval .˛.x0 /; ˇ.x0 // such that ˛n & ˛.x0 / (resp. ˇn % ˇ.x0 /), where ˛.x0 /; ˇ.x0 / are defined in (3). Assume also that ˛n < x0 < ˇn for all n 2 N. Note that ˛n ;ˇn % A P-a.s. The process M n D .Mtn /t 2Œ0;1/ , defined by Z Mtn D
t ^˛n ;ˇn
b.Yu /dWu
for all t 2 Œ0; 1/;
0
is a local martingale. Therefore its stochastic exponential E .M n / is also a local martingale and, since ˛n ;ˇn < A P-a.s., the equality ZN t ^˛n ;ˇn D E .M n /t holds P-a.s. for all t 0. For any m 2 N define the stopping time m D infft 0I ZN t mg. The stopped process ZN m ^˛n ;ˇn is a bounded local martingale and hence a martingale. Furthermore note that m % 1 P-a.s. as m tends to infinity. To prove that ZN m is a martingale for each m 2 N, note that the process ZN is stopped
46
A. Mijatovi´c et al.
^ at A , which implies the almost sure limit limn!1 ZN t m ˛n ;ˇn D ZN t m for every m ^˛n ;ˇn m P-a.s., the conditional dominated convergence t 2 Œ0; 1/. Since ZN t theorem implies the martingale property of ZN m for every m 2 N. This concludes the proof. t u
Define the process S D .St /t 2Œ0;1/ by Z
A
St D exp 0
1 b .Yu / dWu 2
Z
A 0
b .Yu / du 1ft A ;R A b 2 .Yu /du<1g : 2
0
(10)
Note that for any t 2 Œ0; 1/, P-a.s. on the set fA tg we have A < and hence the integrals in (10) are well-defined. We can express Z as Z D ZN S:
(11)
It is clear from this representation that Z is not necessarily a continuous process. Moreover, since that paths of S are non-decreasing, we have EŒZt jFs ZN s EŒSt jFs ZN s Ss D Zs : It follows that Z is a non-negative supermartingale and we can define Z1 D lim Zt : t !1
(12)
Note further that if x0 2 A, we have Z 0. A path of the process Z defined by (8) and (12) is equal to a path of a stochastic exponential if A D 1. Otherwise, if A < 1, it has one of the following forms: R 1. A < and 0 A b 2 .Yt /dt < 1 (see Fig. 1). R A 2. A < 1 and 0 b 2 .Yt /dt D 1 (see Fig. 2). R 3. < A and 0 b 2 .Yt /dt < 1 (see Fig. 3).
3 Main Results The case A D ; was studied by Mijatovi´c and Urusov in [8]. We generalize their result for the case where A ¤ ;.
3.1 The Case A D ; In this case we have
b2 2 L1loc .J /: 2
(13)
Martingale Property of Generalized Stochastic Exponentials
47
t Fig. 1 IfR A < , then the process Z is positive up to time A and is equal to zero afterwards. If the integral 0 A b 2 .Yt /dt is finite, then Zt approaches a positive value as t approaches A . Therefore, there is a jump at t D A
t R A Fig. 2 If A < 1 and 0 b 2 .Yt /dt D 1, then the process Z is zero after the time A . Since the limit of Zt is zero as t approaches A , there is no jump
The generalized stochastic exponential Z defined by (8) and (12) can now be written as (Z ) Z t ^ 1 t ^ 2 Zt D exp b .Yu / dWu b .Yu / du ; 2 0 0 R where we set Zt D 0 for t on f < 1; 0 b 2 .Yu / du D 1g. Note that in this N case Z is a local martingale by Lemma 1, since A D and hence Z D Z. e governed by the SDE Let us now define an auxiliary J -valued diffusion Y et dt C Y et dW et; et D . C b/ Y dY
e0 D x0 ; Y
48
A. Mijatovi´c et al.
t Fig. 3 If < A , the process Z is stopped after the exit time. Since equal to a positive constant for t
R 0
b 2 .Yt /dt is finite, Zt is
e ; .F e t /t 2Œ0;1/ ; e e F on some probability space .˝; P/. The coefficients C b and satisfy Engelbert–Schmidt conditions since b 2 L1loc .J / (this follows from (13)). Hence the SDE has a weak solution, unique in law and possibly explosive. As with e and assume that the boundary points diffusion Y , we denote by e the exit time of Y are absorbing. For an arbitrary c 2 J we define the scale functions s;e s and their derivatives ;e : 2.y/ dy ; x 2 J; 2 c .y/ Z x 2b.y/ e .x/ D .x/ exp dy ; x 2 J; .y/ c Z x s.x/ D .y/ dy; x 2 JN ; Z .x/ D exp
Z
x
(14)
c x
e s.x/ D
e .y/ dy;
x 2 JN :
c
For any a 2 .l; r, L1loc .a/ denotes the set of allR Borel measurable functions f W a J ! R such that there exists b < a, b 2 J , with b jf .x/j dx < 1. Similarly, we define the set L1loc .aC/ for any a 2 Œl; r/. We say that the endpoint r is good if s.r/ < 1
and
.s.r/ s/b 2 2 L1loc .r/: 2
Martingale Property of Generalized Stochastic Exponentials
49
It is equivalent to show that e s.r/ < 1
.e s.r/ e s/b 2 2 L1loc .r/: e 2
and
The endpoint l is good if s.l/ > 1
and
.s s.l//b 2 2 L1loc .lC/; 2
e s.l/ > 1
and
.e s e s.l//b 2 2 L1loc .lC/: e 2
or equivalently
If an endpoint is not good, we say it is bad. The good and bad endpoints were introduced in [8], where one can also find the proof of equivalences above. We will use the following terminology: e exits at r means e e Y P.e < 1; limt %e Y t D r/ > 0; e exits at l means e e Y P.e < 1; limt %e Y t D l/ > 0. Define
Z e v.x/ D
x
c
e s.x/ e s.y/ dy; e .y/ 2 .y/
x 2 J;
(15)
and e v.r/ D lim e v.x/; x%r
e v.l/ D lim e v.x/ x&l
(note thate v is decreasing on .l; c and increasing on Œc; r/). Feller’s test for explosions (see [5, Chap. 5, Theorem 5.29]) tells us that: e exits at the boundary point r if and only if 1. Y e v.r/ < 1: It is equivalent to check (see [2, Chap. 4.1]) e s.r/ < 1
and
e s.r/ e s 2 L1loc .r/: e 2
e exits at the boundary point l if and only if 2. Y e v.l/ < 1;
(16)
50
A. Mijatovi´c et al.
which is equivalent to e s.l/ > 1
and
e s e s.l/ 2 L1loc .lC/: e 2
e Remark 2. The endpoint r (resp. l) is bad whenever one of the processes Y and Y exits at r (resp. l) and the other does not. The following theorems are reformulations of Theorems 2.1 and 2.3 in [8]. Theorem 1. Let the functions ; and b satisfy conditions (1), (2) and (13). Then e does not exit at the bad endpoints. the process Z is a martingale if and only if Y Theorem 2. Let the functions ; and b satisfy conditions (1), (2) and (13). Then Z is a uniformly integrable martingale if and only if one of the conditions .a/ .d / below is satisfied: .a/ .b/ .c/ .d /
b D 0 a.e. on J with respect to the Lebesgue measure; r is good and e s.l/ D 1; l is good and e s.r/ D 1; l and r are good.
3.2 The Case A ¤ ; In the rest of the paper we assume that x0 … A; since x0 2 A implies that Z 0. The following example shows that even when A is not empty we can get a martingale or a uniformly integrable martingale defined by (8) and (12). Example 1. .i / Let us consider the state space J D R, the coefficients of the SDE D 0; D 1, the starting point of the diffusion x0 > 0 and the function b such that b.x/ D x1 for x 2 R n f0g and b.0/ D 0. Then A D f0g and Yt D Wt ; W0 D x0 . Using Itô’s formula and the fact that Brownian motion does not exit at infinity, we get for t < 0 Z
t
Zt D exp 0
D
1 1 dWu Wu 2
Z
t 0
1 du Wu2
1 Wt x0
and Zt D 0 for t 0 . Hence, Zt D
1 x0 Wt ^0
that is a martingale.
Martingale Property of Generalized Stochastic Exponentials
51
.i i / Using the same functions ; and b as above on a state space J D .1; x0 C 1/ we get 1 Zt D Wt ^0;x0 C1 ; x0 which is a uniformly integrable martingale. Let for any x 2 J n A the points ˛.x/; ˇ.x/ 2 A [ fl; rg be as in (3). Then 2 L1loc .˛.x/; ˇ.x//. Therefore, on .˛.x/; ˇ.x// functions ; and b satisfy the same conditions as in the previous subsection. e with For any starting point x0 2 J n A we can define an auxiliary diffusion Y state space .˛.x0 /; ˇ.x0 // driven by the SDE b2 2
et dW et; et D . C b/ Y et dt C Y dY
e0 D x0 ; Y
e ; .F e t /t 2Œ0;1/ ; e e F on some probability space .˝; P/. There exists a unique weak solution of this equation since coefficients satisfy the Engelbert–Schmidt conditions. As in the previous subsection we can define good and bad endpoints. Let functions ;e ; s;e s and e v be defined by (14), (15) and (16) with c 2 .˛.x0 /; ˇ.x0 //. We say that the endpoint ˇ.x0 / is good if s.ˇ.x0 // < 1
and
.s.ˇ.x0 // s/b 2 2 L1loc .ˇ.x0 //: 2
It is equivalent to show and sometimes easier to check that e s.ˇ.x0 // < 1
and
s/b 2 .e s.ˇ.x0 // e 2 L1loc .ˇ.x0 //: 2 e
The endpoint ˛.x0 / is good if s.˛.x0 // > 1
and
.s s.˛.x0 ///b 2 2 L1loc .˛.x0 /C/; 2
and
.e s e s.˛.x0 ///b 2 2 L1loc .˛.x0 /C/: e 2
or equivalently e s.˛.x0 // > 1
If an endpoint is not good, we say it is bad. The following theorem plays a key role in all that follows. Theorem 3. Let ˛ W J n A ! A [ fl; rg be the function defined by (3) and assume that the starting point x0 2 J n A of diffusion Y satisfies ˛.x0 / > l. Denote ˛0 D ˛.x0 / 2 A. Then:
52
A. Mijatovi´c et al.
.a/ .x A .b/ .x A
R 2 ˛0 / b 2 .x/ 2 L1loc .˛0 C/ ” 0 ˛0 b 2 .Yt / dt < 1 P-a.s. on f˛0 D < 1g; R ˛ 2 ˛0 / b 2 .x/ … L1loc .˛0 C/ ” 0 0 b 2 .Yt / dt D 1 P-a.s. on f˛0 D < 1g.
Remark 3. (i) The assumption ˛.x0 / > l means that Theorem 3 deals with the situation where the set A contains points to the left of the starting point x0 . (ii) Clearly, Theorem 3 has its analogue for ˇ0 D ˇ.x0 / < r, i.e. the case when A contains points to the right of x0 . The deterministic criterion in this case takes the form b2 .x/ 2 L1loc .ˇ0 / 2 Z ˇ 0 ” b 2 .Yt / dt < 1 P-a.s. on fˇ0 D A < 1g;
(17)
b2 .x/ … L1loc .ˇ0 / 2 Z ˇ 0 b 2 .Yt / dt D 1 P-a.s. on fˇ0 D A < 1g: ”
(18)
.ˇ0 x/
0
.ˇ0 x/
0
(iii) Note that P.˛0 D A < 1/ > 0. Indeed, if ˇ0 D r, then f˛0 D A < 1g D f˛0 < 1g; if ˇ0 < r, then f˛0 D A < 1g D f˛0 < ˇ0 g. In both cases P.˛0 D A < 1/ > 0 by [2, Theorem 2.11]. (iv) Since the diffusion Y starts at x0 , P-a.s. we have the following implications ˛0 ; ˇ0 2 A H) A D ˛0 ;ˇ0 ; ˛0 2 A; ˇ0 D r H) A D ˛0 ; ˛0 D l; ˇ0 2 A H) A D ˇ0 :
(19) (20) (21)
The case ˛0 D l; ˇ0 D r cannot occur since A ¤ ;. (v) Note that right-hand sides of the equivalences in .a/ and .b/ inR Theorem 3 are ˛ not negations of each other. If it does not hold that the integral 0 0 b 2 .Yt / dt is finite P-a.s. on the set f˛0 D A < 1g, then it must be infinite on some subset of f˛0 D A < 1g of positive probability, which may be strictly smaller than P.˛0 D A < 1/. For any starting point x0 2 J n A of diffusion Y , define the set
Z
B.x0 / D x 2 J \ f˛0 ; ˇ0 gI
x
b .Yt / dt D 1 P-a.s. on fx D A < 1g : 2
0
Note that B.x0 / is contained in A and that it contains at most two points. Theorem 3 implies that ˛0 2 B.x0 / if and only if the deterministic condition in (b) is satisfied.
Martingale Property of Generalized Stochastic Exponentials
53
Similarly ˇ0 2 B.x0 / is equivalent to the deterministic condition in (18). Therefore Theorem 3 yields a deterministic description of the set B.x0 /. We can now give a deterministic characterisation for a generalized stochastic exponential Z to be a local martingale and a true martingale. Theorem 4. .i / The generalized stochastic exponential Z is a local martingale if and only if ˛.x0 /; ˇ.x0 / 2 B.x0 / [ fl; rg. .ii/ The generalized stochastic exponential Z is a martingale if and only if Z is a local martingale and at least one of the conditions (a)-(b) below is satisfied and at least one of the conditions (c)-(d) below is satisfied: e does not exit at ˇ.x0 /, i.e.e (a) Y v.ˇ.x0 // D 1 or, equivalently, we have e s.ˇ.x0 // D 1
or
e s.ˇ.x0 // e s 1 e s.ˇ.x0 // < 1 and … L .ˇ.x // I 0 loc e 2
(b) ˇ.x0 / is good, e does not exit at ˛.x0 /, i.e.e (c) Y v.˛.x0 // D 1 or, equivalently, we have e s.˛.x0 // D 1
or
e s e s.˛.x0 // 1 e s.˛.x0 // > 1 and … Lloc .˛.x0 /C/ I e 2
(d) ˛.x0 / is good. Remark 4. Part (ii) of Theorem 4 says that Z is a martingale if and only if the e can exit only at the good endpoints. .˛.x0 /; ˇ.x0 //-valued process Y Proof. .i / We can write Z D ZN S as in (11). The process ZN is a continuous local martingale by Lemma 1. Suppose that Z is a local martingale. Then S can be written as a sum of two local martingales and therefore, it is also a local martingale. It follows that S is a supermartingale (since it is non-negative). Since A > 0 (we are assuming that x0 2 J n A) and S0 D 0, S should be almost surely R equal to 0. By definition (10) of S this happens if and only if P A < 1; 0 A b 2 .Yu /du < 1 D 0, which is by the definition of the set B.x0 / and (19)–(21) equivalent to ˛.x0 /; ˇ.x0 / 2 B.x0 / [ fl; rg. N Since .ii/ To get at least a local martingale S needs to be zero P-a.s. Then Z D Z. the values of Y on Œ0; A / do not exit the interval .˛.x0 /; ˇ.x0 //; the conditions of Theorem 1 are satisfied and the result follows. t u Similarly, we can characterize uniformly integrable martingales. We can use characterization in Theorem 2 for the process ZN defined by (9). As above, for ˛.x0 /; ˇ.x0 / 2 B.x0 / [ fl; rg the process Z defined by (8) and (12) coincides with N Otherwise, Z is not even a local martingale. Z. Theorem 5. The process Z is a uniformly integrable martingale if and only if Z is a local martingale and at least one of the conditions .a/ .d / below is satisfied: .a/ b D 0 a.e. on .˛.x0 /; ˇ.x0 // with respect to the Lebesgue measure;
54
A. Mijatovi´c et al.
.b/ ˛.x0 / is good and e s.ˇ.x0 // D 1; .c/ ˇ.x0 / is good and e s.˛.x0 // D 1; .d / ˛.x0 / and ˇ.x0 / are good. The following remark simplifies the application of Theorems 4 and 5 in specific situations. Remark 5. If ˛.x0 / 2 B.x0 /, then ˛.x0 / is not a good endpoint. Indeed, if s.˛.x0 // > 1, then we can write .s.x/ s.˛.x0 /// b2 .s.x/ s.˛.x0 ///b 2 .x/ D .x ˛.x0 // 2 .x/: 2 .x/ .x/ .x ˛.x0 //.x/ The first fraction is bounded away from zero, since it is continuous for x > ˛.x0 / and has a limit equal to 1 as x approaches ˛.x0 /. Since ˛.x0 / 2 B.x0 /, (b) of 2 2 0 ///b Theorem 3 implies .x ˛.x0 // b 2 .x/ 62 L1loc .˛.x0 /C/. Therefore .ss.˛.x 62 2 1 Lloc .˛.x0 /C/ and the conclusion follows. Similarly, ˇ.x0 / 2 B.x0 / implies that ˇ.x0 / is not a good endpoint.
4 Proof of Theorem 3 For the proof of Theorem 3 we first consider the case of a Brownian motion. Let W y be a Brownian motion with W0 D x0 . Denote by Lt .W / the local time of W at time t and level y. Let us consider a Borel function b W R ! R and set A D fx 2 R W b 2 … L1loc .x/g: We assume that x0 … A and define ˛0 ; ˇ0 (˛0 < ˇ0 ) so that ˛0 ; ˇ0 2 A [ f1; 1g and x0 2 .˛0 ; ˇ0 / R n A: W We additionally assume that ˛0 > 1. Below we use the notations xW ; x;y , and W A for the stopping times defined by (5), (6) and (7) respectively with Y replaced by W .
Lemma 2. If .x ˛0 /b 2 .x/ 2 L1loc .˛0 C/, then Z
˛W0 0
b 2 .Wt / dt < 1 P-a.s. on f˛W0 D AW g:
Remark 6. In the setting of Lemma 2 we have P.˛W0 < 1/ D 1 since Brownian motion reaches every level in finite time almost surely. Therefore the events f˛W0 D AW g and f˛W0 D AW < 1g are equal. Cf. with the formulation of Theorem 3.
Martingale Property of Generalized Stochastic Exponentials
55
Proof. Let .ˇn /n2N be an increasing sequence such that x0 < ˇn < ˇ0 and ˇn % ˇ0 . By [11, Chap. VII, Corollary 3.8] we get "Z
˛W0 ^ˇWn
E 0
# 0 b .Wt / dt D 2 ˇˇnn x ˛0
Z
x0
2
C
.y ˛0 /b 2 .y/ dy
˛0
0 2 ˇx0n˛ ˛0
Z
ˇn
.ˇn y/b 2 .y/ dy
x0
for every ˇn . Both integrals are finite since b 2 2 L1loc .˛0 ; ˇ0 / and .x R ˛W ^ W ˛0 /b 2 .x/ 2 L1loc .˛0 C/: Thus, EŒ 0 0 ˇn b 2 .Wt / dt < 1 and therefore R ˛W0 ^ˇWn 2 b .Wt / dt < 1 almost surely for every n. It remains to note that P-a.s. 0 on f˛W0 D AW g we have ˛W0 < ˇWn for sufficiently large n. This concludes the proof. t u R ˛W0 2 Lemma 3. If 0 b .Wt / dt < 1 on a set U with P.U / > 0, then .x ˛0 /b 2 .x/ 2 L1loc .˛0 C/. Proof. The idea of the proof comes from [4]. Using the occupation times formula we can write Z
˛W0
Z
0
˛0
Let us define a process Ry D Z
˛W0
Z
1
b 2 .Wt / dt D
x0
y
b 2 .y/L W .W / dy ˛0
y 1 L y˛0 ˛W
0
0
y
b 2 .y/L W .W / dy: ˛0
.W /. Then R is positive and we have
Z b 2 .Wt / dt
˛0
x0
Ry .y ˛0 /b 2 .y/ dy:
(22)
˛0
By [11, Chap. VI, Proposition 4.6], Laplace transform of Ry is EŒexpf Ry g D
1 1 C 2
for every y:
Hence, every random variable Ry has exponential distribution with EŒRy D 2. Denote by L an indicator function of a measurable set. We can write Z EŒLRy D E L
1 0
1fRy >ug du D
Z
1 0
EŒL1fRy >ug du:
By Jensen’s inequality we get a lower bound for the integrand
56
A. Mijatovi´c et al.
EŒL1fRy >ug D EŒ.L 1fRy ug /C .EŒL PŒRy u/C u
D .EŒL C e 2 1/C : Hence,
Z
1
EŒLRy
u
.EŒL C e 2 1/C du D C;
(23)
0
where C is a strictly positive constant if EŒL is strictly positive. R ˛W Let us suppose that we can choose L, so that EŒL 0 0 b 2 .Wt / dt is finite. Using Fubini’s Theorem and inequalities (22) and (23), we get " Z E L
˛W0
# Z b .Wt / dt 2
0
x0
Z EŒLRy .y ˛0 /b 2 .y/ dy C
˛0
x0
.y ˛0 /b 2 .y/ dy:
˛0
Therefore, .y ˛0 /b 2 .y/ 2 L1loc .˛0 C/ if we can find an indicator function L such R ˛W that EŒL is strictly positive and EŒL 0 0 b 2 .Wt / dt is finite. R ˛W Since 0 0 b 2 .Wt / dt < 1 on a set with positive measure, such L exists. Indeed, R ˛W denote by Ln an indicator function of the set Un D f 0 0 b 2 .Wt / dt ng. Then, for R ˛W0 2 every integer n, we have EŒL b .Wt / dt < 1. Since the sequence .Un /n2N n 0 S is increasing, U n2N Un and P.U / > 0, there exists an integer N such that P.UN / > 0 and therefore EŒLN > 0. t u Now we return to the setting of Sect. 2. Proof of Theorem 3. Suppose that 0 and is arbitrary. Since Yt is a continuous local martingale, by the Dambis–Dubins–Schwarz theorem we have Yt D BhY it for a Brownian motion B with B0 D x0 , defined possibly on an enlargement of the initial probability space. Using the substitution u D hY it , we get Z
˛Y0 0
Z
˛Y0
b .Yt / dt D 2
b2 .Yt / dhY it D 2
0
Z
hY i Y
˛0
0
b2 .Bu / du 2
P-a.s.
(24)
Furthermore, it is easy to see from Yt D BhY it that hY i˛Y D ˛B0 0
P-a.s. on f˛Y0 < 1g
(25)
and f˛Y0 D AY < 1g f˛B0 D AB g
P-a.s.
(26)
Now (24)–(26) imply the theorem in the case 0 and arbitrary. It only remains to prove the general case when both and are arbitrary. Let e satisfies SDE et D s.Yt /, where s is the scale function of Y . Then Y Y
Martingale Property of Generalized Stochastic Exponentials
57
et D e et dWt ; dY Y where e .x/ D s 0 .q.x//.q.x// and q is the inverse of s. Define e b D b ı q. Since s is strictly increasing, s.˛0 / > 1 by the assumption Y e˛ D s.Y˛ / D s.˛0 /, it follows that the equality ˛Y D e ˛0 > l, and Y s.˛0 / holds 0 0 0 P-a.s. Then we have Z
˛Y0
b 2 .Yt / dt D
Z e Y
0
s.˛0 /
e et dt: b2 Y
0
Besides, for some small positive " we have Z
s.˛0 C"/ s.˛0 /
Z ˛0 C" 2 e b .y/ s.y/ s.˛0 / b 2 .x/ .x s.˛ dy: // dx D 0 2 e .x/ 2 .y/ s 0 .y/ ˛0
0/ The fraction ss.y/s.˛ 0 .y/.y˛ / is continuous for y > ˛0 and tends to 1 as y & ˛0 . Hence 0 it is bounded and bounded away from zero on .˛0 ; ˛0 C ". It follows that .x 2 b2 ˛0 / b 2 .x/ 2 L1loc .˛0 C/ if and only if .x s.˛0 //e .x/ 2 L1loc .s.˛0 /C/: Then the e 2 result follows from the first part of the proof. t u
Remark 7. It is interesting to note that, in fact, both sets in (26) are P-a.s. equal. One can prove the reverse inclusion using the Engelbert–Schmidt construction of solutions of SDEs and the fact that ˛0 > l.
A Appendix Let Y be a J -valued diffusion starting from x0 with a drift and volatility that satisfy the Engelbert–Schmidt conditions. Let b W J ! R be a Borel measurable function and let .c; d / J , c < x0 < d . Recall that, for any x; y 2 J , the stopping times x and x;y are defined in (5) and (6). We now extend the definition in (6) by setting c;d WD if c D l; d D r; c;d WD c ^ if c > l; d D r; c;d WD ^ d if c D l; d < r. Proposition 1. (i) The condition b2 2 L1loc .c; d / 2 implies that for all t 2 Œ0; 1/, Z
t 0
b 2 .Yu / du < 1 P-a.s. on ft < c;d g:
58
A. Mijatovi´c et al.
(ii) For any ˛ 2 J such that Z
t
b2 2
… L1loc .˛/ we have
b 2 .Yu / du D 1 P-a.s. on f˛ < t < g:
0
Remark 8. Let us note that P.˛ < 1/ > 0 by [2, Theorem 2.11]. Clearly, f˛ < 1g D f˛ < g. Hence there exists t 2 Œ0; 1/ such that P.˛ < t < / > 0. Proof. (i) Using the occupation times formula, P-a.s. we get Z
t
Z
t
b 2 .Yu /du D
0
0
b2 .Yu /dhY iu D 2
Z J
b2 y .y/Lt .Y /dy; 2
t 2 Œ0; /:
(27)
y
P-a.s. on the set ft < c:d g the function y 7! Lt .Y / is cádlág (see [11, Chap. VI, Theorem 1.7] with a compact support in the interval .c; d /. Now the first statement follows from (27). (ii) We have Z
˛C" ˛"
b2 .y/dy D 1 for all " > 0: 2
By [2, Theorem 2.7], we have for any t 0 L˛t .Y / > 0 and L˛ t .Y / > 0 P-a.s. on f˛ < t < g: y
Then, P-a.s. on f˛ < t < g, there exists " > 0 such that the function y 7! Lt .Y / is away from zero on the interval .˛ "; ˛ C "/. It follows from (27) that R t bounded 2 b .Y / du D 1 P-a.s on f˛ < t < g. This concludes the proof. t u u 0 Acknowledgements We would like to thank the anonymous referee for a thorough reading of the paper and many useful suggestions, which significantly improved the paper.
References 1. S. Blei, H.-J. Engelbert, On exponential local martingales associated with strong Markov continuous local martingales. Stoch. Process. Their Appl. 119(9), 2859–2880 (2009) 2. A.S. Cherny, H.-J. Engelbert, Singular Stochastic Differential Equations. Lecture Notes in Mathematics, vol. 1858 (Springer, Berlin, 2005) 3. H.-J. Engelbert, T. Senf, On functionals of a Wiener process with drift and exponential local martingales, in Stochastic Processes and Related Topics (Georgenthal, 1990). Mathematical Research, vol. 61 (Akademie, Berlin, 1991), pp. 45–58 4. T. Jeulin, Semi-martingales et grossissement d’une Filtration. Lecture Notes in Mathematics, vol. 833 (Springer, Berlin, 1980) 5. I. Karatzas, S.E. Shreve, Brownian Motion and Stochastic Calculus. Graduate Texts in Mathematics, vol. 113, 2nd edn. (Springer, Berlin, 1991)
Martingale Property of Generalized Stochastic Exponentials
59
6. N. Kazamaki, On a problem of Girsanov. Tôhoku Math. J. 29(4), 597–600 (1977) 7. R.S. Liptser, A.N. Shiryaev, Statistics of Random Processes, I. Applications of Mathematics (New York), vol. 5, 2nd edn. (Springer, Berlin, 2001) 8. A. Mijatovi´c, M. Urusov, On the martingale property of certain local martingales. Probability Theory and Related Fields, 152(1):1–30 (2012) 9. A.A. Novikov, A certain identity for stochastic integrals. Theor. Probab. Appl. 17, 761–765 (1972) 10. P.E. Protter, Stochastic Integration and Differential Equations. Stochastic Modelling and Applied Probability, vol. 21, 2nd edn. (Springer, Berlin, 2005) Version 2.1 (Corrected third printing) 11. D. Revuz, M. Yor, in Continuous Martingales and Brownian Motion. Grundlehren der Mathematischen Wissenshaften, vol. 293, 3rd edn. (Springer, Berlin, 1999)
Some Classes of Proper Integrals and Generalized Ornstein–Uhlenbeck Processes Andreas Basse-O’Connor, Svend-Erik Graversen, and Jan Pedersen
Abstract We give necessary and sufficient conditions for existence of proper integrals from 0 to infinity or from minus infinity to 0 of one exponentiated Lévy process with respect to another Lévy process. The results are related to the existence of stationary generalized Ornstein–Uhlenbeck processes. Finally, in the square integrable case the Wold-Karhunen representation is given. Keywords Stochastic integration • Lévy processes • Generalized Ornstein– Uhlenbeck processes
1 Introduction Let .; / D .t ; t /t 2R denote a bivariate Lévy process indexed by R satisfying 0 D 0 D 0, that is, .; / is defined on a probability space .˝; F ; P /, has càdlàg paths and stationary independent increments. We are interested in the two integrals Z 1 Z 0 s .a/ W e ds and .b/ W e s ds : (1) 0
1
The first of these has been thoroughly studied, see e.g. [5,6,9,13,15,18], where it is Rt treated as an improper integral, i.e. as the a.s. limit as t ! 1 of 0 e s ds . Recall that Erickson and Maller, [9] Theorem 2, give necessary and sufficient conditions
A.B.-O’Connor Department of Mathematics, The University of Tennessee, 227 Ayres Hall, 1403 Circle Drive, Knoxville, TN 37996-1320, USA e-mail:
[email protected] S.-E. Graversen J. Pedersen () Department of Mathematical Sciences, University of Aarhus, Ny Munkegade, 8000 Aarhus C, Denmark e-mail:
[email protected];
[email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__3, © Springer-Verlag Berlin Heidelberg 2012
61
62
A. Basse-O’Connor et al.
in terms of the Lévy-Khintchine triplet of .; / for the existence of (1)(a) in the improper sense. In the following this integral is considered as a semimartingale integral up to infinity in the sense of e.g. [7] or [2], which we can think of as a proper integral. Theorem 1 shows that the conditions given by Erickson and Maller are also necessary and sufficient for the existence of (1)(a) in the proper sense. To the best of our knowledge the second integral has previously only been studied in special cases, in particular when is deterministic. As we shall see in the next section, is a so-called increment semimartingale in the natural filtration of .; /, that is, the least right-continuous and complete filtration to which .; / is adapted. Integration with respect to increment semimartingales has been studied by the authors in [2], and we use the results obtained there to give necessary and sufficient conditions for the existence of (1)(b); see Theorem 1. As an application, generalized Ornstein–Uhlenbeck processes (and some generalizations hereof) are considered. Recall that a càdlàg process V D .Vt /t 2R is a generalized Ornstein–Uhlenbeck process if it satisfies Z t .t s / t Vt D e Vs C e e u du for s < t: (2) s
See [3, 4, 14] for a thorough study of these processes and their generalizations and references to theory and applications. Assuming t ! 1 as t ! 1 a.s., Theorem 2 shows that a necessary and sufficient condition for the existence of a stationary V is that (1)(b) exists, and in this case V is represented as Z t t Vt D e e u du for t 2 R: (3) 1
This result complements Theorem 2.1 in [14] where the stationary distribution is expressed in terms of an integral from 0 up to infinity. Finally, assuming second moments, Theorem 3 gives the Wold-Karhunen representation of V .
2 Integration with Respect to Increment Semimartingales In this section we first recall a few general results related to integration with respect to increment semimartingales. Afterwards we specialize to integration with respect to . Let .Ft /t 2R denote a filtration satisfying the usual conditions of right-continuity and completeness. Recall from [2] that a càdlàg R-valued process Z D .Zt /t 2R is called an increment semimartingale with respect to .Ft /t 2R if for all s 2 R the process .Zt Cs Zs /t 0 is an .FsCt /t 0 -semimartingale in the usual sense. Equivalently, by Example 4.1 in [2], Z is an increment semimartingale if and only if it induces an L0 .P /-valued Radon measure on the predictable -field P. Note that in general an increment semimartingale is not adapted. Let Z D Z .!I dt dx/ denote the jump measure of Z defined as Z .A/ D #fs 2 R W .s; Zs / 2 Ag for A 2 B.R R0 /;
Some Classes of Proper Integrals and Generalized Ornstein–Uhlenbeck Processes
63
where R0 D R n f0g, and let .B; C; / denote the triplet of Z; see [2]. That is, D .!I dt dx/ is the predictable compensator of Z in the sense of [12], Theorem II.1.8. Moreover, B D B.!I dt/ is a random signed measure on R of finite total variation on compacts satisfying that t 7! B..s; s C t/ is an .FsCt /t 0 -predictable process for all s 2 R and C D C.!I dt/ is a random positive measure on R which is finite on compacts. Finally, for all s < t we have Z Zt Zs D Ztc Zsc C Z C
x ŒZ .du dx/ .du dx/ .s;t fjxj1g
x Z .du dx/ C B..s; t/ .s;t fjxj>1g
where Z c D .Ztc /t 2R is a continuous increment local martingale and for all s 2 R c the quadratic variation of .ZsCt Zsc /t 0 is C..s; sCt/, t 0. Choose a predictable random positive measure D .!I dt/ on R which is finite on compacts, a realvalued predictable process b D bt .!/, a positive predictable process c D ct .!/, and aR transition kernel K D K.t; !I dx/ from .R ˝; P/ into .R; B.R// satisfying 2 R .1 ^ x / K.tI dx/ < 1 and K.tI f0g/ D 0 for all t 2 R such that B.dt/ D bt .dt/;
C.dt/ D ct .dt/;
.dt dx/ D K.tI dx/ .dt/:
As shown in [2] a necessary and sufficient condition for the existence of is that we have the following:
R R
Z Z ˇ ˇ ˇ ˇ . s x/ s .x/ K.sI dx/ˇ .ds/ < 1; ˇ s bs C R
Z R
R
s2 cs .ds/ < 1;
Z Z R
R
1 ^ . s x/2 K.sI dx/ .ds/ < 1;
s dZs
(4) (5)
where W R ! R is a truncation function, i.e. it is bounded, measurable and satisfies .x/ D Rx in a neighborhood of 0. Moreover, when these conditions are satisfied the t process 1 s dZs , t 2 R, is a semimartingale up R to infinity. Here R we use the usual convention that for a measurable subset A of R, A s dZs WD R s 1A .s/ dZs , and R Rt s WD .s;t for s < t. Let us turn to integration with respect to where .; / is a bivariate Lévy process indexed by R with 0 D 0 D 0, that is, plays the role of Z from now on. Denote the Lévy-Khintchine triplet of 1 by . ; 2 ; m /, and let ; denote the covariance of the Gaussian components of and at time t D 1. A similar notation will be .;/ used for all other Lévy processes. Let .Ft /t 2R denote the natural filtration of .;/ .; /. Note that .t /t 0 is a Lévy process in .Ft /t 0 , i.e. for all 0 s < t, .;/ t s is independent of Fs . Using this it is easily seen that the increment .;/ semimartingale triplet of in .Ft /t 2R is, for t > 0, given by .dt/ D dt and 2 .bt ; ct ; K.tI dx// D . ; ; m .dx//. Thus, (4)–(5) provide necessary and sufficient
64
A. Basse-O’Connor et al.
R1 .;/ conditions that 0 s ds exists for an arbitrary .Ft /t 2R -predictable process . This in particular includes (1)(a). When it comes to integrals involving the negative half axis, such as (1)(b), the situation is more complicated since is not a Lévy .;/ process in .Ft /t 2R (see [2], Sect. 5). In fact, a priori it is not even clear that is an increment semimartingale in the natural filtration of .; / (or in any other filtration). However, using an enlargement of filtration result due to Protter and Jacod, see [11], Theorem 5.3 in [2] shows that this is indeed the case, and the triplet of is calculated in an enlargedRfiltration. Theorem 5.5 in [2] provides sufficient conditions for integrals of the form R s ds to exist and as this result will be used throughout the paper we rephrase it as a remark. .;/
Remark 1. Assume EŒ21 < 1. Then for Rany .Ft /t 2R -predictable process R having a.a. paths locally bounded the integral R s ds exists if R .j s j C s2 / ds < 1 a.s.
3 Main Results As above let .; / denote a bivariate Lévy process indexed by R with 0 D 0 D 0. To avoid trivialities assume that none of them are identically equal to 0. Often we will assumeR that t ! 1 a.s. as t ! 1 because if this fails, Theorem 2 in [9] t shows that 0 e s ds does not converge as t ! 1 a.s., implying that (1)(a) does not exist. We need the function A defined, cf. [9] and [14], as Z
x
A .x/ D maxf1; m ..1; 1//g C
m ..y; 1// dy;
x 1:
(6)
1
To study (1)(b) we follow Lindner and Maller [14] and introduce .Lt /t 0 given by Lt D t C
X
.e s 1/s t;
for t 0:
(7)
0<st .;/
The process .Lt ; t /t 0 is then a bivariate Lévy process in the .Ft (see [14], Proposition 2.3) and for all t 0 we have Z
0
e
s
t
D
ds D e
t
Z
t
e
s
D
Z
t
ds D
0
e s dLs :
/t 0 -filtration
(8)
0
D
(Here D denotes equality in distribution.) Indeed, the second equality follows from [14], Proposition 2.3. To prove the first equality note that Z
0
t
e s ds D e .0 t /
Z
0 t
e s t ds ;
Some Classes of Proper Integrals and Generalized Ornstein–Uhlenbeck Processes
65
which by the stationary increments has the same law as e .t 0 /
Z
t
e s 0 ds D e t
0
Z
t
e s ds : 0
The existence of (1)(a) and (1)(b) is characterized in the following. All integrals over infinite intervals are defined in the proper sense of [2], implying in particular that they exist as improper integrals. Theorem 1. Assume t ! 1 as t ! 1 a.s. (1) The following statements are equivalent: (a) RThe integral (1)(a) exists. t (b) 0 e s ds converges in distribution as t ! 1. (c) We have Z log jyj m .dy/ < 1: c A .log jyj/ Œe;e
(9)
(2) The following statements are equivalent: (a) The integral (1)(b) exists. R0 (b) t e s ds converges in distribution as t ! 1. (c) We have Z log jyj mL .dy/ < 1: c A .log jyj/ Œe;e If (1)(b) exists then
R0 1
D
e s ds D
R1 0
(10)
e s dLs .
It should be noted that (1c) coincides with the condition in [9], Theorem 2, for (1)(a) to exist as an improper integral. In the case when t ! 1 as t ! 1 a.s, (1c) implies (2c); this is shown in [14], where also further interesting relations between these conditions can be found. Proof. Throughout the proof assume t ! 1 as t ! 1 a.s. The equivalence between (1b) and (1c) is given in [14], Proposition 2.4. Using (8) it thus follows that (2b) and (2c) are equivalent. Since, as noted above, existence of integrals in the proper sense implies existence as improper integrals it is obvious that (1a) implies (1b) and (2a) implies (2b). We prove the remaining assertions in a few steps. Step 1. Assume there is an > 0 such that m .fjxj > g/ D 0. (That is, has no big jumps, implying in particular square integrability). We show that in this case (1)(a) and (1)(b) both exist. Note that if EŒj1 j < 1 then by the law of large numbers t =t ! EŒ1 as t ! 1 a.s. and in this case EŒ1 > 0. Recall that we assume t ! 1 as t ! 1 a.s. It follows from Kesten’s trichotomy theorem (see e.g. [8], Theorem 4.4) that if
66
A. Basse-O’Connor et al.
EŒj1 j D 1 then limt !1 t =t D 1 a.s. Thus, there is a 2 .0; 1 such that s =s ! as s ! 1 and t =t ! as t ! 1 a.s. In particular Z
Z
0
.e C e s
2s
1
Z
1
/ ds C
.e t C e 2t / dt
0
Z
0
D
.e
s.s =s/
Ce
2s.s =s/
1
/ ds C
1
.e t .t =t / C e 2t .t =t / / dt < 1
0
which by Remark 1 implies the existence of (1)(a) and (1)(b). Step 2. Assume is a compound Poisson process, that is, Z m .R/ < 1;
2 D 0;
D
jxj1
x m .dx/
P and t is given by t D 0<st s for all t > 0 a.s. Assume in addition that (1c) P holds. We show that in this P case s>0 e s js j < 1 a.s. (This clearly implies that (1)(a) exists and equals s>0 e s s ). For this purpose we first prove Z
1 0
Z
1 ^ .e s jxj/ m .dx/ ds < 1
R
a.s.
(11)
R1 The in (11) can be written as 0 g.s / ds, where, for y 2 R, g.y/ D R integral y jxj// m .dx/. Therefore, since g is non-increasing it follows from [9], R .1 ^ .e Theorem 1, that (11) is satisfied if and only if Z .1;1/
y jdg.y/j < 1: A .y/
(12)
Simple manipulations show that Z g.y/ D m .R/
y
Z
1
jxjez
jxj m .dx/e z dz
(13)
and hence the integral in (12) equals Z .1;1/
y A .y/
Z jxjey
jxj m .dx/e y dy:
(14)
Since A is non-decreasing we can use Fubini to rewrite and dominate this integral as
Some Classes of Proper Integrals and Generalized Ornstein–Uhlenbeck Processes
Z
Z
1
jxj log jxj
Œe;ec
ye y dy m .dx/ A .y/ Z D Œe;ec
Z Œe;ec
jxj A .log jxj/
Z
1 log jxj
67
ye y dy m .dx/
jxj e log jxj .1 C log jxj/ m .dx/; A .log jxj/
which by assumption. Now, by P [7], Lemma 3.4, (11) is equivalent to P is finite s s 1 ^ .e j j/ < 1 a.s. Hence js j < 1 a.s. s s>0 s>0 e Step Decompose .t /t 0 as t D 1t C 2t where P3. Proof that (1c) implies (1a). 2 D 0<st s 1fjs j>1g , that is, contains all jumps of magnitude larger than 1. R1 R1 By Step 1, 0 e s d1s exists and by Step 2, 0 e s d2s exists if (1c) is fulfilled. 2t
Step 4. We first prove that (2c) implies (2a). As in the proof of Step 2 we may and will assume that is a compound Poisson process. In this case .Lt /t 0 in (7) is a compound Poisson process as well. By definition of .Lt /t 0 and Step 2 we have X
e s js j D
0<s<1
X
e s jLs j < 1
a.s.
0<s<1
D
Since .t ; t /t 0 D ..t / ; .t / /t 0 it follows that X
D
e s js j D
X
e s js j:
1<s<0
0<s<1
R0 Thus, the right-hand side is finite a.s., implying that the integral 1 e s ds exists P and equals 1<s<0 e s s . if (1)(b) exists then the condition in (10) is satisfied implying by (1) that R 1Finally, s e dL s exists. From (8) it follows that 0 Z
0 1
Z e s ds D lim
0
t !1 t
D
Z
e s ds D lim
t !1 0
t
e s dLs D
Z
1
e s dLs
0
t u
where the first and third equality signs hold a.s.
Next we use the above theorem to study generalized Ornstein–Uhlenbeck processes. For this, consider a bivariate Lévy process .U; / D .Ut ; t /t 2R with U0 D 0 D 0 and assume that none of them are identically equal to 0. Assume in addition that mU .f1g/ D 0, meaning that U has no jumps of size 1. Following .U;/;ex [2], Sect. 5.2, we introduce an extended filtration .Ft /t 2R which is defined as .U;/;ex .U;/ Ft D Ft for t 0 and .U;/;ex
Ft
.U;/
D Ft
_ ..U;/ ..t; 0 A/ W A 2 B.R2 //
for t < 0;
68
A. Basse-O’Connor et al.
where .U;/ is the jump measure of .U; /. By [2], Theorem 5.3, and U are increment semimartingales in the extended filtration, ensuring a well-defined integration theory with respect to these processes. Since the index set is R rather than RC , we define the stochastic exponential of U , E .U / D .E .U /t /t 2R , as the càdlàg process satisfying E .U /0 D 1 and E .U /t E .U /s
.t s/ 2 U 2
D e .Ut Us /
Y
.1 C Uu /e Uu
for s < t:
8 <e Ut 2t U2 Q Uu 0
for t 0;
(15)
s
Put differently, E .U / is given by
for t 0:
(16)
Equation (15) shows that for s 2 R, .E .U /t Cs =E .U /s /t 0 , is the usual stochastic exponential (with index set RC ) of .Ut Cs Us /t 0 , cf. [12], II.8. To describe E .U / in a more convenient way we follow [4] and introduce two important auxiliary processes N D .Nt /t 2R and D .t /t 2R as follows. Let N0 D 0 D 0 and Nt Ns D U ..s; t .1; 1/
for s < t X 2 t s D .Ut Us / C .t s/ ŒUu log j1 C Uu j 2 U C
(17) for s < t:
s
(18) These are essentially the definitions given in [4] except that the index set is R rather b there. As noted in [4], than RC , and our corresponds to the process called U Nt Ns is the number of jumps in U of size less than 1 on the interval .s; t. Moreover, N and are both Lévy processes. .U;/;ex
Lemma 1. The process .t ; Nt /t 2R is .Ft
/t 2R -adapted.
Proof. For t 0 we let s D 0 in (17) and (18), which trivially shows that .Nt ; t / is .U;/;ex FtU -measurable and hence also Ft -measurable. For t < 0 note that (use t D 0 .U;/;ex U and s D t in (17)) Nt D ..t; 0 .1; 1/ implying that Nt is Ft measurable by definition of this -field. Moreover, by a standard argument, Z
.x/ U .du dx/
(19)
.t;0R .U;/;ex
-measurable for all measurable W R ! R for which (19) exists. In is Ft P particular t
measurable with respect to Ft
.
t u
Some Classes of Proper Integrals and Generalized Ornstein–Uhlenbeck Processes
69
The importance of N and is due to the fact that E .U / is given as E .U /t E .U /s
D .1/Nt Ns e .t s / .U;/;ex
By Lemma 1 this shows that E .U / is .Ft jumps of size less than 1, we have E .U /t E .U /s
D e .t s /
for s < t:
(20)
/t 2R -adapted. When U does not have
for s < t:
(21)
In this case [12], II.8, shows that for all s 2 R and t 0, Ut Cs Us D L og.e .Cs s / /t , where L og denotes the stochastic logarithm. Finally, we need the process L D .Lt /t 2R defined as L0 D 0 and Lt Ls D t s C .Œ; U t Œ; U s / X D t s C Uu u C .t s/U;
(22) for s < t:
(23)
s
It follows from (18) that when E .Ut / D e t for t 2 R (that is, U has no jumps of size less than 1), then Lt D Lt , t 0, where the latter is defined in (7). Clearly, L is determined by .U; /. Conversely, is determined by .U; L / since t s D Lt Ls
X Uu L u .t s/U;L 1 C U u s
for s < t:
Note that since L differs from only by a term which is of bounded variation on compacts and is an increment semimartingale in the extended filtration, so is L . .U;/ Similarly it follows that .Lt /t 0 is a semimartingale in .Ft /t 0 . In the following we consider càdlàg processes V D .Vt /t 2R satisfying E .U /s du s E .U /u Z t E .U /t D Vs C E .U /t ŒE .U /u 1 du E .U /s s
Vt D
E .U /t E .U /s
Z Vs C
t
(24) for s < t:
(25)
.U;/;ex
Remark 2. Assume V D .Vt /t 2R is càdlàg and .Ft /t 2R -adapted. Then V is given by (24) if and only if it satisfies the linear stochastic differential equation Vt D Vs C
.Lt
Ls /
Z
t
C
Vu dUu
for s < t:
(26)
s
For a proof, see [4], Proposition 3.2, or [10], Theorem VI(6.8). A detailed study of stationary solutions to (26), including the nasty case mU .f1g/ > 0, can be found in [4].
70
A. Basse-O’Connor et al.
In the case when E .Ut / D e t for t 2 R, (24) and (25) reduce to so-called generalized Ornstein–Uhlenbeck processes: Vt D e
.t s /
Z t u s Vs C e du s
D e .t s / Vs C e t
Z
(27)
t
e u du
for s < t:
(28)
s
R0 Theorem 2. (1) Assume t ! 1 as t ! 1 a.s. The integral 1 E .U /1 u du R0 R D 0 exists if 1 e u du exists. We have 1 E .U /1 D u du R 1 if and only there is a stationary càdlàg 0 E .U /u dLu in case of existence. Moreover, R0 process V D .Vt /t 2R satisfying (24) if and only if 1 E .U /1 u du exists. In this case V D .Vt /t 2R is uniquely determined as Z Vt D E .U /t
t
1
E .U /1 u du ;
t 2 R:
(29)
R1 R1 u d (2) Assume t ! 1 as t ! 1. The integrals 0 E .U /1 u u du and 0 e exist at the same time. There is a stationary càdlàg process V D .V t /t 2R R1 satisfying (24) if and only if 0 E .U /1 u du exists. In this case V D .Vt /t 2R is uniquely determined as Z
1
Vt D E .U /t t
E .U /1 u du ;
t 2 R:
(30)
R1 Recall that necessary and sufficient conditions that the integrals 0 e u du R0 R0 u and du exist are given in Theorem 1. In particular, 1 e u du and 1 e R 1 u R01 e dLu exist at the same time. This is also equivalent to the existence of u dL ; indeed, it is easily verified that jL j D jL j for all t > 0, and t u t 0 e thus the condition in Theorem 1(1c) with D L is equivalent to the one with L . Moreover, of the first part of (1) below it follows that R 1D R 1as in the proof u dL and e E .U / dL exist at the same time. u u u 0 0 The above conditions for existence of V are also given in [4], Theorem 2.1, and so is the representation (30); thus, (2) is completely contained in [4] (except that we use proper integrals) but it is restated here for completeness. When t D t for some non-zero constant , (27) and (28) simplify to a usual Ornstein–Uhlenbeck process; see e.g. [1] and [16]. In this case, (29) and (30) are well known representations of stationary Ornstein–Uhlenbeck processes cf. e.g. [16], Theorem 55. The case when t does not converge to ˙1 as t ! 1 is treated in Theorem 2.1 of [4]. u only differ by a factor of absolute value 1 the Proof. (1) Since E .U /1 u and e two proper integrals exist at the same time. The identity in distribution follows as in
Some Classes of Proper Integrals and Generalized Ornstein–Uhlenbeck Processes
71
the proof of the last assertion of Theorem 1(2), where instead of (8) we use that for t 0, Z
0 t
D
E .U /1 u du D E .U /t
Z
t 0
D
E .U /1 u du D
Z
t 0
E .U /u dLu ;
(31)
where the first equality follows as in the proof of (8) and the second comes from Lemma 3.1 in [4]. Assume V is stationary and satisfies (24). Letting s ! 1 and using that R0 s ! 1 a.s. it follows from (25) and (20) that s E .U /1 u du converges in Rt distribution. From (31) it follows that 0 E .U /u dLu converges in distribution as t ! 1. Theorem 3.6 in [4] shows that the R 0 condition in Theorem 1(2c) (with L replaced by L ) is satisfied, implying that 1 E .U /1 u du exists. R0 1 Conversely, assuming that 1 E .U /u du exists and defining V by (29) it is easily seen that (24) is satisfied. Moreover, since Vt is given a.s. as Z Vt D lim E .U /t h!1
t t h
E .U /t h!1 E .U /t h
E .U /1 u du D lim
Z
t t h
E .U /t h du E .U /u
(32)
and the distribution of the right-hand side does not depend on t, the distribution of Vt is also independent of t. The variable Vt is moreover determined by .t u ; Ut Uu /ut and is hence in particular independent of .uCt t ; UuCt Ut /u0 . From (24) it follows that V is stationary. The proof of (2) is similar except that in (24) and (25) we fix s, rearrange terms to isolate Vs , and let t ! 1. t u In the next theorem we study integrability properties and the so-called WoldKarhunen representation of generalized Ornstein–Uhlenbeck processes. As above we consider the bivariate Lévy process .U; / as well as and L defined in (18) and (22). From now on we assume mU ..1; 1/ D 0, that is, U has no jumps of size 1 or smaller. We will thus not need the process N . Whenever U is integrable let WD EŒU1 and define U D .U t /t 2R as U 0 D 0 and U t U s D Ut Us C .t s/ for s < t. Theorem 3. Assume mU ..1; 1/ D 0. (1) For r > 0 we have U1 2 Lr .P / if and only if EŒe r1 < 1. When U1 2 L1 .P / we have Z h i 1 1 2 exp.EŒU1 / D EŒe D exp C 2 C .e x 1 C x1fjxj1g / m .dx/ : R
(33) (2) Assume L1R 2 L2 .P / and EŒe 21 < 1. Then t ! 1 as t ! 1 a.s and 1 the integral 0 e t dLt exists and is in L2 .P /. Moreover, is strictly positive, the generalized Ornstein–Uhlenbeck process satisfying (27) exists and is square integrable, and
72
A. Basse-O’Connor et al.
Z Vt D
t
e
.t s/
1
dLs
Z
t
C 1
e .t s/ Vs dU s ;
t 2 R:
(34)
Remark 3. Assume L1 2 L2 .P /, EŒe 21 < 1 and EŒL1 D 0. By (34), Z Vt D
t 1
e .t s/ ds ;
t 2 R;
(35)
Rt where, for s < t, t s D Lt Ls s Vu dU u and 0 D 0. The process is square integrable with zero mean and stationary orthogonal increments, implying that (35) is the Wold-Karhunen representation of V . To verify this fix s 2 R. It was noted in the proof of Theorem 2 that Vs is independent of .LsCt Ls ; UsCt Us /t 0 . Using that the distribution of Vs as well as of .LsCt Ls ; Ut Cs Us /t 0 does not depend on s, it follows from (27) that has stationary increments. Since L as well as U are zero mean square integrable Lévy processes and V is square integrable and stationary it follows by definition of that .t Cs s /t 0 is a square integrable martingale in the filtration generated by Vs and .LsCt Ls ; Ut Cs Us /t 0 . In particular this implies that has zero mean and orthogonal increments. For results related to some of the integrability properties above see Proposition 3.1 and Theorem 3.3 in [3]. Proof. (1) Since Ut D L og.t / for t > 0, [12], Corollary II.8.16, shows that Z
Z u mU .du/ D
Z
r
u>1
.e erx 1>1
rx
1/ m .dx/ D
log 2 x> r
.e rx 1/ m .dx/:
R Thus, the left-hand side is finite if and only if jxj>1 e rx m .dx/ is finite. On the other hand, since mU ..1; 1/ D 0, [17], Corollary 25.8, shows that finiteness of the left-hand side is equivalent to U1 2 Lr .P /, and similarly finiteness R rx of jxj>1 e m .dx/ is equivalent to EŒe r1 < 1. Now assume EŒe 1 < 1. The second equality in (33) follows from [17], Theorem 25.17. Moreover, since Ut D L og.t / for t > 0, [12], Theorem II.8.10, implies that Ut D t C 2t 2 C Œ.e x 1 C x/ t ;
t 0:
Recalling the Lévy-Itô decomposition of : t D x1fjxj1g . /t C Œx1fjxj>1g t C t C Gt ;
t 0;
where G denotes the (mean zero) Gaussian component of and .dt dx/ D m .dx/ dt, it follows that Ut D x1fjxj1g . /t t Gt C 2t 2 CŒ.e x 1Cx1fjxj1g / t ;
t 0:
Some Classes of Proper Integrals and Generalized Ornstein–Uhlenbeck Processes
73
All terms on the right-hand side have finite mean, and the first and third term have mean zero, implying Z EŒU1 D C 12 2 C .e x 1 C x1fjxj1g / m .dx/: R
In particular this gives (33). p (2) Since EŒe 1 EŒe 21 < 1 it follows from (33) that > 0. It is shown in [14], Proposition 4.1, that t ! 1 a.s. Recall ([17], Theorem 25.17) that EŒe 2t D .EŒe 21 /t for t 0. Using that L1 is square integrable we can decompose .Lt /t 0 as Lt D Mt C c1 t where c1 is a constant and .Mt /t 0 is a square integrable martingale with hM it D c2 t for some c2 0. Since i Z 1 hZ 1 E e 2t dhM it D .EŒe 21 /t d.c2 t/ < 1 R1
0
0
the integral 0 e t dMt exists and is square integrable. Moreover, since s is independent of u s for all 0 s < u and EŒe 1 ; EŒe 21 < 1, we get h Z
1
E
e s ds
Z
2 i
1
Z Z
s
1
Z
s
1
EŒe s u du ds
D2
0
0
Z
1
EŒe 2s .u s / du ds
D2 0
Z
1
1
D2 0
.EŒe 21 /s .EŒe 1 /us du ds < 1
s
R1
R1 and hence the integrals 0 e s d.c1 s/ and 0 e s dLs exist and are square integrable. By Theorem 2 and the remarks following it the generalized Ornstein– Uhlenbeck process V D .Vt /t 2R exists and is square integrable. From (26) we have Z t Z t Vt D Vs C Œ.Lt Ls / C Vu dU u Vu du for s < t: s
s
Thus, by [10], Theoreme VI(6.8), Vt D e
.t s/
Z
t
Vs C
e s
.t u/
dLu
Z
t
C
e .t u/ Vu dU u
for s < t:
(36)
s
Since V is a stationary square integrable process, is positive and U and L are square integrable, it follows from Remark 1 that .Vu e u 1fut g/u2R is integrable with respect to U and .e u 1fut g /u2R is integrable with respect to L . Letting s ! 1 in (36) it follows that Z t Z t Vt D e .t u/ dLu C e .t u/ Vu dU u for t 2 R: 1
1
t u
74
A. Basse-O’Connor et al.
References 1. O.E. Barndorff-Nielsen, N. Shephard, Non-Gaussian Ornstein-Uhlenbeck-based models and some of their uses in financial economics. J. R. Stat. Soc. Ser. B Stat. Methodol. 63(2), 167–241 (2001) 2. A. Basse-O’Connor, S.-E. Graversen, J. Pedersen, A unified approach to stochastic integration on the real line. Thiele Research report 2010-08 http://www.imf.au.dk/publication/publid/880, 2010 3. A. Behme, Distributional properties of solutions of d Vt D Vt d Ut C dLt with Lévy Noise. Adv. Appl. Prob. 43, 688–711 (2011) 4. A. Behme, A. Lindner, R.A. Maller, Stationary solutions of the stochastic differential equation d Vt D Vt d Ut C dLt with Lévy Noise. Stochast. Process. Appl. 121(1), 91–108 (2011) 5. J. Bertoin, A. Lindner, R. Maller, On continuity properties of the law of integrals of Lévy processes. In Séminaire de Probabilités XLI. Lecture Notes in Mathematics, vol. 1934 (Springer, Berlin, 2008), pp. 137–159 6. P. Carmona, F. Petit, M. Yor, Exponential functionals of Lévy processes. In Lévy Processes (Birkhäuser Boston, Boston, MA, 2001), pp. 41–55 7. A. Cherny A. Shiryaev, On stochastic integrals up to infinity and predictable criteria for integrability. In Séminaire de Probabilités XXXVIII. Lecture Notes in Mathematics, vol. 1857 (Springer, Berlin, 2005), pp. 165–185 8. R.A. Doney, R.A. Maller, Stability and attraction to normality for Lévy processes at zero and at infinity. J. Theor. Probab. 15(3), 751–792 (2002) 9. K.B. Erickson, R.A. Maller, Generalised Ornstein-Uhlenbeck processes and the convergence of Lévy integrals. In Séminaire de Probabilités XXXVIII. Lecture Notes in Mathematics, vol. 1857 (Springer, Berlin, 2005), pp. 70–94 10. J. Jacod, Calcul Stochastique et Problèmes de Martingales. Lecture Notes in Mathematics, vol. 714 (Springer, Berlin, 1979) 11. J. Jacod, P. Protter, Time reversal on Lévy processes. Ann. Probab. 16(2), 620–641 (1988) 12. J. Jacod, A.N. Shiryaev, Limit Theorems for Stochastic Processes, vol. 288 of Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences], 2nd edn. (Springer, Berlin (2003) 13. H. Kondo, M. Maejima, K. Sato, Some properties of exponential integrals of Lévy processes and examples. Electron. Commun. Probab. 11, 291–303 (electronic) (2006) 14. A. Lindner, R. Maller, Lévy integrals and the stationarity of generalised Ornstein-Uhlenbeck processes. Stochast. Process. Appl. 115(10), 1701–1722 (2005) 15. A. Lindner, K. Sato, Continuity properties and infinite divisibility of stationary distributions of some generalized Ornstein-Uhlenbeck processes. Ann. Probab. 37(1), 250–274 (2009) 16. A. Rocha-Arteaga, K. Sato, Topics in Infinitely Divisible Distributions and Lévy Processes, vol. 17 of Aportaciones Matemáticas: Investigación [Mathematical Contributions: Research] (Sociedad Matemática Mexicana, México, 2003) 17. K. Sato, Lévy Processes and Infinitely Divisible Distributions, vol. 68 of Cambridge Studies in Advanced Mathematics (Cambridge University Press, Cambridge, 1999). Translated from the 1990 Japanese original, Revised by the author 18. M. Yor, Exponential Functionals of Brownian Motion and Related Processes (Springer Finance. Springer, Berlin, 2001). With an introductory chapter by Hélyette Geman, Chapters 1, 3, 4, 8 translated from the French by Stephen S. Wilson
Martingale Representations for Diffusion Processes and Backward Stochastic Differential Equations Zhongmin Qian and Jiangang Ying
Abstract In this paper we explain that the natural filtration of a continuous Hunt process is continuous, and show that martingales over such a filtration are continuous. We further establish a martingale representation theorem for a class of continuous Hunt processes under certain technical conditions. In particular we establish the martingale representation theorem for the martingale parts of (reflecting) symmetric diffusion in a bounded domain with a continuous boundary. Together with an approach put forward in (Liang et al., Ann. Probab.), our martingale representation theorem is then applied to the study of initial and boundary problems for quasi-linear parabolic equations by using solutions to backward stochastic differential equations over the filtered probability space determined by reflecting diffusions in a bounded domain with only continuous boundary. Keywords Backward SDE • Dirichlet form • Hunt process • Martingale • Natural filtration • Non-linear equations
AMS Classification: 60H10, 60H30, 60J45
Z. Qian () Mathematical Institute, University of Oxford, Oxford OX1 3LB, England e-mail:
[email protected] J. Ying Institute of Mathematics, Fudan University, Shanghai, China e-mail:
[email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__4, © Springer-Verlag Berlin Heidelberg 2012
75
76
Z. Qian and J. Ying
1 Introduction Let B D .Bt /t 0 be the Brownian motion in Rd started with an initial distribution . The natural filtration .Ft /t 0 (called the Brownian filtration, for a definition, see [3]) is continuous. More importantly the Brownian motion has the martingale representation property: any martingale over .˝; F ; Ft ; P / can be expressed as an Itô integral against Brownian motion. In particular all martingales on the filtered probability space .˝; F ; Ft ; P / are continuous. The (predictable) martingale representation property of a family of martingales has been studied in a more extended setting, and several general results have been obtained. For example, Jacod and Yor [16] have discovered the equivalence between the martingale representation property and the extremal property of martingale measures. Jacod [15] also obtained the martingale representation property in terms of the uniqueness of some martingale problems, and further present criteria in terms of predictable characteristics. These results have greatly illuminated the subject matter. When applied to specific situations, further work and indeed hard estimates are often required. As a matter of fact, we still have very limited examples of martingales and filtrations which possess martingale representation property (see, e.g. [1, 15, 26, 33]). The renewed interest in recent years in the martingale presentation property has been motivated not only by its own right, but also by its important applications in the mathematical finance [11, 31], backward stochastic differential equations and their applications in some non-linear partial differential equations, see also for example [1, 5, 17, 23, 27, 32] and the reference therein. An intimate question is the continuity of natural filtrations generated by semimartingales and diffusion processes. A great knowledge about them has been obtained in the past. For example, a complete characterization of the natural filtrations generated by simple jump processes and Lévy processes in terms of their sample paths is known. Much information has been obtained for a class of Markov processes. We know, from the fundamental work by Blumenthal, Chung, Dynkin, Getoor, Hunt, Meyer etc., (see for example [4, 8, 25], in particular Hunt [14]), that the natural filtration of a Feller process is right continuous and quasi-left continuous. On the other hand, to the best knowledge of the present authors, there are no general conditions in literature to guarantee the martingales over the natural filtration of a Markov process to be continuous. A reasonable conjecture is that the natural filtration of a diffusion process (a continuous strong Markov process) should be continuous, so are the martingales over the natural filtration. Such a result is plausible but remains to prove. In this paper, we show that all martingales over the natural filtration of a continuous Hunt process are continuous. Our proof follows a key idea originated from Blumenthal [3], formulated carefully in Meyer [25]. The main result of the present article is a martingale representation theorem for a class of continuous Hunt processes which satisfy a technical condition called the Fukushima representation property (FRP). The term FRP is so named, because
Martingale Representations for Diffusion Processes
77
for symmetric diffusions on a finite dimensional space, under a few regularity conditions, due to Fukushima’s chain rule for energy measures, the FRP will be satisfied automatically. We are therefore able to provide a large class of diffusion processes which have the martingale representation property. As a consequence of our main result, we establish the martingale representation theorem for symmetric diffusion processes on a domain, with Dirichlet or Neumann boundary condition. More precisely, let D Rd be a bounded domain with a continuous boundary @D. Consider the symmetric diffusion in D with a formal infinitesimal generator LD
d 1 X @ ij @ a .x/ i j 2 i;j D 1 @x @x
in D
subject to the Dirichlet or Neumann boundary condition (for precise meaning, see Sects. 4 and 5 below), where .aij / are only Borel measurable and satisfies the uniform elliptic condition. To make it clear, let X D .˝; F ; Ft ; Xt ; t ; P x /: be the symmetric diffusion associated with the Dirichlet form .E ; F / where E .u; v/ D
1 2
Z
d X D i;j D 1
aij
@u @v dx @x j @x i
and the Dirichlet space F D H01 .D/ or H 1 .D/ depending on the Dirichlet or Neumann boundary condition. Here we use the same letter F to denote the filtration as well as the Dirichlet space: we hope it should be clear from the context which one F stands for. F , Ft are the natural filtrations generated by the symmetric diffusion .Xt /t 0 . It happens in this case that the coordinate functions uj .x/ D x j belong to the local Dirichlet space Floc , and Xt D .Xt1 ; ; Xtd / has the Fukushima’s decomposition j
j
j
j
Xt X0 D Mt C At
P x -a.e. j D 1; ; d
for all x 2 D (or D in the Neumann boundary condition case) except for a zero capacity set with respect to the Dirichlet form .E ; F /, where M 1 , , M d are continuous martingales additive functionals and A1 , , Ad are continuous additive functionals with zero energy. The following martingale representation theorem follows from our main result. Theorem 1. Under the above assumptions, for any initial distribution which has no charge on capacity zero sets, the family .M 1 , , M d / of martin gales over .˝; F ; Ft ; P / has the martingale representation property: for any square integrable martingale N D .Nt /t 0 over .˝; F ; Ft ; P / there are unique
78
Z. Qian and J. Ying
.Ft /-predictable processes F 1 , , F d such that Nt N0 D
d Z X j D1 0
t
Fsj dMsj .
We must point out a subtle difference for the martingale representation property used in this article to that used in the semimartingale theory, such as Jacod and Yor [16]. Here the filtration F is generated by the diffusion process, not necessarily coincide with the filtration generated by the family of martingale parts M j , and therefore the integrands F j are predictable processes with respect to the diffusion process. Theorem 1 may be applied to the symmetric diffusions in Rd with Dirichlet form .E ; F / where F D H01 .Rd /. This special case has been proved in [33] and [1]. In [33], Zheng has pointed out that the martingale part of symmetric diffusion .Xt /t 0 in Rd with infinitesimal generator being a uniform second order elliptic operator in divergence form has the martingale predictable representation property and described a proof based on the results on the Dirichlet process p.t; Xt / obtained in Lyons and Zheng [22] and [21], where p.t; x/ is the probability density function of Xt under the stationary distribution. More precisely, Lyons and Zheng [21] have extended Fukushima’s representation theorem for martingale additive functionals to a class of processes which has a form f .t; Xt /, where f has finite spacetime energy. Their results in particular yield that p.t; Xt / is a Dirichlet process in the sense of Föllmer [12], and its martingale part can be expressed as an Itô integral against .M 1 ; ; M d /, which, together with the Markov property, allows to show that for D f1 .Xt1 / fn .Xtn / the conditional expectation E.jFt / is again a Dirichlet process which can be expressed as an Itô’s integral against .M 1 ; ;RM d /, where the expectation is taken against the stationary distribution P m ./ D Rd P x ./dx. A routine procedure based on the Doob’s maximal inequality allows to prove the martingale representation theorem for the symmetric diffusion in Rd with the generator L an elliptic operator of second order. Apparently not knowing the work [33], in an independent work Bally, Pardoux and Stoica [1], among other things, a detailed proof has been provided. The technical difficulty with the proof described above lies in the fact that even for a smooth function f , f .Xt / may not be a semimartingale, so that Itô’s calculus can not be applied. Instead of considering random variables with product form such as D f1 .Xt1 / fn .Xtn / which linearly span a vector space dense in Lp .˝; F ; P /, we utilize a linear vector C spanned by those random variables which have a product form D 1 n with Z
1
j D
e ˛j t fj .Xt /dt
0
where fj are bounded Borel measurable functions, and ˛j > 0 for j D 1; ; n. According to Meyer [25], C is dense in Lp .˝; F ; P /. The important feature is
Martingale Representations for Diffusion Processes
79
that U ˛ f .Xt / is a semimartingale for any ˛ > 0 and a bounded Borel function f , where U ˛ is the resolvent of the transition semigroup. Moreover, in the symmetric case, U ˛ f always belongs to the Dirichlet space (when f is bounded and square integrable) and thus Fukushima’s representation theorem for martingale additive functionals can be applied to extend the representation to any martingales. The martingale representation theorem for the symmetric diffusion process .Xt / in a domain D allows us to study the following type of backward stochastic differential equation (BSDE) d Yt D f .t; Yt ; Zt /dt C
d X
j
j
Zt dMt
j D1
with a terminal condition that YT D 2 L2 .˝; FT ; P /, and thus gives a probability representation for weak solutions to the initial and boundary value problem for non-linear parabolic equations. The existence and uniqueness of solutions to the BSDE follows from exactly the same approach as for the Brownian motion case, which is the pioneering work in BSDE done by Pardoux and Peng in [27]. We however describe an approach put forward in [20], which allows us to devise an alternative probability representation for the initial and boundary problem of the corresponding semi-linear parabolic equation d @ 1 X @ ij @ u a u C f .t; u; ru/ D 0 @t 2 i;j D1 @x j @x i
ˇ @ ˇ u.t; / D 0 in a bounded domain with only continuous boundary. subject to @ @D For related topics of BSDEs associated with general elliptic operators, we mention in particular the papers Lejay [19], Bally, Pardoux and Stoica [1], Rozkosz [30] and [28] and the references therein. The paper is organized as follows. In Sect. 2, we develop further an idea put forward in Meyer [25], and show both the natural filtrations and martingales over the natural filtrations for a continuous Hunt process are continuous. This result is hardly new but it seems not appear in the literature yet. In order to prove this result, we devise an important while elementary formula for E .jFt /, which may be considered as a refined version of a classical formula devised firstly by Hunt and Blumenthal for potentials and multiple potential case by Meyer. In Sect. 3, we establish the main result of the paper: a martingale representation theorem for a continuous Hunt process under technical assumptions called the Fukushima representation property, and give some examples in which our result may apply. In Sect. 4, we outline the existence and uniqueness of solutions to backward stochastic differential equations over the natural filtered probability space over a reflecting symmetric diffusion in a bounded domain with non-smooth diffusion coefficients and non-smooth boundary, and finally we apply the theory of BSDE to the study of the initial and (Neumann) boundary problem of a non-linear parabolic equation in a
80
Z. Qian and J. Ying
bounded domain with only continuous boundary. We believe these results are new even for reflecting Brownian motion in a domain with non-smooth boundary. The first version of this paper was posted on ArXiv a year ago. Subsequently many colleagues have kindly informed the authors that it is known to experts that the subspace of potential martingales is dense in the martingale space over a Hunt process, a result goes back to Kunita and Watanabe [18]. A proof for symmetric diffusions may be found in Fukushima et al. [13]. By using this fact, we can prove Theorem 2 without using the multiple potential formula (6) which is the main technical step in our proof. The first author would like to thank Professor Le Jan and Professor Emery for the discussions on this topic and the literature about filtrations and martingale representation property.
2 Martingales over the Filtrations of Continuous Hunt Processes Consider a Markov process .˝; F 0 ; Ft0 ; Xt ; t ; P x / in a state space E 0 D E [ f@g, where E is a locally compact separable metric space E, with transition probability function fPt .x; / W t 0g, i.e., Z E x ff .Xt Cs /jFs0 g D
f .z/Pt .Xs ; d z/ .
(1)
E
In (1), E x on the left-hand side stands for the (conditional) expectation with respect to the probability measure P x , and the right-hand side may be abbreviated as Pt f .Xs / where Z Pt f .x/ D
f .z/Pt .x; d z/ E
which is well defined for a bounded or non-negative Borel measurable function f . The family of kernels .Pt /t >0 is called the transition semigroup associated with the Markov process .Xt /t 0 . Without specification, .Ft0 /t 0 is the filtration generated by .Xt /t 0 , that is Ft0 D fXs W s tg for t 0 and F 0 D fXs W s 0g. For a -finite measure on .E; B.E// (where B.M / always represents the Borel -algebra on a topological space M ) Z P ./ D
P x ./.dx/; 2 F 0 ;
E
defines a measure on .˝; F 0 /. If is a probability, then .Xt /t 0 is Markovian under P with transition semigroup .Pt /t >0 and initial distribution , in the sense that Z E ff .Xt Cs /jFs0 g D
f .z/Pt .Xs ; d z/ P -a.e. E
Martingale Representations for Diffusion Processes
81
and P fX0 2 Ag D .A/ for any A 2 B.E/, where E is the (conditional) expectation against the probability P . Denote by P.E/ the space of all probability measures on .E; B.E//. If 2 P.E/, F denotes the completion of F 0 under P , and Ft is the smallest 0 -algebra containing Ft and all sets in F with zero probability. .Ft /t 0 is called the natural filtration of the Markov process .Xt /t 0 with initial distribution . Let Ft D
\
Ft
2P.E/
and .Ft /t 0 is called the natural filtration determined by the Markov process X D .˝; F 0 ; Ft0 ; Xt ; t ; P x /: If .Gt /t 0 is a filtration, i.e. an increasing family of -algebras on a common sample space, then Gt C D \s>t Gs for t 0 and Gt D fGs W s < tg for t > 0. The filtration is called right (resp. left) continuous if Gt C D Gt for all t 0 (resp. Gt D Gt for all t > 0). The sample function properties of a Markov process .Xt /t 0 and the continuity properties of its natural filtration had been studied by Blumenthal, Dynkin, Getoor, Hunt, Meyer etc. The fundamental results have been established via the regularity of the transition probability function fPt .x; / W t > 0g. Their work achieved the climax for Markov processes with Feller transition semigroups. As matter of fact, the continuity of the filtration .Ft / (or .Ft /) does not follow that of sample function .Xt /t 0 . For example, a right continuous Markov process does not necessarily lead to the right continuity of its natural filtration .Ft / (or .Ft /). The same claim applies to the left continuity. In fact, the regularity of natural filtrations has much to do with the nature of the Markov property, such as strong Markov property. Let C1 .E/ (resp. C0 .E/) denote the space of all continuous functions f on E which vanish at infinity @, i.e. limx!@ f .x/ D 0 (resp. with compact support). Recall that a transition semigroup .Pt /t >0 on .E; B.E// is Feller, if for each t > 0, Pt preserves C1 .E/ and limt #0 Pt f .x/ D f .x/ for each x 2 E and f 2 C1 .E/. For a given Feller semigroup .Pt /t >0 on .E; B.E//, there is a Markov process X D .˝; F 0 ; Ft0 ; Xt ; t ; P x / with the Feller transition semigroup .Pt /t >0 such that the sample function t ! Xt is right continuous on Œ0; 1/ with left hand limits on .0; 1/. In this case, we call .˝; F 0 ; Ft0 ; Xt ; t ; P x / a Feller process on E. For a Feller process .˝; F 0 ; Ft0 ; Xt ; t ; P x /, the natural filtration .Ft /t 0 for any 2 P.E/ and as well as .Ft /t 0 are right continuous. .Xt /t 0 and .Ft /t 0 are also quasi-left continuous, that is, if Tn is an increasing family of .Ft /-stopping times, and Tn " T , then limn!1 XTn D XT on fT < 1g and FT D fFTn W n 2 Ng. Therefore accessible .Ft /-stopping times are predictable.
82
Z. Qian and J. Ying
An .Ft /-stopping time T is totally inaccessible if and only if P fT < 1g > 0 and XT ¤ XT on fT < 1g P -a.e. Similarly, T is accessible if and only if XT D XT P -a.e. on fT < 1g. Hence X has only inaccessible jump times. What we are mainly concerned in this article is Hunt processes. Hunt processes are right continuous, strong Markov processes which are quasi-left continuous. These processes are defined in terms of sample functions, rather than transition semigroups, see [4] and [8] for details. It is well-known that Feller processes are stereotype of Hunt processes or the latter is an abstraction of the former. We are interested in the martingales over the filtered probability space .˝; F , Ft ; P x /, and we are going to show that, if .˝; F 0 ; Ft0 ; Xt ; t ; P x / is a Hunt process which has continuous sample function, then any martingale on this filtered probability space is continuous, a result one could expect for the natural filtration of a diffusion process. Indeed, this result was proved more or less by Meyer in his Lecture Notes in Mathematics 26, “Processus de Markov”. Meyer himself credited his proof to Blumenthal and Getoor, more precisely a calculation done by Blumenthal [3]. However it is surprising that the full computation, which yields more information about martingales over the natural filtration of a Hunt process, was not reproduced either in the new edition of Meyer’s “Probabilités et Potentiels” or Chung’s “Lectures from Markov Processes to Brownian Motion”, although it was mentioned in [9] where Chung and Walsh gave an alternative proof of Meyer’s predictability result, so that Blumenthal’s computation is no longer needed. However, it is fortunate that Blumenthal’s calculation indeed leads to a proof of a martingale representation theorem we are going to establish for certain Hunt processes, see Sect. 3. Let us first describe an elementary calculation, originally according to Meyer [25] due to Blumenthal. Let X D .˝; F 0 ; Ft0 ; Xt ; t ; P x / be a Hunt process in a state space E 0 D E [ f@g with the transition semigroup .Pt /t 0 , where @ plays a role of cemetery. Let fU ˛ W ˛ > 0g be the resolvent of the transition semigroup .Pt /t 0 : Z
1
U .x; A/ D ˛
e ˛t Pt .x; A/dt
0
and .U ˛ /˛>0 the corresponding resolvent (operators), i.e. Z U ˛ f .x/ D
f .z/U ˛ .x; d z/ E Z 1
D 0
e ˛t Pt f .x/dt
Martingale Representations for Diffusion Processes
83
for bounded or nonnegative Borel measurable function f on E. To save words, we use Bb .E/ to denote the algebra of all bounded Borel measurable functions on E. Obviously, C1 .E/ Bb .E/. Let K.E/ Bb .E/ be a vector space which generates the Borel -algebra B.E/. Let C L1 .˝; F ; P / (for any initial distribution ) be the vector space spanned by all D 1 n for some n 2 N, Z
1
e ˛j t fj .Xt /dt
j D 0
where ˛j are positive numbers, fj 2 K.E/, j D 1; ; n. Meyer [25] proved that C is dense in L1 .˝; F ; P / for a Hunt process. Since this density result will play a crucial role in what follows, we include Meyer’s a proof for completeness and for the convenience of the reader. The key observation in the proof is the following result from real analysis. Lemma 1. Let T > 0. Let K denote the vector space spanned by all functions e˛ .t/ D e ˛t , where ˛ > 0, then K is dense in C Œ0; T equipped with the uniform norm. The lemma follows from Stone–Weierstrass’ theorem. Lemma 2 (P. A. Meyer). For any initial distribution and p 2 Œ1; 1/, C is dense in Lp .˝; F ; P /. Proof. First, by utilizing Doob’s martingale convergence theorem, it is easy to show that the collection A of all random variables which have the following form g1 .Xt1 / gn .Xtn /, where n 2 N, 0 < t1 < < tn < 1 and gj 2 K.E/, is dense in Lp .˝; F ; P /. Let H be the linear space spanned by all D 1 n , where Z
1
j D
gj .Xt /'j .t/dt, 0
where gj 2 K.E/ and 'j 2 C Œ0; 1/ with compact supports. According to the previous lemma, for every " > 0 we may choose j 2 K such that j'j .t/
j .t/j
Z
for some > 0. Let
< "e t for all t 0
1
j D
gj .Xt /
j .t/ dt.
0
Then Q D 1 n 2 C , and j j j j
1 jjgj jj1 "
84
Z. Qian and J. Ying
where jj jj1 is the supermum norm. It follows that p Q p n max jjgj jjnp "p Ej j 1
p j
and thus belongs to the closure of C . Finally it is clear that any element g1 .Xt1 / gn .Xtn / D lim k1 kn k!1
Z
where kj D
1 0
gj .Xt /'jk .t/dt
and 'jk has compact support and 'jk ! ıtj weakly. We thus have completed the proof. u t Let be any fixed initial distribution. If f is a bounded Borel measurable function on E and ˛ > 0, then Z
1
D
e ˛t f .Xt /dt 2 L1 .˝; F ; P /.
0
˚ Consider the martingale Mt D E jFt where t 0. According to an elementary computation in the theory of Markov processes, Z
1
Mt D E 0
Z
t
D Z
e ˛s f .Xs /ds C E
t 0
Z
e ˛s f .Xs /ds C e ˛t
Z
1
e ˛s f .Xs /dsjFt
t 1
e ˛s Ps f .Xt /ds
0 t
D
0
D Z
e ˛s f .Xs /dsjFt
e ˛s f .Xs /ds C e ˛t U ˛ f .Xt /.
0
It is known that if X D .Xt /t 0 is a Hunt process,then for any ˛ > 0 and bounded Borel measurable function f , U ˛ f is finely continuous, i.e., t ! U ˛ f .Xt / is right continuous. Moreover if X is a continuous Hunt process, it follows from a result proved by Meyer that t ! U ˛ f .Xt / is continuous, and therefore, the martingale Mt D E fjFt g is continuous. We record Meyer’s result as a lemma here. This result was proved in [25] for Hunt processes (see T15 THEOREME, page 89, [25]). A simpler proof for Feller processes may be found on page 168, [10]. Lemma 3 (Meyer). Let .Xt /t 0 be a Hunt process, f 2 Bb .E/, ˛ > 0 and h D U ˛ f be a potential. Then
Martingale Representations for Diffusion Processes
h.Xt / D h.X /t
85
8t > 0 P -a.e.
for any initial distribution . P. A. Meyer pointed out that the previous computation can be carried out equally for randomR variables on .˝; F 0 / which have a product form D 1 n where 1 each j D 0 e ˛j s fj .Xs /ds. Let n denote the permutation group of f1; ; ng. R1 Lemma 4 (Blumenthal and Meyer). Let D ˚1 n where j D 0 e ˛j s fj .Xs /ds, ˛j > 0 and fj 2 Bb .E/, and Mt D E jFt . Then Mt D
n X
k Z Y
X
t
! e ˛ji s fji .Xs /ds F.j1 ; ;jk ; ;jn / .Xt /
(2)
i D1 0
kD0 .j1 ; ;jk ; ;jn /2n
where ( F.j1 ; ;jk ; ;jn / .x/ D E
n Y
x
e
˛jl t
Z
!)
1
e
˛jl s
fjl .Xs /ds
.
(3)
0
lDkC1
˚ Proof. The task is to calculate the conditional expectation Mt D E jFt . The idea is very simple: splitting each j into Z
t
j D
e ˛j s fj .Xs /ds C e ˛j t
Z
0
1
e ˛j s fj .Xs ı t /ds
0
so that D
n X
k Z Y
X
e
˛jl t
Z
1
e
e
˛ji s
i D1 0
kD0 .j1 ; ;jk ; ;jn /2f1; ;ng n Y
!
t
˛jl s
fji .Xs /ds !
fjl .Xs ı t /ds .
0
lDkC1
By using the Markov property one thus obtains ˚ Mt D E jFt D
n X
k Z Y
X
kD1 .j1 ; ;jk ; ;jn /2f1; ;ng
(
E
n Y lDkC1
e
˛jl t
Z
e
˛ji s
fji .Xs /ds
i D1 0
1
e 0
!
t
˛jl s
fjl .Xs ı t /ds
!
) jFt
86
Z. Qian and J. Ying
D
n X
k Z Y
X
kD0 .j1 ; ;jk ; ;jn /2f1; ;ng
(
n Y
E Xt
Z
e ˛jl t
1
t
! e ˛ji s fji .Xs /ds
i D1 0
!)
e ˛jl s fjl .Xs /ds
.
(4)
0
lDkC1
t u Our only contribution in this aspect is the following formula, which allows to prove not only that all martingales over the natural filtration of a continuous Hunt process are continuous, but also a martingale representation theorem in the next section. Lemma 5. Let ˛j be positive numbers and fj 2 Bb .E/ for j D 1; ; k. Consider Z F .x/ D
Z
e
Z
Pk
j D1 ˛j sj
E ˝k
0<s1 <<sk <1
f1 .z1 / fk .zk /Ps1 .x; d z1 /
Ps2 s1 .z1 ; d z2 / Psk sk1 .zk1 ; d zk /ds1 dsk . Then
F D U ˛1 CC˛k f1 .U ˛2 CC˛k f2 .U ˛k fk / .
(5)
Proof.R To see why it is true, we begin with the case that k D 1. In this case 1 F D 0 e ˛s Ps f ds D U ˛ f . If k D 2, then “
e ˛2 s2 e ˛1 s1 Ps1 .f1 Ps2 s1 f2 / ds1 ds2
F D 0<s1 <s2 <1
Z
1
Z
D Z
0
e ˛1 t Pt
Z
0
1
e
˛1 t ˛2 t
e
Z Pt f1
0
1
e
˛2 s
Ps f2 ds dt
0 1
D
e ˛2 s f1 Pst f2 ds dt
t
1
D Z
e ˛1 t e ˛2 s Pt .f1 Pst f2 / dsdt
t 1
D Z
1
e .˛1 C˛2 /t Pt .f1 U ˛2 f2 / dt
0
D U ˛1 C˛2 .f1 U ˛2 f2 / and by an induction argument, for a general case. Indeed, if k > 2, then
Martingale Representations for Diffusion Processes
Z
Z
F .x/ D
e
j D1
˛j sj ˛kC1 skC1
e
0<s1 <<skC1 <1
Z
Pk
87
f1 .z1 / fk .zk /PskC1 sk fkC1 .zk /
E ˝k
Ps1 .x; d z1 /Ps2 s1 .z1 ; d z2 / Psk sk1 .; zk1 ; d zk /ds1 dsk dskC1 Z Z Z Pk D e j D1 ˛j sj f1 .z1 / E ˝k
0<s1 <<sk <1
Z
1
fk .zk /
e
˛kC1 skC1
sk
PskC1 sk fkC1 .zk /dskC1
Ps1 .x; d z1 /Ps2 s1 .z1 ; d z2 / Psk sk1 .; zk1 ; d zk /ds1 dsk Z Z Z P kj D1 ˛j sj D e f1 .z1 / E ˝k
0<s1 <<sk <1
fk .zk /e ˛kC1 sk
Z
1
e ˛kC1 t Pt fkC1 .zk /dt
0
Ps1 .x; d z1 /Ps2 s1 .z1 ; d z2 / Psk sk1 .; zk1 ; d zk /ds1 dsk Z Z Pk1 D e j D1 ˛j sj .˛kC1 C˛k /sk 0<s1 <<sk <1
Z
E ˝k
f1 .z1 / fk .zk /U ˛kC1 fkC1 .zk /dt t u
and the formula follows the induction assumption. Lemma 6. Let f1 ; ; fk 2 Bb .E/, ˛j positive numbers, and 80 19 k Z 1 = < Y F .x/ D E x @ e ˛j s fj .Xs /ds A : ; : 0 j D1
Then X
F D
U ˛j1 CC˛jk fj1 .U ˛j2 CC˛jk fj2 .U ˛jk fjk /
fj1 ; ;jk g2k
where k is the permutation group of f1; ; kg. Proof. We have Z 1 0
e ˛1 s1 e ˛k sk f1 .Xs1 / fk .Xsk /ds1 dsk
0
Z
1
D
1
0
Z
Z
1
F .x/ D E x
0
e ˛1 s1 e ˛k sk E x ff1 .Xs1 / fk .Xsk /g ds1 dsk
(6)
88
Z. Qian and J. Ying
D
X fj1 ; ;jk g2k
Z
Z
e ˛1 s1 e ˛k sk
0<sj1 <<sjk <1
E x ff1 .Xs1 / fk .Xsk /g ds1 dsk Z Z X D e ˛j1 s1 e ˛jk sk fj1 ; ;jk g2k
0<s1 <<sk <1
fj1 ; ;jk g2k
0<s1 <<sk <1
˚ E x fj1 .Xs1 / fjk .Xsk / ds1 dsk Z Z Z X D e ˛j1 s1 e ˛jk sk fj1 .z1 / fjk .zk / E ˝k
Ps1 .x; d z1 /Ps2 s1 .z1 ; d z2 / Psk sk1 .zk1 ; d zk /ds1 dsk t u
and (6) follows from Lemma 5. From now on, we assume that X D .˝; F 0 ; Ft0 ; Xt ; t ; P x /
is a continuous Hunt process in E 0 D E [ f@g with the transition semigroup .Pt /t 0 . In other words, it is a Hunt process with continuous sample paths. Therefore, .Xt /t 0 is a diffusion process in E, i.e. .Xt /t 0 possesses the strong Markov property with continuous sample function. Under our assumptions, any finite .Ft / stopping time is accessible and thus predictable, and therefore FT D FT . In particular, .Ft / is left continuous, and thus the filtration .Ft / is continuous for any initial distribution . Since any martingale on .˝; F ; Ft ; P / has a right continuous modification, by a martingale we always mean a martingale with right continuous sample function. Lemma 7. Suppose D 1 n where each j has the following form Z 1 j D e ˛j s fj .Xs /ds 0
˚ where ˛j > 0 and fj 2 Bb .E/. Let Mt D E jFt . Then .Mt /t 0 is a bounded continuous martingale on .˝; F ; Ft ; P /. Proof. According to Lemma 4, we need only to show that for function of the following type ( !) Z 1 n Y x ˛jl t ˛jl s F .x/ D E e e fjl .Xs /ds ; lDkC1
0
t 7! F .Xt / is continuous. By Lemma 6, F is an ˛-potential, so that it is finely continuous, and together with Lemma 3, it implies that t ! F .Xt / is continuous, which completes the proof. t u
Martingale Representations for Diffusion Processes
89
We now state the main result of this section. For simplicity, a square integrable martingale .Mt /t 0 over .˝; M ; Mt ; P / means Mt D E.jMt / with 2 L2 .˝; M ; P /. This is equivalent to say supt >0 EŒMt2 < 1. Theorem 2. Let X D .˝; F 0 ; Ft0 ; Xt ; t ; P x / be a continuous Hunt ˚process in E, and 2 P.E/. If 2 L2 .˝; F ; P /, then the martingale Mt D E jFt is continuous, that is, square-integrable martingales on .˝; F ; Ft ; P / are continuous. Therefore local martingales over .˝; F , Ft ; P / are continuous. Proof. We can choose a sequence n 2 C such that n ! in L2 . Doob’s maximal inequality implies that, if necessary by considering a subsequence, the martingales fE .n jFt / W t 0g converges (almost surely at least along a subsequence) to fE .jFt / W t 0g uniformly on any finite interval of t 0. It is shown in Lemma 7 that for each n, the martingale E .n jFt / is continuous and thus the square integrable martingale fE .jFt / W t 0g must be continuous. By the localization technique, it follows thus that local martingales on .˝; F , Ft ; P / are continuous. t u
3 Martingale Representation for Continuous Hunt Process In this section we assume that X D .˝; F 0 ; Ft0 ; Xt ; t ; P x / is a continuous Hunt process in the state space E 0 D E [ f@g with transition semigroup fPt .x; dy/ W t 0g, where E is a locally compact separable metric space. Let 2 P.E/ be an initial distribution. If ˛ > 0 and f 2 Bb .E/ then M ˛;f denotes the continuous martingale Z 1 ˛;f Mt D E e ˛s f .Xs /dsjFt . 0
Recall that, if u is an ˛-potential, i.e., u D U ˛ f where f 2 Bb .E/, then u.Xt / u.X0 / is a continuous semimartingale on .F ; Ft ; P /. and possesses Doob– Meyer’s decomposition Œu
Œu
u.Xt / u.X0 / D Mt C At Z
where Œu
Mt
t
D 0
and Lu D ˛u f .
Œu
e ˛s dMs˛;f , At D
Z
t
Lu.Xs /ds 0
90
Z. Qian and J. Ying
We make the following assumptions on a continuous Hunt process X with an initial distribution 2 P.E/. Assumptions. There is an algebra (a vector space which is closed under the multiplication of functions) K.E/ Bb .E/ which generates the Borel -algebra B.E/ and is invariant under U ˛ for ˛ > 0, and there are finite many continuous martingales M 1 ; ; M d over .˝; F ; Ft ; P / such that the following conditions are satisfied: (1) For any potential u D U ˛ f where ˛ > 0 and f 2 K.E/, the martingale part M Œu of the semimartingale u.Xt / u.X0 / has the martingale representation in terms of .M 1 ; ; M d /, that is, there are predictable processes F1 , , Fd on .˝; F ; Ft / such that Œu
Mt
D
d Z X
t
Fsj dMsj P -a.e.
j D1 0
(2)
(7)
j hM ; M i it is strictly positive definite.
The first assumption means that the martingale M Œu with u being a potential may be represented. The second condition ensures that the representation (7) is unique. In this case we say that Fukushima representation property holds for X with initial law and martingales f.M .1/; ; M .d / /g. The Fukushima representation property is mainly an abstraction of the chain rule for the martingale part of u.Xt /. Indeed, if Xt D .Xt1 ; ; Xtd / is a d -dimensional Brownian motion and u is an ˛-potential with ˛ > 0, then u is smooth and by Itô’s formula u.Xt / u.X0 / D
d Z X
t
j D1 0
so that Œu
Mt
D
@u .Xs /dXsj C @x j
d Z X
t
j D1 0
Z
t 0
1 u.Xs /ds 2
@u .Xs /dXsj . @x j
One can easily see that the Brownian motion satisfies the Fukushima representation property. Theorem 3 (Martingale representation). Let 2 P.E/. If Fukushima representation property holds for X with initial law and a finite set of martingales .M 1 ; ; M d /, then for any square-integrable martingale N D .Nt /t 0 on .˝; F ; Ft ; P /, there are unique predictable processes .Fti / such that Nt N0 D
d Z X i D1
t 0
Fsi dMsi
P -a.e.
Martingale Representations for Diffusion Processes
91
Proof. The uniqueness follows from condition 2) in the Fukushima representation. We prove the existence. Take 2 L2 .˝; F ; P / such that Nt D E fjFt g. Since 2 C is dense in L .˝; F ; P /, so we first prove the martingale representation for 2 CR. By the linearity, we only need to consider the case that D 1 n where 1 j D 0 e ˛j s fj .Xs /ds for ˛j > 0 and fj 2 K.E/. In this case, according to 4, Lemmas 5 and 6 X Nt D E fjFt g D Ztm m
where the sum is a finite one, and for each m, Z m D Zt has the following form Ztm D Vtm um .Xt / (the superscript m will be dropped if no confusion may arise), where k Z Y 0
Vt D
t
e ˇi s gi .Xs /ds
i D1 0
and Z u.x/ D
Z
e
Pk
j D1
Z ˇj sj E ˝k
0<s1 <<sk <1
h1 .z1 / hk .zk /Ps1 .x; d z1 /
Ps2 s1 .z1 ; d z2 / Psk sk1 .zk1 ; d zk /ds1 dsk for some k 0 and k, ˇi > 0 and functions gi , hj are bounded and continuous. According to Lemma 5 u D U ˇ1 CCˇk h1 .U ˇ2 CCˇk h2 .U ˇk hk / . In particular, u is again a potential which has a form u D U ˛ g for g D h1 .U ˇ2 CCˇk h2 .U ˇk hk / / 2 K.E/ and ˛ D ˇ1 C C ˇk . Hence u.Xt / is a continuous semimartingale with decomposition Œu Œu u.Xt / u.X0 / D Mt C At where AŒu is continuous with finite variation, and due to the Fukushima representation property d Z t X Œu Gsj dMsj Mt D j D1 0
j
for some predictable processes G . In particular, each Z m is a continuous semimartingale. Since, by Theorem 2, N is a continuous martingale, so that Nt D
X m
the continuous martingale part of Vtm um .Xt /.
92
Z. Qian and J. Ying
Therefore we are interested in the martingale part of Zt D Vt u.Xt /. Since V is a finite variation process, so according to Itô’s formula Z t Z t Zt D Z0 C u.Xs /dVs C Vs d u.Xs / Z
0
0
Z
t
D Z0 C
t
u.Xs /dVs C 0
Z
Œu
Vs dMt 0
t
u.Xs /dVs C
Œu
Vs dAt C
0
t
C
0
Z
t
D Z0 C
Z Œu Vs dAt
0
d Z t X i D1
0
Vs Gsi dMsi
so that the martingale part of Zt is d Z X i D1
t
Therefore
Vs Gsi dMsi .
0
Nt D E fjFt g D
d Z tX X i D1
0
Vsm Gsm;i dMsi
m
which shows the martingale representation. Suppose now 2 L2 .˝; F ; P /. Choose a sequence n 2 C such that n ! .n/ in L2 .˝; F ; P /. Let Nt D E .n jFt / and Nt D E .jFt /. According to Doob’s maximal inequality, if necessary by passing to a subsequence, we can .n/ .n/ assume that Nt converges to Nt uniformly on any finite interval. Nt has the martingale representation .n/ Nt
.n/ N0
D
d Z X j D1 0
t
F .n/js dMsj
so that .n/
hNt
.m/
.n/
.m/
Nt ; Nt Nt i d Z t X D .F .n/is F .m/is /.F .n/js F .m/js /d hM i ; M j is . i;j D1 0
Since .hM i ; M j it / is positive, it follows that .F .n/1 ; ; F .n/d / converges to predictable processes .F 1 ; ; F d / under the norm jj.F 1 ; ; F d /jj D
Z N
1 d X 1 X i j i j E F F d hM ; M i s . s s 2N i;j D1 0 N D1
(8)
Martingale Representations for Diffusion Processes
Then Nt N0 D
d Z X j D1 0
93
t
Fsj dMsj . t u
This theorem claims that as long as every martingale of resolvent type is representable, so is any martingale. When is the Fukushima representation property satisfied? There are many examples. In the remain of this section, we shall give three interesting examples in symmetric situation. Brownian motion with any initial distribution is certainly an example. Indeed, for 1 Brownian motion in Rd , we may choose K.E/ D C1 .Rd / (the space of smooth functions which vanish at infinity), then for f 2 K.E/, U ˛ f is smooth, and (7) follows from Itô’s formula applying to U ˛ f . Theorem 3 gives a new proof for classical martingale representation theorem. The second example is the reflecting Brownian motion. As Example 1.6.1 in [13], we consider Dirichlet form . 12 D; H 1 .D// on L2 .D/ where D is the classical Dirichlet integral and D is a bounded domain on Rd . We further assume that any x 2 @D has a neighborhood U such that D \ U D f.xi / 2 Rd W xd > F .x1 ; ; xd 1 /g \ U for some continuous function F . Then C01 .D/ (the space of restriction to D of functions in C01 .Rd /) is dense in H 1 .D/ (see [24] for details), i.e., . 12 D; H 1 .D// is a regular Dirichlet form on L2 .D/. The corresponding continuous Hunt process X D .Xt ; P x / is called the reflecting Brownian motion. For x D .x i / 2 Rd , we use ui .x/ D x i , 1 i d , to denote the coordinate functions. Then ui 2 F and we denote by M i D M Œui the martingale part in Fukushima’s decomposition. It can be seen from Corollary 5.6.2 [13] that for any u 2 C01 .D/, Œu
Mt
D
d Z X i D1
t
@u .Xs /dMsi ; P x -a.s. for q.e. x 2 D; @xi
0
where q.e. means “quasi-everywhere”, i.e., except a set of zero-capacity. Then a routine approximation procedure shows that for any u 2 H 1 .D/, there exist Borel measurable functions ffi W 1 i d g on D such that Œu
Mt
D
d Z X i D1
0
t
fi .Xs /dMsi , P x -a.s. for q.e. x 2 Rd :
Therefore the reflecting Brownian motion has Fukushima representation property, by choosing K.D/ to be the space of bounded measurable functions and any initial distribution charging no set of zero capacity, i.e., a smooth distribution, because an exceptional set exists in above representation as is always when the process is
94
Z. Qian and J. Ying
constructed through a Dirichlet form. In particular, if the boundary is Lipschitz, then the transition function has density [2] and in this case, the exceptional set may be erased and Fukushima representation property holds for X with any initial law or starting at each point. Notice that under the current condition, the reflecting Brownian motion X itself is not necessarily a semimartingale. The readers who are interested may refer to [2, 6, 7, 29] about when a reflecting BM is a semimartingale and the corresponding Skorohod decomposition. It should be pointed out that, although the martingale part of the reflected Brownian motion is a Brownian motion, but the martingale representation property does not follow from the classical representation property for Brownian motion. The reason is that, as long as the boundary is not sufficiently smooth, the natural filtration .Ft /t 0 is much bigger in general than the natural filtration generated by the martingale part .M 1 ; ; M d / of X . Another example our main result may apply is symmetric diffusions in a domain killed at boundary. Actually Theorem 6.2.2 in [13] tells us that every continuous symmetric Hunt process with a smooth core enjoys the Fukushima representation property. More precisely let D be a domain of Rd with continuous boundary @D and m a Radon measure on D. Let X be a continuous Hunt process which is symmetric with respect to m and .E ; F / the associated Dirichlet form on L2 .D; m/, which has C01 .D/ as a core. For x D .x i / 2 Rd , we use ui .x/ D x i , 1 i d , to denote the coordinate functions. Then ui 2 Floc and we denote by M i D M Œui the martingale part in Fukushima’s decomposition. Let i;j D hM i ;M j i ; 1 i; j d; the smooth measure associated with CAF hM i ; M j i. Then E is expressed as d Z X
E .u; v/ D
i;j D1 D
@u @u di;j .x/; u; v 2 C01 .D/: @x i @x j
As asserted in Theorem 6.2.2 [13], for any initial smooth distribution (i.e. a probability on .D; B.D// having no charge on capacity zero sets) and u 2 F , the martingale part M Œu in Fukushima’s decomposition of u may be represented as Œu
Mt
D
d Z X i D1
t 0
fi .Xs /dMsi
P -a.e.
where f1 ; ; fd 2 B.D/. If we take K.E/ D L2 .E; m/ \ Bb .D/, X satisfies the Fukushima representation property. In these examples, fM i g are the martingales corresponding to coordinate functions so we call them coordinate martingales. To have the uniqueness, some kind of non-degenerateness is needed. We say that X is non-degenerate if the condition (2) in Fukushima representation property is satisfied: .hM i ; M j i/1i;j d is positive.
Martingale Representations for Diffusion Processes
95
Corollary 1. Assume that X is either the reflecting Brownian motion on a bounded domain or a non-degenerate symmetric Hunt diffusion on a domain D Rd as stated above. Then the Fukushima representation property is satisfied and therefore the martingale representation holds in the sense of Theorem 3 with coordinate martingales and for a given initial distribution charging no sets of zero capacity. From this result, we may recover the martingale representation established in [1] and [33], where X is a diffusion process corresponding to non-degenerate symmetric elliptic operator on Rd . Without essential difference, the conclusion holds also for reflecting diffusions on such domain with generator being a symmetric uniformly elliptic differential operator of second order as introduced in the beginning of next section.
4 Backward Stochastic Differential Equations In this section we consider backward stochastic differential equations which can be used to provide probability representations for weak solutions of the initial and boundary value problem of a quasi-linear parabolic equation. Let D Rd be a bounded domain with a continuous boundary @D, D D D [@D the closure of D. Let d 1 X @ ij @ LD a 2 i;j D1 @x j @x i be an elliptic differential operator of second order, where a D .aij / is a positivedefinite, symmetric, matrix-valued function on D, a D .aij / is Borel measurable, and satisfies the elliptic condition:
jj2
d X
aij .x/i j 1 jj2
8 D .i / 2 Rd
i;j D1
for all x 2 D for some constant > 0. Consider the Dirichlet form .E ; F / on L2 .D; dx/, where Z X d @u @v 1 aij j i dx (9) E .u; v/ D 2 D i;j D1 @x @x and F D H 1 .D/. Let ˝ be a space of all continuous paths in D, .Xt /t 0 the coordinate process on ˝, F 0 D fXs W s 0g, Ft0 D fXs W s tg for each t 0, and .t /t 0 shift operators on ˝. Let X D .˝; F 0 ; Ft0 ; Xt ; t ; P x /
96
Z. Qian and J. Ying
be the canonical realization of the symmetric diffusion process in the state space D associated with the Dirichlet space .E ; F /, which is called a reflecting symmetric diffusion in D. The coordinate functions uj .x/ D x j (j D 1; ; d ) belong to the local Dirichlet space Floc , so that j
j
j
j
Xt X0 D Mt C At
P x -a.e. j D 1; ; d
(10)
for all x 2 D except for a capacity zero set, where M j D M Œfj etc. Let S1 .D/ denote the space of all probability 2 P.D/ which has no charge on zero capacity sets (with respect to the Dirichlet form .E ; H 1 .D// defined by (9). According to Theorem 3, for any initial distribution 2 S1 .D/, the family of martingales fM j W j D 1; ; d g over .˝; F ; Ft ; P / has the martingale representation property: for any square-integrable martingale N D .Nt /t 0 on .˝; F ; Ft ; P /, there are unique predictable processes .Fti / such that Nt N0 D
d Z X i D1
t 0
Fsi dMsi
P -a.e.
Let us work with a fixed smooth initial distribution 2 S1 .D/ and the filtered probability space .˝; F ; Ft ; P /. Consider the following backward stochastic differential equation d Yti D f i .t; Yt ; Zt /dt C
d X
ij
j
Zt dMt , YTi D i
(11)
i;j D1
i D 1; ; d 0 , where T > 0, i 2 L2 .˝; FT ; P / are given terminal values, and f i are Lipschitz functions: there is a constant C1 0
jf i .t; y; z/j C1 .1 C t C jyj C jzj/ and Q zQ/j C1 .jy yj Q C jz zQj/ jf i .t; y; z/ f i .t; y; 0
0
for all t 0, y; yQ 2 Rd , z; zQ 2 Rd d . One seeks for a solution pair .Y; Z/ which solves the following integral equation Z Yti
T
D
f .s; Ys ; Zs /ds
i
i
t
d Z X j D1 t
T
Zsij dMsj
(12)
for t 2 Œ0; T . The integral equation (12) has a unique solution pair .Y; Z/ such that Y i is a continuous semimartingale, and Z ij are predictable processes satisfying Z
T
E 0
d X k;lD1
akl .Xs /Zsi l Zski ds < 1.
Martingale Representations for Diffusion Processes
97
This can be demonstrated by employing the Picard iteration for .Y; Z/ as in the case of Brownian motion (see [27]). Another approach, proposed in a paper by Lyons et al. [20] which applies to a general filtered probability space, may be described as follows. The idea is to rewrite the integral equation (12) into a functional differential equation for the variation process part V of Y . Let Y D N V where V is a finite variation process, and
Nti
D
N0i
d Z X
t
j D1 0
Zsij dMsj .
On the other hand
Nt D E f C VT jFt g. Since Y is a continuous semimartingale, its decomposition is unique up to an initial value. The integral equation (12) leads to that Z
T
Vt D
f i .s; Ys ; Zs /ds C NT t
conditioned on Ft and we obtain Z T ˚ i f .s; Ys ; Zs /dsjFt C Nt E jFt Vt D E Z D E
T
f Z
D E
t i
.s; Ys ; Zs /dsjFt
˚ C E VT jFt
t T
f .s; Ys ; Zs /ds i
VT jFt
Z
t
C
0
f i .s; Ys ; Zs /ds: 0
Therefore the integral equation (12) is equivalent to Z
t
Vt V0 D
f i .s; Ys ; Zs /ds
(13)
0
where
˚ Yt D Y .V /t D N.V /t Vt , N.V /t D E C VT jFt
and Zt D Z.V /t is determined by the martingale representation theorem N.V /it
N.V /i0
D
d Z X j D1 0
t
Z.V /ijs dMsj .
Equation (13) thus may be written as a functional equation Z
t
Vt V0 D
f i .s; Y .V /s ; Z.V /s /ds 0
(14)
98
Z. Qian and J. Ying
where Y .V / and Z.V / are considered as functionals of V . The Picard iteration applies to (14) we have
Theorem 4. If 2 L2 .˝; FT ; P / and f i are Lipschitz continuous, then there is a unique pair .Y; Z/ such that Y is a continuous semimartingale which solves BSDE (11). For a complete proof of Theorem 4, the reader may refer to [20].
5 Non-linear Parabolic Equations We are under the same setting as in the previous section, and use the notations established therein. To motivate our approach, let us begin with the case that a is smooth, and D is bounded domain with a smooth boundary. j In this case X j D .Xt /t 0 in (10), Sect. 3, are continuous semimartingales, j thus A are finite variation processes. Forˇ any h 2 Cb1;2 .Œ0; 1/ D/ satisfying @ ˇ the Neumann boundary condition that @h @ @D D 0, where @ denotes the normal derivative with respect to the Riemann metric .aij / D .aij /1 , we have h.t; Xt / h.0; X0 / D Mth C Aht Z t
where Mth
D h.t; Xt / h.0; X0 / 0
@ C L h.s; Xs /ds @s
(15)
is a martingale under P x , and Z t Aht
D 0
@ C L h.s; Xs /ds. @s
(16)
On the other hand, applying Itô’s formula 0
Z
t
h.t; Xt / h.0; X0 / D 0
C
it thus follows that Mth
D
1 d 2 X 1 @ @ @ C A h.s; Xs /ds aij @s 2 i;j D1 @x i @x j
Z tX d @ h.s; Xs /d.Msj C Ajs / j @x 0 j D1
d Z X j D1 0
t
@ h.s; Xs /dMsj @x j
(17)
Martingale Representations for Diffusion Processes
and Ait
1 D 2
99
Z tX d @ ij a .Xs /ds. j 0 j D1 @x
(18)
Consider a solution u.x; t/ to the initial boundary problem to the non-linear parabolic equation 8 @ ˆ ˆ L u C f .t; u; ru/ D 0, ˆ ˆ @t < u.0; x/ Dˇ '.x/; ˆ ˆ ˆ @u.t; / ˇˇ ˆ : D 0; @ ˇ@D
x 2 Rd ;
(19)
t >0
Then, by (15) and (17) Z
T
h.T; XT / D h.t; Xt / C
t
C
d Z X
T
j D1 t
@ C L h.s; Xs /ds @s
@ h.s; Xs /dMsj @x j
together with the PDE (19) we deduce that Z
T
h.t; Xt / h.T; XT / D
f .T s; h.s; Xs /; rh.s; Xs //ds
t
d Z X
T
j D1 t
@ h.s; Xs /dMsj . @x j
(20)
j
Let Yt D u.T t; Xt / and Zt D @x@j h.t; Xt /. Then the previous equation may be written as Z
T
Yt YT D
f .T s; Ys ; Zs /ds
t
d Z X j D1 t
T
Zsj dMsj
(21)
and YT D u.0; XT / D '.XT /. That is to say that Yt D u.T t; Xt / solves the scalar BSDE d X j j d Yt D fT .t; Yt ; Zt /dt C Zt dMt , YT D '.XT / (22) j D1
where fT D f .T t; y; z/.
100
Z. Qian and J. Ying
˚ For any fixed T > 0, let Y T D YtT W t 2 Œ0; T be the unique solution to the BSDE (22) on .˝; F ; Ft ; P /. Since the solution to BSDE is unique, T Yt D u.T t; Xt /. In particular u.T; X0 / D Y0T , and therefore Z Rd
u.T; x/.dx/ D E Y0T .
(23)
The above argument leading to the probabilistic representation (23) can not be justified in the case that a D .aij / is only Borel measurable or the boundary @D is only continuous, as in this case, .Xt /t 0 is no longer a semimartingale, both (16) and (18) no longer make sense. While, in this case, boundary problem (strong or weak solutions) to the non-linear PDE (19) also need to be interpreted. On the other hand, the BSDE (22), which relies on only the martingale representation, still make sense, thus the representation theorems stated in Sect. 3 can be made as the definition of a solution to (19). This is the approach we will carry out. Consider the initial value problem of the following non-linear parabolic equation in a bounded domain D with a continuous boundary @D 0
1 d X @ 1 @ @ @ aij .x/ i A u C f .t; u; ru/ D 0 @t 2 i;j D1 @x j @x
(24)
subject to the initial and boundary conditions u.x; 0/ D '.x/;
ˇ ˇ @ u.t; /ˇˇ D 0 for t > 0 @ @D
where a D .aij / is Borel measurable, satisfying the uniform ellipticity condition:
d X
j i j2
d X
i D1
i j aij .x/ 1
i;j
d X
j i j2
8. i / 2 Rd ;
i D1
for some constant > 0. Definition 1. The functional on S1 .D/ defined by ! E fY .t; /0 g, denoted by u.t; /, is called the stochastic solution of the initial and boundary problem of (24), where for each t > 0 and 2 S1 .D/, Y .t; / D .Ys /st is the unique solution to the BSDE ( P j j d Ys D f .t s; Ys ; Zs /ds C dj D1 Zs dMs , (25) Yt D '.Xt /
on .˝; F ; Ft ; Xt ; t ; P /.
Martingale Representations for Diffusion Processes
101
As a consequence we have Theorem 5. If ' is bounded and Borel measurable on D, and f is Lipschitz continuous, then there is a unique stochastic solution to the non-linear parabolic equation (24) We will study the regularity theory of the stochastic solutions in a separate paper. On the other hand we would like to derive an alternative probability representation of the stochastic solution. Let us apply the approach outlined in [20]. Let Ys D Ns Vs where Ns N0 D
d Z X j D1 0
s
Zrj dMrj .
Then V D .Vs /s 2 Œ0;t is the unique solution to the functional differential equation Z s f .t r; Y .V /r ; Z.V /r /dr, V0 D 0 (26) Vs D 0
where N.V /s D E f'.Xt / C Vt jFs g, Y .V /s D E f'.Xt / C Vt jFs g Vs for s 2 Œ0; t, and Z.V / is given as the density process of N.V / in the martingale representation. In particular
Y .V /0 D E f'.Xt / C Vt jF0 g. We therefore have the following Theorem 6. Let ' be bounded and measurable. For t > 0, let V .t/ be the unique solution to the functional differential equation (26). Then the stochastic solution to the Neumann boundary problem of the non-linear PDE (24) is given by u.t; / D E f'.Xt / C V .t/t g
8 2 S1 .D/.
(27)
Acknowledgements The research of the first author was supported in part by EPSRC grant EP/F029578/1, and by the Oxford-Man Institute. The second author’s research was supported in part by the National Basic Research Program of China (973 Program) under grant No. 2007CB814904, and a Royal Society Visiting grant.
References 1. V. Bally, E. Pardoux, L. Stoica, Backward stochastic differential equations associated to a symmetric Markov process. Potential Anal. 22(1), 17–60 (2005) 2. R.F. Bass, P. Hsu, Some potential theory for reflecting Brownian motion in Hölder and Lipschitz domains. Ann. Probab. 19(2), 486–508 (1991)
102
Z. Qian and J. Ying
3. R.M. Blumenthal, An extended Markov property. Trans. Am. Math. Soc. 85, 52–72 (1957) 4. R.M. Blumenthal, R.K. Getoor, Markov Processes and Potential Theory. Pure and Applied Mathematics, vol 29 (Academic Press, New York, 1968) 5. P. Briand, Y. Hu, BSDE with quadratic growth and unbounded terminal value. Probab. Theory Relat. Fields 136(4), 604–618 (2006) 6. Z.Q. Chen, Pseudo Jordan domains and reflecting Brownian motions. Probab. Theory Relat. Fields 94(2), 271–280 (1992) 7. Z.Q. Chen, P.J. Fitzsimmons, R.J. Williams, Reflecting Brownian motions: quasimartingales and strong Caccioppoli sets. Potential Anal. 2(3), 219–243 (1993) 8. K.L. Chung, Lectures from Markov processes to Brownian motion, Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Science], vol. 249 (Springer, Berlin, Heidelberg, 1982) 9. K.L. Chung, J.B. Walsh, Meyer’s theorem on predictability, Z. Wahrscheinlichkeitstheorie und Verw. Gebiete 29, 253–256 (1974) 10. C. Dellacherie, P.A. Meyer, Probabilités et potentiel, Hermann, Paris, 1975, Chapitres I à IV, Édition entièrement refondue, Publications de l’Institut de Mathématique de l’Université de Strasbourg, No. XV, Actualités Scientifiques et Industrielles, No. 1372. 11. N. El Karoui, S. Peng, M.C. Quenez, Backward stochastic differential equations in finance. Math. Finance 7(1), 1–71 (1997) 12. H. Föllmer, Dirichlet Processes, Stochastic integrals (Proc. Sympos., Univ. Durham, Durham, 1980), Lecture Notes in Mathematics, vol. 851 (Springer, Berlin, Heidelberg, 1981), pp. 476– 478 ¯ 13. M. Fukushima, Y. Oshima, M. Takeda, Dirichlet Forms and Symmetric Markov Processes, de Gruyter Studies in Mathematics, vol. 19 (Walter de Gruyter & Co., Berlin, 1994) 14. G.A. Hunt, Markoff processes and potentials. I, II. Illinois J. Math. 1, 44–93, 316–369 (1957) 15. J. Jacod, in Calcul Stochastique et Problèmes de Martingales. Lecture Notes in Mathematics, vol. 714 (Springer, Berlin, Heidelberg, 1979) 16. J. Jacod and M. Yor, Étude des solutions extrémales et représentation intégrale des solutions pour certains problèmes de martingales. Z. Wahrscheinlichkeitstheorie und Verw. Gebiete 38(2), 83–125 (1977) 17. M. Kobylanski, Backward stochastic differential equations and partial differential equations with quadratic growth. Ann. Probab. 28(2), 558–602 (2000) 18. H. Kunita, S. Watanabe, On square integrable martingales. Nagoya Math. J. 30, 209–245 (1967) 19. A. Lejay, BSDE driven by Dirichlet process and semi-linear parabolic PDE. Application to homogenization. Stochastic Process. Appl. 97(1), 1–39 (2002) 20. G.C. Liang, T. Lyons, Z.M. Qian, Backward stochastic dynamics on a filtered probability space. Ann. Probab. 39(4), 1422–1448 (2011) 21. T.J. Lyons, W.A. Zheng, Diffusion processes with nonsmooth diffusion coefficients and their density functions. Proc. R. Soc. Edinburgh Sect. A 115(3–4), 231–242 (1990) 22. T.J. Lyons, W.A. Zheng, A crossing estimate for the canonical process on a Dirichlet space and a tightness result. Astérisque, 157–158, 249–271 (1988), Colloque Paul Lévy sur les Processus Stochastiques (Palaiseau, 1987) 23. J. Ma, P. Protter, J.M. Yong, Solving forward-backward stochastic differential equations explicitly—a four step scheme. Probab. Theory Relat. Fields 98(3), 339–359 (1994) 24. V.G. Maz’ja, in Sobolev spaces, Springer Series in Soviet Mathematics (Springer, Berlin, Heidelberg, 1985), Translated from the Russian by T. O. Shaposhnikova 25. P.A. Meyer, Processus de Markov. Lecture Notes in Mathematics, No. 26 (Springer, Berlin, Heidelberg, 1967) 26. D. Nualart, W. Schoutens, Chaotic and predictable representations for Lévy processes. Stochast. Process. Appl. 90(1), 109–122 (2000) 27. É. Pardoux, S.G. Peng, Adapted solution of a backward stochastic differential equation. Syst. Control Lett. 14(1), 55–61 (1990) 28. A. Rozkosz, Backward SDEs and Cauchy problem for semilinear equations in divergence form. Probab. Theory Relat. Fields 125(3), 393–407 (2003a)
Martingale Representations for Diffusion Processes
103
29. A. Rozkosz, On a decomposition of symmetric diffusions with reflecting boundary conditions. Stochast. Process. Appl. 103(1), 101–122 (2003b) 30. A. Rozkosz, BSDEs with random terminal time and semilinear elliptic PDEs in divergence form. Studia Math. 170(1), 1–21 (2005) 31. J.M. Yong, X.Y. Zhou, Stochastic controls. Applications of Mathematics (New York), vol. 43 (Springer, Berlin, Heidelberg, 1999), Hamiltonian systems and HJB equations 32. Y.N. Zhang, W.A. Zheng, Discretizing a backward stochastic differential equation. Int. J. Math. Math. Sci. 32(2), 103–116 (2002) 33. W.A. Zheng, On Symmetric Diffusion Processes, Probability theory and its applications in China, Contemp. Math., vol. 118, Am. Math. Soc., Providence, RI, 1991, pp. 329–333
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition Markus Mocha and Nicholas Westray
Abstract In the present article we provide existence, uniqueness and stability results under an exponential moments condition for quadratic semimartingale backward stochastic differential equations (BSDEs) having convex generators. We show that the martingale part of the BSDE solution defines a true change of measure and provide an example which demonstrates that pointwise convergence of the drivers is not sufficient to guarantee a stability result within our framework. Keywords Quadratic semimartingale BSDEs • Convex generators • Exponential moments AMS Classification: 60H10
1 Introduction Since their introduction by Bismut [3] within the Pontryagin maximum principle, backward stochastic differential equations (BSDEs) have attracted much attention in the mathematical literature. In a Brownian framework such equations are usually
This chapter was originally published within the PhD thesis “Utility Maximization and Quadratic BSDEs under Exponential Moments” by Markus Mocha, Humboldt University Berlin, 2012. With kind permission of the author. M. Mocha () Institut für Mathematik, Humboldt-Universität zu Berlin, Unter den Linden 6, 10099 Berlin, Germany e-mail:
[email protected] N. Westray Department of Mathematics, Imperial College, London SW7 2AZ, UK e-mail:
[email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__5, © Springer-Verlag Berlin Heidelberg 2012
105
106
M. Mocha and N. Westray
written d Yt D Zt d Wt F .t; Yt ; Zt / dt;
YT D ;
(1)
where is an FT -measurable random variable, the terminal value, and F is the so called driver or generator. Here .Ft /t 2Œ0;T denotes the filtration generated by the one-dimensional Brownian motion W . Solving such a BSDE corresponds to finding a pair of adapted processes .Y; Z/ such that the integrated version of (1) holds. The presence of the control process Z stems from the requirement of adaptedness for Y together with the fact that Y must be driven into the random variable at time T . One may think of Z as arising from the martingale representation theorem. In the more general semimartingale framework, where the main source of randomness is encoded in a given continuous local martingale M on a filtration .Ft /t 2Œ0;T that is not necessarily generated by M , we have to add an extra orthogonal component N . The corresponding BSDE then takes the form d Yt D Zt dMt C dNt f .t; Yt ; Zt / d hM it gt d hN it ;
YT D :
(2)
Solving (2) now corresponds to finding an adapted triple .Y; Z; N / of processes satisfying the integrated version of (2), where N is a (continuous) local martingale orthogonal to M . We refer to Z M C N as the martingale part of a solution to the BSDE (2). BSDEs of type (1) and (2) have found many fields of application in mathematical finance. The first problems to be attacked by means of such equations included pricing and hedging, superreplication and recursive utility, for the latter see for instance Schroder and Skiadas [30] and Skiadas [32]. For a general overview the reader is directed to the survey articles El Karoui, Peng and Quenez [12] and El Karoui, Hamadéne and Matoussi [10] and the references therein. A second large focus has been on their use in constrained utility maximization. In a Brownian setting Hu, Imkeller and MRuller [16] used the martingale optimality principle to derive a BSDE for the value process characterizing the optimal wealth and investment strategy. Their article can be regarded as an extension of earlier work by Rouge and El Karoui [29] as well as Sekine [31]. In related work in a semimartingale setting Mania and Schweizer [22] used a BSDE to describe the dynamic indifference price for exponential utility. Their stochastic control approach was extended to robust utility in Bordigoni, Matoussi and Schweizer [4] and to an infinite time horizon in the recent article by Hu and Schweizer [17]. We also mention Becherer [2] for further extensions to BSDEs with jumps and Mania and Tevzadze [23] to backward stochastic partial differential equations. This list is by far not exhaustive and additional references can be found in the stated papers. With regards to the theory of BSDEs, existence and uniqueness results were first provided in a Brownian setting by Pardoux and Peng [27] and in a semimartingale setting by El Karoui and Huang [11] both under Lipschitz conditions. These were extended (in the Brownian case) by Lepeltier and San Martín [21] to continuous drivers with linear growth and by Kobylanski [20] to generators which are quadratic as a function of the control variable Z. The corresponding results for the semi-
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
107
martingale case may be found in Morlais [25] and Tevzadze [33], where in the former the main theorems of [16] are extended. In addition a stability result for quadratic BSDEs may also be found in the recent article by Frei [13]. In the situation when the generator has superquadratic growth, Delbaen, Hu and Bao [8] show that such BSDEs are essentially ill-posed. A strong requirement present in the articles [20, 25, 33] is that the terminal condition be bounded. In a Brownian setting Briand and Hu [5,6] have replaced this by the assumption that it need only have exponential moments but in addition the driver is convex in the Z variable. By interpreting the Y component as the solution to a stochastic control problem, Delbaen, Hu and Richou [9] extend their results and show that one can reduce the order of exponential moments required. Let us finally mention a paper of Barrieu and El Karoui [1], prepared independently of the present article. They prove stability theorems for a class of continuous semimartingales whose finite variation processes satisfy a quadratic growth condition. Viewing BSDEs as a subclass they then derive existence and monotone stability results under a slightly weaker exponential moments condition than that given here, however there is no uniqueness. The present article has two main contributions, the first is to extend the existence, uniqueness and stability theorems of [6] and [25] to the unbounded semimartingale case. The motivation here is predominantly mathematical, having results in greater generality increases the range of applications for BSDEs. We remark however, that there are additional practical applications for the results derived here, e.g. related to utility maximization with an unbounded mean variance tradeoff, see Nutz [26] and Mocha and Westray [24], which provides a second motivation for the present work. In order to prove the respective results in the unbounded semimartingale framework technical difficulties related to an a priori estimate must be overcome. This requires an additional assumption when compared to [6] and [25]. As a biproduct of establishing our results we are able to show via an example that the stability theorem as stated in [6] Proposition 7 needs a minor amendment to the mode of convergence assumed on the drivers and we include the appropriate formulation. Our second contribution is to address the question of measure change. It is a classical result that when the generator has quadratic growth in z then the solution processes Y is bounded if and only if the martingale part Z M C N is a BMO martingale. In the present setting, where Y is assumed to satisfy an exponential moments condition only, such a correspondence is lost. However, we are able to show that whilst Z M C N need not be a BMO martingale, see Frei, Mocha and Westray [14] for further discussion and some examples, the stochastic exponential E q.Z M C N / is still a true martingale for q valid in some half-line. However, it is not only mathematically interesting to be able to describe the properties of the martingale part of the BSDE but also relevant for applications, notably in an unbounded setting. For instance, the above result can be used to extend the results of [16] and [25] on utility maximization. Moreover such a theorem may be used in the partial equilibrium framework of Horst, Pirvu and dos Reis [15] where the market price of external risk is given by equilibrium considerations and is typically unbounded.
108
M. Mocha and N. Westray
The paper is organized as follows. In the next section we lay out the notation and the assumptions and state the main results. The subsequent sections contain the proofs. Section 3 gives the a priori estimates together with some remarks on the necessity of an additional assumption, Sect. 4 deals with existence and Sect. 5 includes the comparison and uniqueness results. In Sects. 6 and 7 we prove the stability property as well as providing an interesting counterexample. In Sect. 8, we turn our attention to the measure change problem and finally, in Sect. 9, we give interesting applications of our results to constrained utility maximization and partial equilibrium models.
2 Model Formulation and Statement of Results We work on a filtered probability space .˝; F ; .Ft /0t T ; P/ satisfying the usual conditions of right-continuity and completeness. We also assume that F0 is the completion of the trivial -algebra. The time horizon T is a finite number in .0; 1/ and all semimartingales are considered equal to their càdlàg modification. Throughout this paper M D .M 1 ; : : : ; M d /T stands for a continuous d -dimensional local martingale, where T denotes transposition. We refer the reader to Jacod and Shiryaev [18] and Protter [28] for further details on the general theory of stochastic integration. The objects of study in the present paper will be semimartingale BSDEs considered on Œ0; T . In the d -dimensional case such a BSDE may be written d Yt D ZtT dMt C dNt 1T d hM it f .t; Yt ; Zt / gt d hN it ;
YT D :
(3)
Here is an R-valued FT -measurable random variable and f and g are random predictable functions Œ0; T ˝ R Rd ! Rd and Œ0; T ˝ ! R, respectively. We set 1 WD .1; : : : ; 1/T 2 Rd . The format in which the BSDE (3) encodes its finite variation parts is not so tractable from the point of view of analysis. Therefore we write semimartingale BSDEs by factorizing the matrix-valued process hM i D hM i ; M j ii;j D1;:::;d . This separates its matrix property from its nature as measure. For i; j 2 f1; : : : ; d g we may write hM i ; M j i D C ij A where C ij are the components of a predictable process C valued in the space of symmetric positive semidefinite d d matrices and A is a predictable increasing process. There are many such factorizations (cf. [18] Sect. III.4a). We may choose A WD Pd ˝ i ˛ M so that A is uniformly bounded by KA D =2 and derive the arctan i D1 absolute continuity of all the hM i ; M j i with respect to A from the Kunita– Watanabe inequality. This together with the Radon-Nikodým theorem provides C . Furthermore, we can factorize C as C D B T B for a predictable process B valued in the space of d d matrices. We note that all the results below do not rely on the specific choice of A, but only on its boundedness. In particular, if M D W is
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
109
a d -dimensional Brownian motion we may choose At D t, t 2 Œ0; T , and B the identity matrix. Then A is bounded by KA D T: We let P denote the predictable -algebra on Œ0; T ˝ generated by all the left-continuous adapted processes. The process A induces a measure A on P, the Doléans measure, defined for E 2 P by Z T 1E .t/ dAt : A .E/ WD E 0
Given the above discussion (3) may be rewritten as d Yt D ZtT dMt C dNt F .t; Yt ; Zt / dAt gt d hN it ;
YT D ;
(4)
where again is an R-valued FT -measurable random variable, the terminal condition, and F and g are random predictable functions Œ0; T ˝ R Rd ! R and Œ0; T ˝ ! R respectively, called generators or drivers. This formulation of the BSDE is very flexible, allowing for various applications and being amenable to analysis. Starting with (3) and setting F .t; y; z/ WD 1T Ct f .t; y; z/ D 1T BtT Bt f .t; y; z/ we get (4). The reversion of this procedure is not so clear, however is not relevant in applications. Under boundedness assumptions, existence of solutions to (4) is provided in [25] via an exponential transformation that makes the d hN i term disappear. A necessary condition for this kind of transformation to work properly is dg D 0. In the sequel we thus consider the above BSDE to be given in the form d Yt D ZtT dMt C dNt F .t; Yt ; Zt / dAt
1 d hN it ; 2
YT D ;
(5)
except in specific situations where a solution is assumed to exist. Definition 1. A solution to the BSDE (4), or (5), is a triple .Y; Z; N / of processes valued in R Rd R satisfying (4), or (5), P-a.s. such that: (i) The function t 7! Yt is continuous P-a.s. (ii) The process Z is predictable and M -integrable, in particular RT T 0 Zt d hM it Zt < C1 P-a.s. (iii) The local martingale N is continuous and orthogonal to each component of M , i.e. hN; M i i D 0 for all i D 1; : : : ; d . (iv) We have that P-a.s. Z
T
jF .t; Yt ; Zt /j dAt C hN iT < C1: 0
As in the introduction we call Z M C N the martingale part of a solution.
110
M. Mocha and N. Westray
In what follows we collect together the assumptions that allow for all the assertions of this paper to hold simultaneously. However we want to point out that not all of our results require that every item of Assumption 1 be satisfied, as will be indicated in appropriate remarks. Assumption 1 There exist nonnegative constants ˇ and ˇ, positive numbers ˇf and max.1; ˇ/ together with an M -integrable (predictable) Rd -valued process so that writing Z
T
˛ WD kBk2 and j˛j1 WD
Z ˛t dAt D
0
0
T
Tt d hM it t
we have P-a.s. (i) The random variable jj C j˛j1 has exponential moments of all orders, i.e. for all p > 1 h i < C1: (6) E exp p jj C j˛j1 (ii) For all t 2 Œ0; T the driver .y; z/ 7! F .t; y; z/ is continuous in .y; z/, convex in z and Lipschitz continuous in y with Lipschitz constant ˇ, i.e. for all y1 , y2 and z we have jF .t; y1 ; z/ F .t; y2 ; z/j ˇ jy1 y2 j:
(7)
(iii) The generator F satisfies a quadratic growth condition in z, i.e. for all t; y and z we have jF .t; y; z/j ˛t C ˛t ˇjyj C kBt zk2 : (8) 2 (iv) The function F is locally Lipschitz in z, i.e. for all t; y; z1 and z2 jF .t; y; z1 / F .t; y; z2 /j ˇf kBt t k C kBt z1 k C kBt z2 k kBt .z1 z2 /k: (v) The constant ˇ in (iii) equals zero and then we set cA WD 0: Alternatively, ˇ > 0, but additionally assume that for all t, y and z we have jF .t; y; z/ F .t; 0; z/j ˇ jyj
and
At cA t
for a positive constant cA . If this assumption is satisfied we refer to (5) as BSDE.F; / with the set of parameters .˛; ˇ; ˇ; ˇf ; /. Remark 1. The above items (i)–(iv) correspond to the assumptions made in [6] and [25]. In particular, the BSDEs under consideration are of quadratic type (in the control variable z) and of Lipschitz type in y. Item (v) is new and arises from the fact that the methods used in [25] to derive an a priori estimate may no longer be directly
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
111
applied so that an additional assumption is required. We elaborate further on this topic in Sect. 3. Observe that in the key application of utility maximization, cf. [24], the associated driver is independent of y and hence ˇ D 0 applies. In particular, we need not assume that the mean-variance tradeoff of the underlying market be bounded. Notice that items (ii) and (iii) from above provide jF .t; y; z/j ˛t C ˇjyj C
kBt zk2 ; 2
(9)
for all t, y and z, P-a.s. This is an inequality which does not involve ˛ in the jyj term on the right hand side and which is used repeatedly throughout the proofs. We also define the constant (10) ˇ WD cA ˇ: Before giving the main results of the paper let us introduce some notation. For p 1, S p denotes the set of R-valued, adapted and continuous processes Y on Œ0; T such that #1=p " E
sup jYt jp
< C1:
0t T
The space S 1 consists of the continuous bounded processes. An R-valued, adapted and continuous process Y belongs to E if the random variable Y WD sup jYt j t 2Œ0;T
has exponential moments of all orders. We also recall that Y is called of class D if the family fY j 2 Œ0; T stopping timeg is uniformly integrable. The set of (equivalence classes of) Rd -valued predictable processes Z on Œ0; T ˝ satisfying " Z E
p=2 #1=p
T
Zt d hM it Zt T
0
< C1
is denoted by Mp . Finally, M p stands for the set of R-valued martingales N on Œ0; T , such that i1=p h p=2 kN kM p WD E hN iT < C1: Notice that if the following assumption on the filtration holds the elements of M p are continuous. Assumption 2 The filtration .Ft /t 2Œ0;T is a continuous filtration, in the sense that all local .Ft /t 2Œ0;T -martingales are continuous. The following four theorems constitute the main results of the paper. We mention that only the existence result requires the assumption of the continuity of the filtration.
112
M. Mocha and N. Westray
Theorem 1 (Existence). If Assumptions 1 and 2 hold there exists a solution .Y; Z; N / to the BSDE (5) such that Y 2 E and Z M C N 2 M p for all p 1. Theorem 2 (Uniqueness). Suppose that Assumption 1 holds. Then any two solutions .Y; Z; N / and .Y 0 ; Z 0 ; N 0 / in E M2 M 2 to the BSDE (5) coincide in the sense that Y and Y 0 , Z M and Z 0 M , and N and N 0 are indistinguishable. Theorem 3 (Stability). Consider a family of BSDEs(F n ; n ) indexed by the extended natural numbers n 0 for which Assumption 1 holds true with parameters .˛ n ; ˇ n ; ˇ; ˇf ; /. Assume that the exponential moments assumption (6) holds uniformly in n, i.e. for all p > 1, i h n n sup E e p .j jCj˛ j1 / < C1: n0
If for n 0 .Y n ; Z n ; N n / is the solution in E M2 M 2 to the BSDE(F n ; n ) and if Z
T
j n 0 j C 0
ˇ n ˇ ˇF F 0 ˇ .s; Y 0 ; Z 0 / dAs ! 0 s s
in probability, as n ! C1; (11)
then for each p 1 as n ! C1 " E
exp
ˇ ˇ sup ˇYtn Yt0 ˇ
0t T
!!p # ! 1
and
Z n M C N n ! Z 0 M C N 0 in M p : Theorem 4 (Exponential Martingales). Suppose that Assumption 1 holds, let jqj > =2 and let .Y; Z; N/ 2 E M2 M 2 be a solution to the BSDE (5). Then the local martingale E q .Z M C N / is a true martingale on Œ0; T . Remark 2. The preceding theorems generalize the results of [6] and [25]. For their proofs we combine the localization and -technique from [6] together with the existence and stability results for BSDEs with bounded solutions found in [25]. Similar ideas are used in [17] on a specific quadratic BSDE arising in a robust utility maximization problem where the authors also investigate the measure change problem for their special BSDE, however here we pursue the general theory. We point out that when the BSDE is of quadratic type and jj C j˛j1 does not have sufficiently large exponential moments there are examples where the BSDE admits no solution, see [14]. In particular, we present all the theoretical background for the study of utility maximization under exponential moments, see [24], as well as partial equilibrium, see [15].
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
113
3 A Priori Estimates In this section we show that, under appropriate conditions, solutions to the BSDE (4) satisfy some a priori norm bounds. After giving an important result used in the subsequent sections we motivate Assumption 1 (v) by showing that without such an assumption the method utilized in [25] for the purpose of deriving appropriate a priori bounds fails in the present unbounded case. Let .Y; Z; N / be a solution to (4), suppose that Assumption 1 (iii) and (v) hold and that g is uniformly bounded by =2. Recall ˇ from (10), fix s 2 Œ0; T and set, for t 2 Œs; T ,
Z t ˇ .t s/ ˇ .rs/ e H t WD exp e jYt j C e d h M ir : s
Rt Rt where we have written h M it WD 0 Tr d hM ir r D 0 ˛r dAr . First we show e is, up to integrability, a local submartingale. that H From Tanaka’s formula, d jYt j D sgn.Yt /.ZtT dMt C dNt / sgn.Yt / F .t; Yt ; Zt / dAt C gt d hN it C dLt ; (12) where L is the local time of Y at 0. Itô’s formula then yields " ˇ .t s/ e e sgn.Yt /.ZtT dMt C dNt / C ˇjYt j.cA dt dAt / dHt D Ht e
(13)
C sgn.Yt /F .t; Yt ; Zt / C ˛t C ˇjYt j C e ˇ .t s/ kBt Zt k2 dAt 2 # ˇ .t s/ d hN it C dLt : C sgn.Yt / gt C e 2 An inspection of the finite variation parts shows that under the present assumptions e is a local submartingale, they are nonnegative. In particular, the semimartingale H which leads to the following result. Proposition 1 (A Priori Estimate). Suppose Assumption 1 (iii) and (v) hold and assume that the function g is uniformly bounded by =2, P-a.s. Let .Y; Z; N / be a solution to the BSDE (4) and let the process
Z ˇ T jY j C exp e
T
e 0
be of class D. Then P-a.s. for all s 2 Œ0; T ,
ˇ r
d h M ir
114
M. Mocha and N. Westray
"
ˇˇ # Z T 1 ˇ ˇ .T s/ ˇ .rs/ jYs j log E exp e jj C e d h M ir ˇ Fs : ˇ s
(14)
e as above. Since H e is a local submartingale there Proof. Fix s 2 Œ0; T and set H exists a sequence of stopping times . n /n1 valued in Œs; T , which converges P-a.s. e n is a submartingale for each n 1: We then derive to T , such that H e T ^ n j Fs exp. jYs j/ EŒH "
Z E exp e ˇ .T s/ jYT ^ n j C
T
e
ˇ .rs/
s
ˇˇ # ˇ d h M ir ˇ Fs : ˇ
Letting n ! C1 the claim follows from the class D assumption.
t u
Proposition 1 provides the appropriate a priori estimate, indeed suppose that jj and j˛j1 are bounded random variables and .Y; Z; N / is a solution to (5). If the current assumptions hold and exp.e ˇ T jY j/ is of class D, then Y satisfies jY j e ˇ T .jj C j˛j1 / : 1
(15)
Comparing with (14) this indicates that the inclusion of Assumption 1 (v) allows us to prove similar estimates to the bounded case which enables us to establish existence for the BSDE (5) when jj C j˛j1 has exponential moments of all orders, to be more precise, an order of at least e ˇ T . Contrary to the above let us investigate the method utilized in [25] under Assumption 1 (iii) only, supposing that g be bounded by =2. We set
Z t Ht WD exp e ˇhM is;t jYt j C e ˇhM is;r d h M ir ;
(16)
s
where h M is;t WD h M it h M is D
Rt s
˛r dAr . We derive from Itô’s formula
" dHt D Ht e ˇhM is;t sgn.Yt /.ZtT dMt C dNt / C sgn.Yt /F .t; Yt ; Zt / C ˛t C ˛t ˇjYt j C e ˇhM is;t kBt Zt k2 dAt 2 # ˇhM is;t d hN it C dLt : C sgn.Yt / gt C e 2 Once again, the finite variation parts are nonnegative. We conclude in the same way as for Proposition 1 that the corresponding a priori result holds for H as well. To sum up, we have that under a similar class D assumption, now on
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
Z exp e ˇhM iT jY j C
T
e
ˇhM ir
115
d h M ir ;
0
P-a.s. for all s 2 Œ0; T , "
ˇˇ # Z T 1 ˇ ˇhM is;T ˇhM is;r jYs j log E exp e jj C e d h M ir ˇ Fs : ˇ s
(17)
e from above equals H and there is no difference with the statement If ˇ D 0, then H of Proposition 1. However when ˇ > 0 the estimate (17) is not sufficient for our purposes. We aim at using the a priori estimate to show the existence of solutions to the BSDE (5) in E M2 M 2 using an appropriate approximating procedure. If jj and j˛j1 are bounded random variables there exists a solution .Y; Z; N / to (5) with Y bounded, cf. [25]. With (17) at our disposal we then have the estimate jY j e ˇj˛j1 jj C j˛j1 : 1
(18)
Our goal is to remove the boundedness assumption and to replace it with the assumption on the existence of exponential moments of jjCj˛j1 in the spirit of [6]. However a closer inspection of the a priori estimate from (17) together with (18) already indicates that more restrictive assumptions are necessary. More specifically, when ˇ > 0 we cannot deduce any integrability of exp e ˇj˛j1 jj C j˛j1 when jj and j˛j1 have only exponential moments. This motivates Assumption 1 (v). Note that we could opt for deriving the existence result under the weaker assumption that the above random variable be integrable. In this case, describing the space in which a solution to the BSDE exists is more technical, as would be a statement of uniqueness. See also [1].
4 Existence In the present section we establish Theorem 1 together with some related results on norm bounds of the solution. The proof of existence follows the following recipe. First we truncate h M i to get approximate solutions. Then by using the estimate from Proposition 1 we localize and work on a random time interval so that the approximations are uniformly bounded and we can apply a stability result. Finally we glue together on Œ0; T to construct a solution. The a priori estimates ensure that we may take all limits in the described procedure. Theorem 5 (Existence). Let Assumptions 1 (ii)–(v) and 2 hold and jj C j˛j1 have an exponential moment of order e ˇ T . Then the BSDE (5) has a solution .Y; Z; N / such that
116
M. Mocha and N. Westray
"
ˇˇ # Z T 1 ˇ ˇ .T t / ˇ .rt / jYt j log E exp e jj C e d h M ir ˇ Ft : ˇ t
(19)
Proof. Exactly as in [6] we first assume that F and are nonnegative. For each integer n 1, set ˇ
Z t ˇ n WD inf t 2 Œ0; T ˇˇh M it WD ˛s dAs n ^ T; 0
n WD ^ n, nt WD 1ft ng t and F n .t; y; z/ WD 1ft ng F .t; y; z/. Then F n satisfies Assumption 1 (ii)–(v) with the same constants, but with the processes n and ˛ n where ˛tn WD kBt nt k2 D 1ft ng kBt t k2 D 1ft ng ˛t : R In particular, j˛ n j1 D 0 n ˛s dAs n and Z
Z
T 0
.nt /T d hM it nt D
0
T
kBt nt k2 dAt D j˛ n j1 n;
so we may apply [25] Theorems 2.5 and 2.6 to conclude that there exists a unique solution .Y n ; Z n ; N n / 2 S 1 M2 M 2 to the BSDE (5), where F is replaced by F n and by n . From Proposition 1 we derive jYtn j
"
ˇˇ # Z T 1 ˇ ˇ .T t / n ˇ .rt / n log E exp e j j C e d h M ir ˇ Ft ˇ t ˇ "
ˇ # Z T 1 ˇ ˇ .T t / ˇ .rt / jj C e d h M ir ˇ Ft log E exp e ˇ t ˇ # " ˇˇ 1 log E exp e ˇ T jj C j˛j1 ˇ Ft DW Xt : (20) ˇ
Let n m so that then we have n m and 1ft ng 1ft mg . In particular, n m and F n F m , from which we deduce that the Assumptions 1 (ii)–(v), hence the corresponding assumptions in [25], hold for both F n and F m with the same set of parameters .˛ m ; ˇ; ˇ; ˇf ; / where the additional c in [25] is equal to m. An application of Theorem 2.7 therein now shows that Y n Y m so that .Y n /n1 is an increasing sequence of bounded continuous processes. The next step would be to send n to infinity, however, we do not dispose of a suitable stability result. Indeed we have only [25] Lemma 3.3 which applies for bounded processes under uniform growth assumptions on the drivers, hence we introduce an additional truncation. Let k 1 be a fixed integer and ˇ n o ˇ k WD inf t 2 Œ0; T ˇ Xt k or h M it k ^ T:
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
117
Thanks to the continuity of the filtration the martingale exp.X / is continuous so that the random variable V WD max Xt _ h M iT t 2Œ0;T
is finite P-a.s. We derive that P-a.s. k D T for large k. Due to (20) the sequence .Y n;k /n1 given by Ytn;k WD Ytn^ k ; is uniformly bounded by k. For the martingale parts we define Ztn;k WD 1ft kg Ztn and Ntn;k WD 1ft kg Ntn : An inspection of the respective cases shows that Z Ytn;k
D
Y nk Z
T
t
Zsn;k
T
Z
T
dMs t
dNsn;k
T
C t
1fs k ^ng F .s; Ysn;k ; Zsn;k / dAs
1 C 2
Z
T
d hN n;k is :
t
n"C1
Moreover, Y nk ! supn1 Y nk DW k , where k is bounded by k. Next we appeal to the stability result stated in [25] Lemma 3.3, noting Remark 3.4 therein. Note that this result requires estimates that are uniform in n which is accomplished by the specific choice of the stopping time k . Hence .Y n;k ; Z n;k ; N n;k / converges to .Y 1;k ; Z 1;k ; N 1;k / in the sense that lim E
n!C1
Z lim E
n!C1
and
T 0
! ˇ n;k ˇ 1;k sup ˇYt Yt ˇ D 0;
0t T
T Zsn;k Zs1;k d hM is Zsn;k Zs1;k D0
ˇ ˇ2 ˇ ˇ lim E ˇNTn;k NT1;k ˇ D 0;
n!C1
where the triples .Y 1;k ; Z 1;k ; N 1;k / solve the BSDE T d Yt1;k D Zt1;k dMt C dNt1;k 1ft kg F .t; Yt1;k ; Zt1;k / dAt
1 d hN 1;k it ; 2
Y 1;k D k ; k
118
M. Mocha and N. Westray
on the random horizon ŒŒ0; k Œ0; T . The stopping times k are monotone in k and therefore it follows that Ytn;kC1 D Ytn;k ; ^ k
1ft kg Ztn;kC1 D Ztn;k
and 1ft kg Ntn;kC1 D Ntn;k ;
so that the above convergence yields (for the two last objects in M 2 ) D Yt1;k ; Yt1;kC1 ^ k
1ft k g Z 1;kC1 M t D Z 1;k M t
and 1ft kg Nt1;kC1 D Nt1;k : To finish the proof, we define the processes Yt WD 1ft 1g Yt1;1 C
X
1ft 2 k1 ; k g Yt1;k ;
k2
Zt WD 1ft 1g Zt1;1 C
X
1ft 2 k1 ; k g Zt1;k
k2
and Nt WD 1ft 1g Nt1;1 C
X
1ft 2 k1 ; k g Nt1;k :
k2
By construction this gives a solution to the BSDE d Yt D ZtT dMt C dNt F .t; Yt ; Zt / dAt
1 d hN it ; 2
YT D ;
P since 1ft 1g C k2 1ft 2 k1 ; k g D 1ft 2Œ0;T g P-a.s. More precisely, there is a P-null set N such that for all ! 2 Nc there is a minimal k0 .!/ with k0 .!/ .!/ D T and such that Y 1;k .!/ D k .!/ for all k, which yields that (possibly after another k .!/ modification of N) 1;k0 .!/
YT .!/ D YT
.!/ D k0 .!/ .!/ D sup YTn .!/ D .!/: n1
The bound in (19) holds as we have it for all n and k from (20). In the case when and f are not necessarily nonnegative, we proceed as in [6] by using a double truncation defined by n;m WD C ^ n ^ m, n;m WD 1ft ng C 1ft mg and F n;m WD 1ft ng F C 1ft mg F . t u As an immediate corollary we deduce Corollary 1 (Norm Bounds). (i) Let Assumptions 1 (ii)–(v) and 2 hold and jjCj˛j1 have an exponential moment of order ı > e ˇ T . Then the BSDE (5) has a solution .Y; Z; N / such that e Y 2 S p for p WD eıˇ T > 1.
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
119
When additionally jj C j˛j1 has exponential moments of all orders, i.e. Assumption 1 (i) holds, this solution is such that Y 2 E. In particular, for each p > 1 we have the estimate h i p p
E e pY E exp pe ˇ T jj C j˛j1 : p1
(21)
(ii) Let Assumption 1 (i)–(iii) and (v) hold and suppose there exists a solution .Y; Z; N / to the BSDE (5) such that Y 2 E. Then .Z; N / 2 Mp M p for all p 1, more precisely " Z p=2 #
T T ˇ T jj C j˛j1 ; Zs d hM is Zs C d hN is cp; E exp 4pe E 0
(22) where cp; is a positive constant depending on p and . The estimate (21) then holds as well. Proof. (i) Let .Y; Z; N / be the solution to (5) obtained in Theorem 5. As in the previous section set
Z t e t WD exp e ˇ t jYt j C H e ˇ r d h M ir ;
(23)
0
which is a local submartingale. R Moreover, from the estimate (19), Jensen’s inequality and the adaptedness of 0 e ˇ r d h M ir we deduce that
Z t h ie ˇ t ˇ r e H t D exp. jYt j/ exp e d h M ir 0
ˇˇ #e ˇ t Z T ˇ e ˇ .rt / d h M ir ˇ Ft E exp e ˇ .T t / jj C ˇ t
Z t exp e ˇ r d h M ir
"
0
ˇ #
ˇ ˇ ˇ T jj C j˛j1 ˇ Ft : E exp e ˇ "
Observe that this upper estimate is a uniformly integrable martingale, in particular it e is a true submartingale. Then, via the Doob maximal is of class D and therefore H inequality, we find that for p > 1
120
M. Mocha and N. Westray
#
" p h i p p pY et ep E e EŒH E sup H T p 1 0t T p
p ; (24) E exp pe ˇ T jj C j˛j1 p1
provided the right hand side is finite. In particular, e Y 2 S p and Y 2 E as soon as jj C j˛j1 has exponential moments of all orders, in which case (21) holds. (ii) We first verify that (21) continues to hold when .Y; Z; N / is a solution to (5) with Y 2 E. Observe that we may reformulate the result of Proposition 1 under the condition that
Z exp e ˇ
T
jY j C
eˇ
r
d h M ir
0
be of class D. Repeating the argument from (i) using (14) instead of (19) leads to the same conclusion, since we have the relation ˇ # "
Z t ˇ ˇ e ˇ r d h M ir E exp e ˇ T Y C j˛j1 ˇ Ft ; exp e ˇ t jYt j C ˇ 0 where the right hand side is indeed a process of class D. For the remaining claim, relation (22), define the functions u; v W R ! RC via u.x/ WD 12 .e x 1 x/ and v.x/ WD u.jxj/. We have that v is a C 2 function, so we use Itô’s formula to see that for a stopping time (to be chosen later) Z
t ^
v.Y0 / D v.Yt ^ / 0
u0 .jYs j/ sgn .Ys /.ZsT dMs C dNs /
1 C u .jYs j/ sgn .Ys / F .s; Ys ; Zs / dAs C d hN is 2 0 Z t ^ 1 u00 .jYs j/ ZsT d hM is Zs C d hN is ; 2 0 Z
t ^
0
where use the notation sgn .x/ WD 1fx0g C 1fx>0g and observe that u0 .0/ D 0. Assumption 1 (iii) yields Z v.Y0 / v.Yt ^ / 0
Z
t ^
C 0
1 C 2
Z
u0 .jYs j/ sgn .Ys /.ZsT dMs C dNs /
u0 .jYs j/ ˛s C ˛s ˇjYs j dAs
t ^ 0
t ^
u0 .jYs j/ u00 .jYs j/ ZsT d hM is Zs
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
1 C 2
Z
t ^
121
u0 .jYs j/ sgn .Ys / u00 .jYs j/ d hN is ;
0
since u0 .x/ D 1 .e x 1/ 0 for x 0. Using the relation u0 .x/ u00 .x/ D 1 together with 1 it follows that Z
t ^
0 v.Y0 / v.Yt ^ / 0
Z
t ^
C 0
u0 .jYs j/ sgn .Ys /.ZsT dMs C dNs /
Z 1 t ^ T u0 .jYs j/ ˛s C ˛s ˇjYs j dAs Zs d hM is Zs C d hN is : 2 0 (25)
Suppose first that p 2. Then (22) can be proved using the Burkholder–Davis– Gundy inequalities as follows. From (25) we deduce that 1 2
Z
0
Z 1 Y 1 T jYs j ˛ e C e C ˛ ˇjY j dAs s s s 2 0 ˇZ t ^ ˇ ˇ ˇ 0 T ˇ C sup ˇ u .jYs j/ sgn .Ys /.Zs dMs C dNs /ˇˇ ;
ZsT d hM is Zs C d hN is
0t T
0
where we used the estimates u0 .x/ e x = and v.x/ e x = 2 , valid for x 0. From the inequalities y e y 1 and ˇ we derive
Z
p=2
0
1 1 p=2 Y p=2 e C p=2 e pY j˛j1 p ˇZ ˇp=2 ! ˇ t ^ ˇ ˇ ˇ 0 T ; C sup ˇ u .jYs j/sgn .Ys /.Zs dMs C dNs /ˇ ˇ 0t T ˇ 0
ZsT d hM is Zs C d hN is
2 3p=22
which yields, after taking expectation and applying the estimate jxjp=2 < e p=2 jxj and the Burkholder–Davis–Gundy inequality, " Z E
p=2 # Zs d hM is Zs C d hN is T
0
" Z C cp; E 0
h i cp; E e p=2 Y C e pY e p=2 j˛j1
e
2 jYs j
! # p=4 Zs d hM is Zs C d hN is ; T
where we used the estimate u0 .x/ e x = for x 0. Note that in the above and in what follows cp; > 0 is a generic constant depending on p and that may change
122
M. Mocha and N. Westray
from line to line. We apply the generalized Young inequality, jabj " WD 1 and for " WD cp; . Then, after an adjustment of cp; , " Z
" 2
a2 C
b2 , 2"
for
p=2 #
Zs d hM is Zs C d hN is
E
T
0
h i 1 h i 1 h i cp; E e p=2 Y C E e 2pY C E e p j˛j1 2 2 " Z p=2 # i 1 h pY T C cp; E e C E Zs d hM is Zs C d hN is 2 0 h i h i cp; E e 2pY C E e 2p j˛j1 " Z p=2 # 1 T C E Zs d hM is Zs C d hN is : 2 0 Next define, for each integer n 1, the stopping time ˇZ t
ˇ 2 jYs j T ˇ n WD inf t 2 Œ0; T ˇ e Zs d hM is Zs C d hN is n ^ T: 0
Inserting n into the above calculation and using e a C e b 2e aCb for a; b 0 together with (21), we may rewrite the last estimate as " Z
n
E
p=2 # Zs d hM is Zs C d hN is T
0
ˇ T jj C j˛j1 : cp; E exp 2pe (26)
By Fatou’s lemma, since n ! T as n ! C1, " Z
p=2 #
T
E
Zs d hM is Zs C d hN is T
0
cp; E exp 2pe ˇ T jj C j˛j1
and (22) follows. In the situation where p < 2, q WD 2=p > 1 and we may combine Jensen’s inequality with (26), which is valid for p D 2, to get " Z E
n
p=2 #q Zs d hM is Zs C d hN is T
0
"Z
n
E
#
Zs d hM is Zs C d hN is c2; E exp 4e ˇ T jj C j˛j1 T
0
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
123
from which (22) follows after another application of Fatou’s lemma together with the fact that the right hand side in the inequality above is greater or equal one while 1=q D p=2 < 1. t u Remark 3. We point out that the results of this section do not require that F be convex in z, but only that F be continuous in .y; z/. The reader may have noticed that the continuity of F is not used directly in the proofs. However in Theorem 5 we rely on the results of [25] where continuity is a technical condition needed for an application of Dini’s theorem. In addition our results also apply to the BSDE (4) if g is identically equal to a nonzero constant g =2, in which case we assume without loss of generality that jg j.
5 Uniqueness We now provide a comparison theorem that yields uniqueness of the BSDE triple. The proof makes use of the -technique applied in the context of second order Bellman-Isaacs equations by Da Lio and Ley [7] and subsequently adapted to the framework of Brownian BSDEs in [6]. We extend these ideas to take into account the orthogonal part of the BSDE solution. Theorem 6 (Comparison Principle). Let .Y; Z; N / and .Y 0 ; Z 0 ; N 0 / be solutions to the BSDE (5) with drivers F and F 0 and terminal conditions and 0 , respectively. Suppose in addition that Y 2 E and Y 0 2 E. If P-a.s. for all t 2 Œ0; T , 0
F .t; Yt0 ; Zt0 / F 0 .t; Yt0 ; Zt0 /;
and
and if .F; / satisfies Assumption 1 (i)–(iii) then P-a.s. for each t 2 Œ0; T Y t Y t0 : Proof. Let be a real number in .0; 1/ and set U WD Y Y 0 , V WD Z Z 0 and W WD N N 0 . Consider the process ( s WD
F .s;Ys ;Zs /F .s; Ys0 ;Zs / Us
if Us ¤ 0;
ˇ
if Us D 0:
Rs By Assumption 1 (ii), is bounded by ˇ and we define Rs WD 0 r dAr . Notice that by the boundedness of A we have that jRj ˇ AT ˇKA . From Itô’s formula we deduce Z
T
e Rt Ut D e RT UT t
Z
T
C t
e Rs .VsT dMs C d Ws /
1 d hN is d hN 0 is ; e Rs Fs dAs C 2
124
M. Mocha and N. Westray
where we define Fs WD F .s; Ys ; Zs / F 0 .s; Ys0 ; Zs0 / s Us . We also set F .s/ WD .F F 0 /.s; Ys0 ; Zs0 / 0 and observe that from the convexity of F in z together with (9) we get F .s; Ys0 ; Zs / F .s; Ys0 ; Zs0 /
Zs Zs0 F .s; Ys0 ; Zs0 / D F s; Ys0 ; Zs0 C .1 / 1
Zs Zs0 .1 / F s; Ys0 ; 1
kBs Vs k2 : .1 /˛s C .1 /ˇjYs0 j C 2.1 /
(27)
Another application of the Lipschitz Assumption 1 (ii), yields F .s; Ys ; Zs / F .s; Ys0 ; Zs / D F .s; Ys ; Zs / F .s; Ys0 ; Zs / C F .s; Ys0 ; Zs / F .s; Ys0 ; Zs / D s Us C F .s; Ys0 ; Zs / F .s; Ys0 ; Zs / s Us C .1 /ˇjYs0 j:
(28)
Combining (27) and (28) we see that Fs D F .s; Ys ; Zs / F .s; Ys0 ; Zs0 / C F .s/ s Us D ŒF .s; Ys ; Zs / F .s; Ys0 ; Zs / C ŒF .s; Ys0 ; Zs / F .s; Ys0 ; Zs0 / C F .s/ s Us .1 / ˛s C 2ˇjYs0 j C
kBs Vs k2 C F .s/: 2.1 /
(29)
exp.ˇKA / Let WD > 0 and Pt WD exp e Rt Ut > 0. The logic is now similar 1
to how we derived the a priori estimates, namely to show that, by removing an appropriate drift, P is a (local) submartingale. By Itô’s formula, for t 2 Œ0; T , Z Pt D PT
Z
T
Ps e
Rs
t
Z C t
T
Ps e Rs
.Vs dMs C d Ws / C
T
T
Ps e t
Rs
Fs
e Rs kBs Vs k2 2
1
e Rs d hN is d hN 0 is : d hW is C 2 2
! dAs
(30)
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
125
To simplify notation set
e R kBV k2 G WD P e R F and 2
Z
e Rs 1 d hW is C d hN is d hN 0 is :
Ps e Rs H WD 2 2 0
(31)
(32)
Let us first investigate the finite variation process H . We claim that H is decreasing, indeed for all r; u 2 Œ0; T , r u, we have Z
Z
u
u
d hW is D r
d hN is 2 d hN; N 0 is C 2 d hN 0 is :
r
Applying the Kunita–Watanabe and Young inequalities, Z
Z
u
Z
u
d hW is
u
d hN is C
r
Z
Z
d hN is C r
0
Z
r
Z
u
D .1 /
1=2
r
Z
u
d hN is
2
0
d hN is
r u
u
d hN is
r u
1=2 Z
u
d hN is 2
r
Z
0
2
u
d hN is C r
0
d hN is r
d hN is d hN 0 is :
r
In particular, since 1 and jRj ˇKA we have, Z
u
e d hW is 1
Z
r
Z
u
u
d hW is
Rs
r
d hN is d hN 0 is ;
r
which shows that the process H is decreasing and hence the third integral in (30) is nonpositive. Next we consider the other finite variation integral in (30). Combining (29), F 0 and the boundedness of R we have
e R G D P e R F kBV k2 P e R .1 / ˛ C 2ˇjY 0 j PJ; (33) 2 where
J WD e 2ˇKA ˛ C 2ˇjY 0 j 0:
126
M. Mocha and N. Westray
Z
We set
t
Dt WD exp
Js dAs
and
e t WD Dt Pt : P
0
Partial integration yields e t D Dt Gt dAt dHt C Pt e Rt VtT dMt C d Wt C Pt Dt Jt dAt dP (34) D Dt .Pt Jt Gt / dAt dHt C Pt e Rt VtT dMt C d Wt and we conclude that the finite variation parts in the latter expression are nonnegative. We can now use the following stopping time argument to derive ˇ ˇ DT (35) Pt E PT ˇˇ Ft : Dt Namely, consider the stopping time ˇZ u
ˇ e 2s e 2Rs VsT d hM is Vs C d hW is n ^ T; n WD inf u 2 Œt; T ˇˇ 2 P t
where n 1 is an integer. Observe that n ! T as n ! C1 due to the integrability assumptions on ˛, Y and Y 0 , as well as the boundedness of A. Then (34) provides the estimate ˇ Z n ˇ Js dAs P n ˇˇ Ft Pt E exp t
Z D E exp
n t
ˇ ˇ e 2ˇKA ˛s C 2ˇjYs0 j dAs P n ˇˇ Ft :
In view of the current integrability and boundedness assumptions we can send n to infinity and deduce (35). Notice that we also have 0 .1 /jj C , where WD 0 0. Then together with the definition of P the inequality (35) shows that ! e ˇKA CRt 0 exp Yt Yt 1
ˇ # "
Z T RT ˇ 2ˇKA 0 0 ˇ E exp ˛s C 2ˇjYs j dAs exp e ˇ Ft e ˇ t ˇ " #
Z T ˇ ˇ ˛s C 2ˇjYs0 j dAs exp e 2ˇKA jj ˇ Ft : E exp e 2ˇKA ˇ t Thus, we can derive the estimate
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
Yt
Yt0
127
"
ˇˇ # Z T 1
ˇ 2ˇKA 0 log E exp e jj C ˛s C 2ˇjYs j dAs ˇ Ft ; ˇ t
which follows from the above by checking the cases Yt Yt0 0 and Yt Yt0 < 0 separately, noting that R C ˇKA 0. Once again, by the integrability assumptions on , ˛ and Y 0 and the boundedness of A, the conditional expectation on the right hand side is finite, P-a.s. Taking " 1 then gives Yt Yt0 and the continuity of Y and Y 0 yields the claim. t u The following corollary is then immediate. Corollary 2 (Uniqueness). Let Assumption 1 (i)–(iii) hold and let .Y; Z; N / and .Y 0 ; Z 0 ; N 0 / be two solutions to the BSDE (5) with Y 2 E and Y 0 2 E. Then Y and Y 0 , Z M and Z 0 M , and N and N 0 are indistinguishable. In addition .Z M; N / and .Z 0 M; N 0 / both belong to M p M p for all p 1. Proof. By Theorem 6 and Corollary 1 (ii) only the assertion regarding the indistinguishability of the martingale part remains. Itô’s formula gives P-a.s. 0 D .YT YT0 /2 D .Y0 Y00 /2 C 2 Z
T
C 0
Z
T
D 0
Z 0
T
.Yt Yt0 / d.Yt Yt0 /
.Zt Zt0 /T d hM it .Zt Zt0 / C d hN N 0 it
.Zt Zt0 /T d hM it .Zt Zt0 / C d hN N 0 it ;
from which Z M Z 0 M and N N 0 .
t u
6 Stability It follows from the previous results that the BSDE (5) has a unique solution in E under appropriate Lipschitz and convexity assumptions on the driver F and an exponential moments condition on the terminal value and process ˛. In the present section we show that a stability result for such BSDEs also holds. More precisely, given a sequence of terminal values and a sequence of drivers such that the exponential moments condition is fulfilled uniformly, and such that they both converge to a fixed terminal value and a fixed generator in a suitable sense, then we gain convergence on the level of the respective BSDE solutions exactly as in [6]. Theorem 7 (Stability). Let .F n /n0 be a sequence of generators for the BSDE (5) such that Assumption 1 (ii)–(iii) and (v) hold for each F n with the set of parameters .˛ n ; ˇ n ; ˇ; ˇf ; /. If . n /n0 are the associated random terminal values then suppose that, for each p > 0,
128
M. Mocha and N. Westray
i h n n sup E e p .j jCj˛ j1 / < C1:
(36)
n0
Let .Y n ; Z n ; N n / be the solution to the BSDE (5) with driver F n and terminal condition n such that Y n 2 E for all n 0. If Z
T
j n 0 j C 0
ˇ n ˇ ˇF F 0 ˇ .s; Y 0 ; Z 0 / dAs ! 0 s s
in probability, as n ! C1; (37)
then for each p > 0, " lim E
exp
n!C1
" Z lim E
n!C1
!!p # D 1 and
0t T
p=2 #
T 0
ˇ ˇ sup ˇYtn Yt0 ˇ
.Zsn
Zs0 /T
d hM is .Zsn
Zs0 /
D 0:
C d hN N is n
0
Remark 4. Let us briefly indicate how the above stability theorem differs from those in the literature, [13] Theorem 2.1 and [25] Lemma 3.3. The key points are that first in our conditions the parameters ˛ n and ˇ n are allowed to depend on n, whereas in [13] and [25] they are assumed independent of n. Second, we assume a uniform exponential moments condition, as opposed to a uniform boundedness condition in the cited references. Finally, in the unbounded setting we require the mode of convergence assumed above, this is in contrast to the setting of [13] Theorem 2.1 where the weaker notion of pointwise convergence is sufficient for a stability result to hold. Proof. Note that Assumption 1 (i) holds for each n thanks to (36). Exactly as in the statement of Corollary 1 we deduce that for each p 1 " sup E
exp
n0
ˇ ˇ sup ˇYtn ˇ
!!p
Z C
0t T
p=2 #
T
0
.Zsn /T
d hM is Zsn
C d hN is n
< C1:
Hence the sequences in n of random variables exp
ˇ ˇ sup ˇYtn ˇ
!!p
Z
T
and
0t T
0
p=2 .Zsn /T d hM is Zsn C d hN n is
are uniformly integrable for all p 1. By the Vitali convergence theorem, it is thus sufficient to prove that ˇ ˇ sup ˇYtn Yt0 ˇ C
0t T
Z
T 0
.Zsn Zs0 /T d hM is .Zsn Zs0 / C d hN n N 0 is ! 0
in probability when n tends to infinity.
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
129
We split the proof into four steps. The first two steps construct one-sided estimates for the difference of Y n and Y 0 proceeding very similarly to the proof of the comparison result. In the third step we combine the aforementioned estimates to show that Y n Y 0 converges to zero uniformly on Œ0; T in probability, i.e. in ucp. Finally, we use this result to show the required convergence of the martingale parts. Step 1. First fix 2 .0; 1/ and n 1 and proceed in the same way as in the proof of Theorem 6 by defining the same objects U , V , W , , R, F , , P , G, and H , subject to the following modification. All the objects X 0 , X 2 fY; Z; N; F g, with a prime 0 are replaced by the respective object X 0 with a superscript 0. All the objects X 2 fY; Z; N; F; ˛g without a prime are replaced by the respective object X n with a superscript n, e.g. U WD Y n Y 0 . We observe that the above objects depend on n however we omit this dependence for brevity. In addition set n F .s/ D .F n F 0 /.s; Ys0 ; Zs0 /. From (29) and (31), G P e R .1 / ˛ n C 2ˇjY 0 j C n F
j n F j j n F j 2ˇKA n 0 D PJ n C e 2ˇKA P ; P ˛ C 2ˇjY j C e 1
1
where, consistent with our modification, J n WD e 2ˇKA ˛ n C 2ˇjY 0 j 0: Observe that in contrast to the proof of the comparison theorem, the difference n F of the drivers cannot be bounded above by zero. Considering
Z Dtn WD exp
t 0
Jsn dAs
and
e nt WD Dtn Pt P
and applying partial integration yields n n e nt j F j dAt D Dtn dPt C Pt dDtn C e 2ˇKA P e nt j F j dAt e nt C e 2ˇKA P dP 1
1
n T j F j n n 2ˇKA Rt Gt dAt dHt C Pt e Vt dMt Cd Wt : D Dt Pt Jt Ce Pt 1
We again conclude that the finite variation parts in the last expression are nonnegative. We now use the stopping time argument from the proof of Theorem 6 to derive " Pt
Dtn Pt
E
DTn PT
e 2ˇKA C 1
Z
T t
ˇ ˇ
#
ˇ
:
ˇ Dsn Ps j n F .s/j dAs ˇ Ft
(38)
130
M. Mocha and N. Westray
From the boundedness of and the definition P D exp e R U we derive, for s 2 Œ0; T , "
!# e 2ˇKA 0 n jYt j C jYt j Ps sup exp DW n . / and 1
0t T
ˇ ˇ ˇ ˇ ˇ 0 ˇ e2ˇKA ˇ n e2ˇKA ˇ n 0ˇ 0ˇ nˇ ˇ _ PT exp 1 exp 1
DW n . /: Using the boundedness of , the inequalities log.x/ x, (38) and 1 Dsn DTn we then find Ytn Yt0 .1 /jYt0 j C Ytn Yt0 D .1 /jYt0 j C Ut 1
exp ˇKA Rt log.Pt / D .1 /jYt0 j C .1 /jYt0 j ˇ # " Z T ˇ 2ˇKA e 1
ˇ E DTn n . / C DTn n . / j n F .s/j dAs ˇ Ft : C ˇ 1
t (39) Step 2. With regards to the converse inequality we proceed as in the proof of [6] Proposition 7 so that finally Yt0 Ytn .1 /jYtn j ˇ # " Z T ˇ 2ˇKA e 1
ˇ E DTn n . / C DTn n . / j n F .s/j dAs ˇ Ft ; C ˇ 1
t (40) where J n and thus D n , n and n . / are as inˇ Step 1. ˇ Step 3. Let us now prove that sup0t T ˇYtn Yt0 ˇ n1 converges to zero in probability. Summing up (39) and (40) we deduce ˇ ˇ n ˇ ˇ ˇY Y 0 ˇ .1 / jY 0 j C jY n j C 1 E D n n . /ˇ Ft t t t t T ˇ " Z C e 2ˇKA E DTn n . /
T
t
ˇ # ˇ ˇ j n F .s/j dAs ˇ Ft : ˇ
We note that by the usual assumptions on the filtration and by continuity of Y n and Y 0 this holds for all t, P-a.s. Applying the Doob, Markov and Hölder inequalities, we deduce the existence of some nonnegative constants c1 , c2 , independent of , as
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
131
well as c3 . / such that, for " > 0, ! ˇ n ˇ c1 .1 / c2 .1 / n 2 1=2 0ˇ ˇ C E . / P sup Yt Yt " " " 0t T " Z 2 #1=2 T c3 . / n E C j F .s/j dAs ; " 0
(41)
where the latter inequality is due to the fact that, by our assumptions, the sequences sup0t T jYt0 j C jYtn j n1 , .DTn /n1 and . n . //n1 are bounded in all Lp .P/ spaces, p 1. In addition, for the application of Doob’s inequality, we used that j n F .s/j jF n .s; Ys0 ; Zs0 / F n .s; 0; Zs0 /j C jF n .s; 0; Zs0 /j C jF 0 .s; 0; Zs0 /j C jF 0 .s; Ys0 ; Zs0 / F 0 .s; 0; Zs0 /j 2ˇjYs0 j C ˛sn C ˛s0 C kBs Zs0 k2 ; which in turn also implies that for all p 1 the family
RT 0
j n F .s/j dAs
(42) p n1
is uniformly integrable due to Corollary 1 and (36). Here, for reasons explained in Sect. 7 we deviate from [6]. The Vitali convergence theorem and (37) imply RT p n that 0 j n F .s/j dA s ! 0 inall L .P/ spaces. Observe that . / converges in probability to exp e 2ˇKA j 0 j as n goes to infinity. This convergence is also in all Lp .P/ spaces because of the uniform integrability assumption on . n /n1 . More precisely, for all p 1, we have "
!# pe 2ˇKA ˇˇ n ˇˇ 0 C j j sup EŒ . / sup E exp 1
n1 n1 n
p
!#1=2 " !#1=2 2pe 2ˇKA ˇˇ n ˇˇ 2pe 2ˇKA ˇˇ 0 j E exp < C1: sup E exp 1
1
n1 "
From (41) we then deduce that for all 2 .0; 1/ ! ˇ n ˇ c1 .1 / 0ˇ ˇ lim sup P sup Yt Yt " " n!C1 0t T C
i 12 c2 .1 / h E exp 2e 2ˇKA j 0 j : "
We then send to 1 to conclude that lim P
n!C1
! ˇ n ˇ 0ˇ ˇ sup Yt Yt " D 0:
0t T
132
M. Mocha and N. Westray
Step 4. Let us now turn to the last assertion of the theorem. We derive from Itô’s formula that # "Z T
E
.Zsn Zs0 /T d hM is .Zsn Zs0 / C d hN n N 0 is
0
"
ˇ ˇ E . / C 2 sup ˇYtn Yt0 ˇ n
Z
T
0 2
0t T
ˇ ˇ n ˇF .s; Y n ; Z n /F 0 .s; Y 0 ; Z 0 /ˇ dAs s
0
s
s
#
s
ˇZ ˇ# ˇ T ˇ ˇ n ˇ ˇ ˇ C E sup ˇYt Yt0 ˇ ˇ d hN n is d hN 0 is ˇ ˇ ˇ 0t T 0 "
after observing that the local martingale arising therein is in fact a true martingale thanks to the present integrability assumptions, cf. Corollary 1. By (9), ˇ n ˇ ˇF .s; Y n ; Z n / F 0 .s; Y 0 ; Z 0 /ˇ ˛ n C ˛ 0 C ˇjY n j C ˇjY 0 j s
s
s
s
s
C
s
s
s
kBs Zsn k2 C kBs Zs0 k2 : 2 2
Applying Hölder’s inequality, the formula (22) and the condition (36) we recognize (the expectation of the squares of) the integrals from the right hand side above as n 0 uniformly bounded ˇ nresult0now ˇ follows2from the fact that ! and (in n). The ˇ ˇ that by Steps 1–3 sup0t T Yt Yt ! 0 in L .P/. To sum up, we conclude in particular that Z
T 0
P
.Zsn Zs0 /T d hM is .Zsn Zs0 / C d hN n N 0 is ! 0 as n ! C1; t u
which completes the proof.
Remark 5. As previously discussed the sense of convergence given here differs from that in [6] where the pointwise convergence of the drivers is assumed, namely A -a.e. for all y and z we have
lim F n .; y; z/ D F 0 .; y; z/:
n!C1
(43)
We provide an example in the next section showing that this condition is not sufficient in the present setting so that the statement of [6] Proposition 7 needs a small modification.
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
133
7 Stability Counterexample Suppose our filtration is the augmentation of the filtration generated by a one dimensional Brownian motion W so that we may set At D t and B 1. The measure A now becomes the product of P and the Lebesgue measure on Œ0; T . In this setting BSDEs take the following form d Yt D Zt d Wt F .t; Yt ; Zt / dt;
YT D :
(44)
A solution now consists of a pair .Y; Z/ such that Y has continuous paths, Z is RT RT a predictable process with 0 Zt2 dt < C1 P-a.s., 0 jF .t; Yt ; Zt /j dt < C1 P-a.s. and such that (44) holds, P-a.s. RT P Suppose our condition 0 jF n F 0 j .s; Ys0 ; Zs0 / ds ! 0 is replaced by (43), n 0 i.e. F converges to F pointwise .t; !/-almost everywhere on Œ0; T ˝, where the A -null set does not depend on .y; z/. One may ask whether this is sufficient for Theorem 7 to hold, in particular if P
sup jYtn Yt0 j ! 0:
0t T
(45)
We now present an example to show that this is in fact not the case. The example resembles the standard counterexample to the dominated convergence theorem and shows that such a stability statement (under the present assumptions) already fails to hold in an essentially deterministic situation. Consider T > 1 together with parameters F 0 ˛ 0 0 0. Then all the assumptions in [6] and in the present paper are satisfied and the unique solution to the BSDE (44) with parameters .F 0 ; 0 / is given by .Y 0 ; Z 0 / 0, up to appropriate null sets. Furthermore, for integers n 1, define the terminal values n 0 and drivers F n ˛ n n 1Œ0; 1 ˝ 0: n
n
Observe that F does not depend on y or on z and is constant in !, hence determinRT istic. In particular j˛ n j1 D 0 ˛sn ds D 1, independently of ! and n, which shows that again all the assumptions in [6] and in the present paper are satisfied by each pair .F n ; n /. The unique solution to the BSDE (44) with parameters .F n ; n / is given P-a.s. by Z n 0, more precisely the zero element in L2 .Œ0; T ˝/, and Ytn D .1 nt/ 1Œ0; 1 ˝ .t; /: n
We deduce that Y n is nonnegative, nonincreasing and that Y0n D 1, independent of n, P-a.s. It now follows that, P-a.s. for all n 1, sup0t T jYtn Yt0 j D Y0n D 1, from which
134
M. Mocha and N. Westray
! lim
n!C1
sup
0t T
jYtn
Yt0 j
D1
P-a:s:
(46)
However, by construction, limn!C1 F n D 0 D F on .0; T ˝, hence A -a.e. independently of y and z, so that (43) holds. Since (45) and (46) cannot hold simultaneously, the condition in (43) is not sufficient for a stability theorem to hold under the present assumptions. We remark that this phenomenon is not dependent on the non differentiability of the paths of the Y n as one can choose F n to be arbitrarily smooth in t. Indeed, independent of !, take a smooth nonnegative function on Œ0; T that is identically zero on . n1 ; T and integrates to one over Œ0; T . The corresponding Y n in the BSDE solution are smooth in t and lead to the same contradiction. The problem arising in the proof of [6] Proposition 7 can be observed from (41) and (42). More specifically, the authors require L2 .P/-convergence of the random RT variables 0 j n F .s/j ds however they only dispose of an estimate on the product space Œ0; T ˝ of the form j n F j 2ˇjY 0 j C ˛ n C ˛ 0 C kZ 0 k2 ; together with uniform integrability assumptions that are on the level of ˝, with the t-component integrated away. There is no guarantee that the pointwise convergence of j n F j on the product space Œ0; T ˝ will transform to pointwise convergence RT of the integrals 0 j n F .s/j ds on ˝, which is necessary to utilize the uniform integrability assumptions. This is the insight behind the present example and motivates the modified condition. We now move on to look at whether the martingale part of our BSDE solution determines a change of measure.
8 Change of Measure In this section we show that under the exponential moments assumption the martingale part of a solution .Y; Z; N / to the BSDE (4) defines a measure change. In particular, we do not need to show that Z M C N is a BMO martingale, which is a stronger statement that may indeed not hold, see [14] for some examples and related discussion. Here we do not require that the driver F be convex in z. Our proof is based upon Kazamaki [19] Lemmas 1.6 and 1.7 which we state here for martingales on compact time intervals. f is a martingale on Œ0; T such that Lemma 1 ([19] Lemmas 1.6 and 1.7). If M
f C 1 hf sup M i E exp M < C1; (47) 2 stopping time valued in Œ0;T
f is a true martingale on Œ0; T . Moreover, if for a real number ¤ 1, then E M condition (47) holds for some > 1 then it holds for all 2 .1; /. We deduce the following result.
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
135
Theorem 8. Let Assumption 1 (iii) hold, j˛j1 have all exponential moments, q be a real number with jqj > =2 and .Y; Z; N / a solution to the BSDE (4) such that g is bounded by =2, Y 2 E and Z M C N is a martingale. If ˇ > 0 we also require that Y j˛j1 has of all orders or that (7) holds with exponential moments fixed y2 D 0. Then E q .Z M C N / is a true martingale on Œ0; T . In particular, when < 2, E .Z M C N / is a true martingale. Remark 6. In [22] Proposition 7 the authors show that the martingale part of solutions to the BSDE (3) with bounded first component and M a BMO martingale also belongs to the class of BMO martingales so that it yields a measure change. Our theorem may thus be seen as a generalization of this result to the case in which Y is not necessarily bounded. We mention that it follows from the proof of this theorem that the assumption of all exponential moments may be weakened to requiring exponential moments of some specific order. f WD q.Z Proof. We apply Lemma 1 with M Q M C N / for some fixed jqj Q > =2. First, we assume that ˇ > 0 and that Y j˛j1 has exponential moments of all orders. Considering
2 1 log G .t/ WD q hZ M C N it Q .Z M /t C Nt C qQ 2 for > 0 we get from the BSDE (4) and the growth condition in (8),
Z t Z t Q Yt Y0 C F .s; Ys ; Zs / dAs C gs d hN is log G .t/ D q 0
C qQ 2
0
1 hZ M C N it 2
jqj.Y Q C jY0 j/ C jqjj˛j Q Q Y j˛j1 1 C jqjˇ
jqj Q 1 C hZ M C N it : C jqj Q 2 2
Noting that jqj Q C 2
1 jqj Q 0 ” DW q0 ; 2 2jqj Q
we have that P-a.s. for all t 2 Œ0; T , G .t/ exp jqj Q Y C jY0 j exp jqjj˛j Q Q Y j˛j1 ; 1 C jqjˇ
(48)
for all q0 . By the exponential moments assumption on Y , j˛j1 and Y j˛j1 , we conclude from Hölder’s inequality that sup stopping time valued in Œ0;T
E G . / < C1
(49)
136
M. Mocha and N. Westray
for all q0 > 1=2. It now follows from Lemma 1 that E .q.Z Q M C N // is a true martingale for all 2 Œq0 ; 1/nf1g. The second part of this lemma ensures that in fact E .q.Z Q M C N // is a true martingale for all > 1. Thus, if jqj > =2 we apply this result for some fixed jqj Q 2 .=2; jqj/ and WD q=qQ D jq=qj Q > 1 to conclude that indeed E .q.Z M C N // is a true martingale. Now if ˇ > 0 and (7) holds with fixed y2 D 0, we use (9) to derive, similarly to the above, log G .t/ jqj.Y Q C jY0 j/ C jqjj˛j Q Q Y AT 1 C jqjˇ
jqj Q 1 C hZ M C N it C jqj Q 2 2
so that the claim follows from the boundedness of AT using exactly the same arguments. The reasoning from above also applies when Assumption 1 (iii) holds with ˇ D 0, without any further conditions. t u
9 Applications In the final section we explore two applications of our results, specifically focusing on utility maximization and partial equilibrium. Our contribution is to show that the standard results continue to hold when the usual boundedness assumptions are replaced by appropriate exponential moments conditions, allowing for more generality.
9.1 Constrained Utility Maximization Under Exponential Moments In the context of the constrained utility maximization problem with power utility the following BSDE appears, cf. [25] Sect. 4.2.1, d Yt D ZtT dMt C dNt F .t; Zt / dAt
1 d hN it ; 2
YT D 0;
where the generator is given by
2 2 p.p 1/ z t z t p.1 p/ F .t; z/ D inf Bt C Bt 1 p 2C 2 1p 2 1 C kBt zk2 : 2
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
137
In the above 1 p 2 .0; C1/ is the investor’s relative risk aversion and refers to investment strategies in a stock whose returns are driven by the continuous local martingale M under the market price of risk process and must be valued in the closed constraint set C . Writing the infimum in terms of the distance function, which is Lipschitz continuous, one can show that the driver F satisfies Assumption 1 (ii)–(v), so that there exist constants c and cz such that jF .t; z/j c kBt t k2 C cz kBt zk2 : RT When we enforce that the mean-variance trade off 0 Tt hM it t has all exponential moments, an assumption weaker than that of boundedness given in the cited literature, we are in the current framework and see that the BSDE admits a unique solution in E M2 M 2 . The crucial step in [16] and [25] is, given a solution triple .Y; Z; N /, to construct the relevant optimizers; this is the process of verification. Using the theorems of the present paper, it is possible to show that one can repeat the reasoning of [16] and [25] and that similar results continue to hold for more general classes of market price of risk processes under appropriate trading constraints such as bounded short-selling and borrowing. A biproduct of the analysis described here is a direct link between solutions to the utility maximization problem and solutions to the BSDE in an exponential moments setting, building on [26]. This allows for a detailed study of the stability of the utility maximization problem, undertaken in [24], which is important in many applications.
9.2 Partial Equilibrium and Market Completion Under Exponential Moments We now briefly describe the partial equilibrium framework of [15] in which structured securities that are written on nontradable assets are priced via a market clearing condition. The agents in this economy have preferences which are given by dynamic convex risk measures. The risk they are exposed to is given by two sources. The first is encoded in a financial market in which frictionless trading in a stock S is possible. The second is a non-financial risk factor R that can only be dealt with via a derivative written on this external factor. It is assumed that this derivative completes the market, in fact it is shown that in equilibrium the market is complete. More specifically, while the market price S of financial risk is given exogenously the market price R of external risk is determined via an equilibrium condition. This states that when the derivative is priced according to the pricing rule arising from .S ; R / the agents’ aggregated demand matches the fixed supply. The demand is in this setting given by the solutions to the agents’ individual risk minimization problems and is a function of R .
138
M. Mocha and N. Westray
To ease the exposition we put ourselves in a representative agent setting where the agent’s preferences are of entropic type, i.e. their utility function is exponential. Then the following BSDE for the agent’s dynamic risk Y appears d Yt D ZtT d Wt
1 S 2 .t / 2St Zt1 .Zt2 /2 dt; 2
YT D H;
where W is a two dimensional Brownian motion representing the two sources of risk, Z D .Z 1 ; Z 2 / is the corresponding control process to Y and H is the agent’s endowment. Under suitable exponential moments assumptions the present article O to the above BSDE. Once we provides the existence of a unique solution .YO ; Z/ 2 O check that Z defines a valid pricing rule, i.e. that E .S ; ZO 2 / W is a true martingale, we know that the equilibrium market price R of external risk is given by R WD ZO 2 . To conclude we can generalize [15], full details will appear elsewhere.
References 1. P. Barrieu, N. El Karoui, Monotone stability of quadratic semimartingales with applications to general quadratic BSDEs and unbounded existence result. arXiv:1101.5282v2, 2012 2. D. Becherer, Bounded solutions to backward SDE’s with jumps for utility optimization and indifference hedging. Ann. Appl. Probab. 16(4), 2027–2054 (2006) 3. J.-M. Bismut, Conjugate convex functions in optimal stochastic control. J. Math. Anal. Appl. 44, 384–404 (1973) 4. G. Bordigoni, A. Matoussi, M. Schweizer, A stochastic control approach to a robust utility maximization problem. In Stochastic Analysis and Applications, vol. 2 of Abel Symposium (Springer, Berlin, 2007), pp. 125–151 5. P. Briand, Y. Hu, BSDE with quadratic growth and unbounded terminal value. Probab. Theory Relat. Fields 136(4), 604–618 (2006) 6. P. Briand, Y. Hu, Quadratic BSDEs with convex generators and unbounded terminal conditions. Probab. Theory Relat. Fields 141(3–4), 543–567 (2008) 7. F. Da Lio, O. Ley, Uniqueness results for second-order Bellman-Isaacs equations under quadratic growth assumptions and applications. SIAM J. Control Optim. 45(1), 74–106 (electronic) (2006) 8. F. Delbaen, Y. Hu, X. Bao, Backward SDEs with superquadratic growth. Probab. Theory Relat. Fields (2010) DOI 10.1007/s00440-010-0271-1 9. F. Delbaen, Y. Hu, A. Richou, On the uniqueness of solutions to quadratic BSDEs with convex generators and unbounded terminal conditions. Ann. Inst. Henri Poincaré Probab. Stat. 47(2), 559–574, (2011) 10. N. El Karoui, S. Hamadène, A. Matoussi, Backward stochastic differential equations and applications. In R. Carmona (ed.), Indifference Pricing: Theory and Applications, Princeton Series in Financial Engineering (Princeton University Press, Princeton, NJ, 2009), pp. 267–320 11. N. El Karoui, S.-J. Huang, A general result of existence and uniqueness of backward stochastic differential equations. In Backward Stochastic Differential Equations (Paris, 1995–1996), vol. 364 of Pitman Res. Notes Math. Ser. (Longman, Harlow, 1997), pp. 27–36 12. N. El Karoui, S. Peng, M.C. Quenez, Backward stochastic differential equations in finance. Math. Finance 7(1), 1–71 (1997) 13. C. Frei, Convergence results for the indifference value based on the stability of BSDEs, Working Paper (2011)
Quadratic Semimartingale BSDEs Under an Exponential Moments Condition
139
14. C. Frei, M. Mocha, N. Westray, BSDEs in utility maximization with BMO price of risk, arXiv:1107.0183 (2011) 15. U. Horst, T. Pirvu, G. dos Reis, On securitization, market completion and equilibrium risk transfer. Math. Finance Econ. 2(4), 211–252 (2010) 16. Y. Hu, P. Imkeller, M. Müller, Utility maximization in incomplete markets. Ann. Appl. Probab. 15(3), 1691–1712 (2005) 17. Y. Hu, M. Schweizer, Some new BSDE results for an infinite-horizon stochastic control problem. In G.D. Nunno, B. Øksendal (eds), Advanced Mathematical Methods for Finance (Springer, Berlin, 2011), pp. 367–395 18. J. Jacod, A.N. Shiryaev, Limit Theorems for Stochastic Processes, vol. 288 of Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences], 2nd edn. (Springer, Berlin, 2003) 19. N. Kazamaki, Continuous Exponential Martingales and BMO. Lecture Notes in Mathematics, vol. 1579 (Springer, Berlin, Heidelberg, 1994) 20. M. Kobylanski, Backward stochastic differential equations and partial differential equations with quadratic growth. Ann. Probab. 28(2), 558–602 (2000) 21. J.P. Lepeltier, J. San Martín, Backward stochastic differential equations with continuous coefficient. Statist. Probab. Lett. 32(4), 425–430 (1997) 22. M. Mania, M. Schweizer, Dynamic exponential utility indifference valuation. Ann. Appl. Probab. 15(3), 2113–2143 (2005) 23. M. Mania, R. Tevzadze, Backward stochastic partial differential equations related to utility maximization and hedging. J. Math. Sci. 153, 291–380 (2008) 24. M. Mocha, N. Westray, The stability of the constrained utility maximization problem – a BSDE approach arXiv:1107.0190 (2011) 25. M.-A. Morlais, Quadratic BSDEs driven by a continuous martingale and applications to the utility maximization problem. Finance Stoch. 13(1), 121–150 (2009) 26. M. Nutz, The Bellman equation for power utility maximization with semimartingales. Ann. Appl. Probab. arXiv:0912.1883 (2012 in press) 27. É. Pardoux, S.G. Peng, Adapted solution of a backward stochastic differential equation. Syst. Control Lett. 14(1), 55–61 (1990) 28. P.E. Protter, Stochastic Integration and Differential Equations, vol. 21 of Applications of Mathematics (New York), 2nd edn. Stochastic Modelling and Applied Probability (Springer, Berlin, 2004) 29. R. Rouge, N. El Karoui, Pricing via utility maximization and entropy. Math. Finance 10(2), 259–276 (2000) INFORMS Applied Probability Conference (Ulm, 1999) 30. M. Schroder, C. Skiadas, Optimal consumption and portfolio selection with stochastic differential utility. J. Econ. Theory 89(1), 68–126 (1999) 31. J. Sekine, On exponential hedging and related quadratic backward stochastic differential equations. Appl. Math. Optim. 54, 131–158 (2006) 32. C. Skiadas, Robust control and recursive utility. Finance Stoch. 7(4), 475–489 (2003) 33. R. Tevzadze, Solvability of backward stochastic differential equations with quadratic growth. Stochast. Process. Appl. 118(3), 503–515 (2008)
The Derivative of the Intersection Local Time of Brownian Motion Through Wiener Chaos Greg Markowsky
Abstract Rosen (Séminaire de Probabilités XXXVIII, 2005) proved the existence of a process known as the derivative of the intersection local time of Brownian motion in one dimension. The purpose of this paper is to use the methods developed in Nualart and Vives (Publicacions Matematiques 36(2):827–836, 1992) in order to give a simple new proof of the existence of this process. Some related theorems and conjectures are discussed.
1 Introduction The study of the derivative of the intersection local time of Brownian motion essentially began with the work of Rogers and Walsh in [8–10]. There, the following functional was studied Z t A.t; Bt / D 1Œ0;1/ .Bt Bs /ds: 0
A formal application of It¯o’s Lemma to A.t; Bt / gives rise to the following expression: Z 1Z t ı00 .Bt Bs /dsdt: (1) 0
0
x2 2"
Let p" .x/ D ."/d=21p2 e be the Gaussian density with mean 0 and variance " in any dimension d . In [11] Rosen proved the following, among other things.
G. Markowsky () Monash University, VIC 3800, Australia e-mail:
[email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__6, © Springer-Verlag Berlin Heidelberg 2012
141
142
G. Markowsky
Theorem 1. Let Bt be 1-dimensional Brownian motion. Then ˛"0 D
Z
1 0
Z
t 0
p"0 .Bt Bs /dsdt
converges in L2 to a random variable ˛ 0 as " ! 0. Note that "0 converges weakly to ı00 , so ˛ 0 can be interpreted as a realization of the formal expression (1). ˛ 0 has attracted some attention in the time since Rosen’s paper. In [4], a Tanaka formula for ˛ 0 was proved, and in the process another proof of Theorem 1 was obtained. Yet another proof of Theorem 1, together with an interesting stochastic integral representation of ˛ 0 , was given in [3]. In [15], a study was begun into A and ˛ 0 driven by fractional Brownian motion. In a different research thread, in [7] the Wiener chaos expansion was used in the study of various local times of Brownian motion. In particular, a new proof of the following well-known theorem, known as Varadhan’s renormalization (see [13]), was given. Theorem 2. Let Bt be 2-dimensional Brownian motion. Then Z L" D .p" .Bt Bs / EŒp" .Bt Bs //dsdt 0<s
converges in L2 to a random variable as " ! 0. The primary purpose of this note is to show how the technique developed in [7] can be used to give a simple proof of Theorem 1. The next section is devoted to this, and gives the explicit Wiener chaos expansion of ˛ 0 as well. The final section is devoted to the discussion of related questions involving higher derivatives and dimensions.
2 The Wiener Chaos Expansion of ˛0 For more information on the Wiener chaos expansion, see [6] or [14]. Given symmetric f 2 L2 .Œ0; 1/n / let Z 1 Z tn Z t2 In .f / D nŠ ::: f .t1 ; t2 ; : : : ; tn /d Wt1 : : : d Wtn 0
0
0
˝n
and for f 2 L .Œ0; 1// let In .f / denote the multiple stochastic integral In applied to the tensor product f ˝n D f .t1 /f .t2 / : : : f .tn /. Following [7], we can deduce the following Wiener chaos expansion for ˛ 0 , in the process obtaining another proof of Theorem 1. 2
The Derivative of the Intersection Local Time of Brownian Motion
143
Theorem 3. 1 X
˛0 D
cm I2m1 1 C
mD1
where
1 1 1 ; m3=2 m3=2 m3=2 .v v / .1 v / .v /
.1/m cm D p 2 .m 1/Š 2m1 .m 1=2/ .m 3=2/
and
v D v1 _ : : : _ v2m1 v D v1 ^ : : : ^ v2m1 :
Proof (Theorems 1 and 3). By Stroock’s formula, given in [12] and [6, p. 35], p"0 .Wt /
1 X 1 EŒp".nC1/ .Wt /In .1Œ0;t ˝n /: D nŠ nD0
(2)
The n-th (probabilists’) Hermite polynomial is given by Hn .x/ D .1/n e x
2 =2
d n x 2 =2 .e /: dx n
As a consequence of this, and as is shown in [7], x p".n/ .x/ D .1/n "n=2 p" .x/Hn p : "
(3)
Plugging (3) into (2) gives p"0 .Wt / D
W .1/n t EŒp .W /H p In .1Œ0;t ˝n /: " t nC1 .nC1/=2 nŠ " " nD0
1 X
(4)
The expectation in the last expression of (4) is calculated in [7]. It is equal to 0 when n is even, and equal to ."/m .2m/Š p 2 2m mŠ .t C "/mC1=2 when n D 2m 1 is odd. We arrive at p"0 .Wt / This leads to
D
1 X
.1/mC1 p I2m1 .1Œ0;t ˝2m1 /: m1 .t C "/mC1=2 2 .m 1/Š 2 mD1
144
G. Markowsky
˛"0 D
1 X
.1/mC1 p 2 .m 1/Š 2m1 mD1
Z
1
Z
0
t 0
I2m1 .1.s;t ˝2m1 / dsdt: .t s C "/mC1=2
(5)
We want to let " ! 0 in this expression. In order to do so, we will apply the following lemma from [7], which is a consequence of the dominated convergence theorem. Lemma 1. Let F" P be a family of square integrable random variables with chaos 1 " " 2 expansions F" D nD0 In .fn /. If fn converges in L to some function fn as " ! 0 for each n, and if 1 X nD0
supfnŠjjfn" jj22 g < 1;
P1
then F" converges in L2 to F D
(6)
"
nD0 In .fn /
as " ! 0.
We must therefore show that 1 X nD0
" supf.2m 1/Šjjf2m1 jj22 g < 1;
(7)
"
" " determined by (5). For fixed m and ", .2m 1/Šjjf2m1 jj22 is given by with f2m1
.2m 1/Š 2..m 1/Š/2 22m2 D
Z
.2m 1/Š 2..m 1/Š/2 22m2
s
Z
h1.s;t ; 1.u;v i2m1 dsdtdudv .t s C "/mC1=2 .v u C "/mC1=2
s
j.s; t \ .u; vj2m1 dsdtdudv: .t s C "/mC1=2 .v u C "/mC1=2
(8)
Using Stirling’s approximation, p .2m 1/Š D O. m/ 2 2m2 2..m 1/Š/ 2
(9)
To estimate the integral, we break it into two regions as Z 2 0
.v s/2m1 dsdtdudv ." C t s/mC1=2 ." C v u/mC1=2 Z .t s/m3=2 C2 dsdtdudv: (10) mC1=2 0
The Derivative of the Intersection Local Time of Brownian Motion
145
We set " D 0 in order to obtain an upper bound for this expression. To bound the second integral for large m, perform the linear transformation .u; a; b; c/ D .u; su; t s; v t/ to get Z
Z
1 0
1u 0
Z Z
1ua 0 1
Z
0
1 0
Z
1uab 0
Z
1 0
b m3=2 dcdbdad u .a C b C c/mC1=2
(11)
b m3=2 dcdadb .a C b C c/mC1=2
k .m 1=2/.m 3=2/
with k a constant. To obtain the final inequality in (11) we twice employed the trivial bound Z 1 1 1 dx (12) n n1 .x C h/ .n 1/h 0 valid for n > 1. To estimate the first integral in (10), substitute .u; a; b; c/ D .u; su; v s; t v/ to get Z
1
0
Z
1u
0
Z
1ua 0
Z
1uab
b 2m1 (13) .a C b C c/mC1=2 .a C b/mC1=2 0 Z 1Z 1Z 1 b 2m1 dcdadb mC1=2 .a C b/mC1=2 0 0 0 .a C b C c/
k .m 1=2/.2m 1/
for large m, again with k a fixed constant. Combining (11) and (13) we see that (10) " is O.m2 /. Taking into account (9), we see that .2m 1/Šjjf2m1 jj22 is O.m3=2 /. This is summable, so we can indeed apply Lemma 1 to let " ! 0 in (5). This shows that ˛"0 converges in L2 , and to arrive at the form for the Wiener chaos decomposition given in the statement of the theorem, note that
0
˛ D
1 X
Z
.1/mC1
p 2 .m 1/Š 2m1 mD1 D
1 X
1 0
.1/mC1
Z
t 0
I2m1 .1.s;t ˝2m1 / dsdt .t s/mC1=2
p I2m1 2 .m 1/Š 2m1 mD1
Z
1 0
Z
t 0
1.s;t ˝2m1 dsdt : (14) .t s/mC1=2
Now, for 0 < s < t < 1, 1.s;t ˝2m1 .v1 ; : : : ; v2m1 / D 1.v1 _:::_v2m1 ;1 .t/1Œ0;v1 ^:::^v2m1 .s/:
146
G. Markowsky
Using the notation given in (2), we therefore have Z
1 0
Z
t 0
1.s;t ˝2m1 dsdt .t s/mC1=2 Z 1 Z v D 1.v;1 .t/ 0
Z
0
1 dsdt .t s/mC1=2
1 1 m1=2 dt m1=2 .t v / .t/ v 1 1 1 1 D : 1C .m 1=2/.m 3=2/ .v v /m3=2 .1 v /m3=2 .v/m3=2 D
1 m 1=2
1
Plugging this into (14) gives
˛0 D
1 X
.1/mC1 p 2 .m 1/Š 2m1 .m 1=2/ .m 3=2/ mD1 1 1 1 : I2m1 1 C .v v/m3=2 .1 v/m3=2 .v/m3=2
This completes the proof of Theorems 1 and 3.
t u
3 Higher Dimensions and Derivatives In [2], these methods were extended considerably in order to prove a number of results concerning fractional Brownian motion. As relates to Brownian motion, a proof was given for the following central limit theorem, which was originally proved by Yor in [16]. Theorem 4. Let Bt be 3-dimensional Brownian motion. Then Z 1 p .p" .Bt Bs / EŒp" .Bt Bs //dsdt log.1="/ 0<s
converges in law to a normal law as " ! 0.
The Derivative of the Intersection Local Time of Brownian Motion
147
" Note that ıp ıx is the partial derivative with respect to one of the space variables, x. It seems likely that the methods from [2] can be adapted to yield a proof of Theorem 5, as well as of the following conjecture.
Conjecture 1. (a) Let Bt be 1-dimensional Brownian motion. Then, for some > 0, 1 log.1="/
Z 0<s
.p"00 .Bt Bs / EŒp"00 .Bt Bs //dsdt
converges in law to a normal law as " ! 0, where p"00 is the second derivative in the space variable of p" . (b) Let Bt be a d-dimensional Brownian motion. Then, if D is a (possibly mixed) partial derivative of order n 4 d , Z "
.Dp" .Bt Bs / EŒDp" .Bt Bs //dsdt
.nCd /=23=2 0<s
converges in law to a normal law as " ! 0. We should mention that part (b) of the conjecture has already been proved for n D 0 in [1] and [2]. The remaining cases of the conjecture come from examining the scaling properties of the integrals which result from expanding in Wiener chaos. Unfortunately, however, the details seem to become a bit more onerous with the presence of the derivatives, and handling the relevant integrals which result would seem to require some new techniques. Acknowledgements I would like to thank Paul Jung, David Nualart, and Jay Rosen for helpful conversations. I would also like to thank the referee for comments which improved the exposition. I am grateful for support from the Priority Research Centers Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science, and Technology (Grant #2009-0094070) and from Australian Research Council Grant DP0988483.
References 1. J.Y. Calais, M. Yor, Renormalization et convergence en loi pour certaines intégrales multiples associées au mouvement Brownien dans Rd . In Séminaire de Probabilités XXI. Lecture Notes in Mathematics, vol. 1247 (Springer, Berlin, Heidelberg, 1987) 2. Y. Hu, D. Nualart, Renormalized self-intersection local time for fractional Brownian motion. Ann. Probab. 33(3), 948–983 (2005) 3. Y. Hu, D. Nualart, Central limit theorem for the third moment in space of the Brownian local time increments. Electron. Commun. Prob. 15, 396–410 (2009) 4. G. Markowsky, Proof of a Tanaka-like formula stated by J. Rosen in Séminaire XXXVIII. Séminaire de Probabilités XLI, 2008, pp. 199–202 5. G. Markowsky, Renormalization and convergence in law for the derivative of intersection local time in R2 . Stochast. Proces. Appl. 118(9), 1552–1585 (2008) 6. D. Nualart, The Malliavin Calculus and Related Topics (Springer, Berlin, Heidelberg, 1995)
148
G. Markowsky
7. D. Nualart, J. Vives, Chaos expansions and local times. Publicacions Matematiques 36(2), 827–836 (1992) 8. L.C.G. Rogers, J.B. Walsh, A.t; Bt / is not a semimartingale. Seminar on Stochastic Process, Progr. Probab., 24, Birkhäuser Boston, Boston, MA, 1991, pp. 275–283 9. L.C.G. Rogers, J.B. Walsh, The intrinsic local time sheet of Brownian motion. Prob. Theory Relat. Fields 88(3), 363–379 (1991) 10. L.C.G. Rogers, J.B. Walsh, Local time and stochastic area integrals. Ann. Probab. 19(2), 457–482 (1991) 11. J. Rosen, Derivatives of self-intersection local times. Séminaire de Probabilités XXXVIII. Lecture Notes in Mathematics, vol. 1857, 2005, pp. 263–281 12. D. Stroock, Homogeneous chaos revisited. Séminaire de Probabilités XXI. Lecture Notes in Mathematics, vol. 1247, 1987, pp. 1–7 13. S.R.S. Varadhan, Appendix to “Euclidean quantum field theory” by K. Symanzik. Local Quant. Theory (1969) 14. N. Wiener, The homogeneous chaos. Am. J. Math. 60(4), 897–936 (1938) 15. L. Yan, X. Yang, Y. Lu, p-variation of an integral functional driven by fractional Brownian motion. Stat. Prob. Lett. 78(9), 1148–1157 (2008) 16. M. Yor, Renormalisation et convergence en loi pour les temps locaux d’intersection du mouvement Brownien dans R3 . Séminaire de Probabilités XIX. Lecture Notes in Mathematics, vol. 1123, 1985, pp. 350–365
On the Occupation Times of Brownian Excursions and Brownian Loops Hao Wu
Abstract We study properties of occupation times by Brownian excursions and Brownian loops in two-dimensional domains. This allows for instance to interpret some Gaussian fields, such as the Gaussian Free Fields as (properly normalized) fluctuations of the total occupation time of a Poisson cloud of Brownian excursions when the intensity of the cloud goes to infinity. Keywords Conformal invariance • Brownian excursion measure • Brownian loop measure • Green’s function
1 Introduction Conformal invariance of planar Brownian motion has been derived and exploited long ago by Paul Lévy [8]. See also B. Davis (Annals of Proba 1979) in particular his derivation of Picard’s big theorem. More recently, conformal invariance turned out to be an instrumental idea in the study of various critical models from statistical physics in the plane (see for instance [4, 16] and the references therein). Two basic important conformally invariant measures on random geometric objects are the Brownian excursion measure and the Brownian loop measure. Let us now very briefly describe these measures and the meaning of conformal invariance relatively to these measures. For each open domain D with non-polar boundary in the plane, one can define these two measures in D respectively denoted by D and D . These are infinite but -finite measures on Brownian-type paths with particular properties:
H. Wu () Département de Mathématiques, Université Paris-Sud, 91405 Orsay Cedex, France e-mail:
[email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__7, © Springer-Verlag Berlin Heidelberg 2012
149
150
H. Wu
• D is supported on the set of Brownian excursions .Bt ; t / in D i.e. Brownian paths such that B0 and B are in @D, while B.0; / D. • D is supported on the set of Brownian loops .Bt ; t / i.e. Brownian paths in D such that B0 D B . In fact, in both cases, it is useful to view these paths up to monotone reparametrization (in the loop-case, one views the time-set modulo i.e., there is no “starting point” on the loop). Then, it turns out (see [5, 15] for details) that for any conformal map ˚ from D onto ˚.D/, the image measures of D and D under ˚ are exactly ˚.D/ and ˚.D/ . These two measures on loops and on excursions allow in some sense to get rid of the dependence of the measure on Brownian paths with respect to its starting point, see for instance the discussion in [16]. In the present text, we shall focus on the following type of results (here and in the sequel, dx or dy will denote the area measure, and x or y will always denote points in the plane): Proposition 1. Suppose that D is a simply connected domain and that A and B are two open proper subsets of D. Then, Z
Z
D . 0
Z
and
ds1A .s / Z
ds1A .s / 0
0
dx dy GD .x; y/
(1)
AB
0
D .
Z ds1B .s // D 4 Z
ds1B .s // D
dx dy .GD .x; y//2 ;
(2)
AB
where .s ; 0 s / is a Brownian excursion in (1) and a Brownian loop in (2), GD .x; y/ denotes the usual Green’s function in D (with Dirichlet boundary conditions). The Brownian excursion measure and the loop measure are infinite measures, but they can be used to define random conformally invariant collections of excursions and loops (i.e. under a probability measure) by a Poissonization procedure. As explained in [16], both these Poissonian clouds are of interest and useful in the context of random planar conformally invariant curves of SLE-type: The “excursion clouds” give rise to the restriction measures [15], while the “loop-soups (loop clouds)" are related to Conformal Loop Ensembles (see [14]). It is natural to study the cumulative occupation time of these random collections of Brownian paths. The previous proposition can then be viewed as a description of the covariance structure of these cumulative occupation times (even if as we shall explain later, things are slightly more complicated in the case of the loop measure because cumulative occupation times are infinite, so that a renormalization procedure is needed). By the classical central limit theorem, in the asymptotic regime where the intensity of these clouds goes to infinity, the fluctuations of these occupation times converge (if properly normalized of course) to a Gaussian process with the same covariance structure. This will in particular enable us to interpret the
On the Occupation Times of Brownian Excursions and Brownian Loops
151
Gaussian Free Field in terms of fluctuations of occupation times of high-intensity clouds of Brownian excursions. Note that in [7], a different and more direct (as it involves no asymptotic) relation between the loop-soup occupation times and the Gaussian Free Field (or rather its square) is pointed out. Here is how the present paper is structured: In Sect. 2, we review various very elementary facts concerning Green’s functions, their conformal invariance and their relation to Brownian motion and the Gaussian Free Field. In Sect. 3, we recall the definition of the Brownian excursion measure, we derive (1) and deduce from it the interpretation of the Gaussian Free Field as asymptotic fluctuations of the Excursions occupation time measure. In passing, we note a representation of the solution to the standard Dirichlet problem using Brownian excursions, that does not seem so well-known despite its simplicity. Section 4 is the counterpart of Sect. 3 for Brownian loops instead of Brownian excursions. Finally, in Sect. 5, we briefly mention a generalization of the previous results using some clouds of interacting pairs of excursions (via their intersection local-time) that exhibits some relations between loops and excursions. We will focus on two-dimensional domains, but many of our statements (in particular those on Brownian excursions) are also valid in higher dimensions. However, as the reader will see, we choose to base our proofs on conformal invariance, so that another approach would be needed to derive the results in dimensions greater than two. We should also point out that the statements are in fact valid in non-simply connected domains, but again, some of our proofs, in particular those dealing with the loop-measure, would need to be changed in order to cover non-simply connected planar domains (as we will use explicit expressions for the unit disc).
2 Review of Basic Notions 2.1 Generalities We first recall some classical facts about Brownian motion and its relation to harmonic functions, see for instance [1, 10, 11] for further details or background. Suppose that D is a bounded planar domain, and that it has a smooth boundary. Then, for any point x in D, the distribution of the exit position from D by a Brownian motion started at x has a continuous density with respect to the surface measure .d z/ on @D, called the Poisson kernel, that we will denote by hD .x; z/ for z 2 @D. In other words, the exit distribution is hD .x; z/.d z/. This Poisson kernel is closely related to the solutions of the Dirichlet problem in D (i.e., to find a harmonic function u in D, that is continuous on D and equal to some prescribed continuous function f on the boundary of D). Indeed, the solution to the Dirichlet problem, if it exists, is given by
152
H. Wu
Z u.x/ D
.d z/hD .x; z/f .z/ D Ex .f .Z //
where Z is a planar Brownian motion started from x under the probability measure Px and denotes its exit time from D. The Green’s function in D, is the unique function in D D, such that for each x 2 D, y 7! GD .x; y/ is harmonic, vanishes on @D, and satisfies GD .x; y/ 1 log.1=jx yj/ when y ! x. Alternatively, one can think of GD .x; y/dy as the expected time spent by Z in the infinitesimal neighborhood of y before exiting D. More precisely, if A denotes an open set, the expected time spent by the Brownian motion Z (started from Z0 D x) in A before exiting D is Z
Z
dt1A .Zt // D
Ex .
dyGD .x; y/: A
0
The Green’s function is closely related to the Poisson problem (i.e. to find a C 2 function u in D such that u D 2g, where g is some given continuous function in D, with the property that u is continuous on D and equal to 0 on @D). Under mild assumptions on D, the solution to this problem exists, is unique, and Z
Z
u.x/ D
dyGD .x; y/g.y/ D Ex . D
dtf .Zt //: 0
Not surprisingly, the Poisson kernel is closely related to the Green’s function. More precisely, if n D nz;D is the inwards pointing normal vector at z 2 @D, then, as goes to 0, GD .x; z C n/ 2 hD .x; z/: In the case of the unit disc U WD fx W jxj < 1g in the complex plane, the Poisson kernel and the Green’s function can be explicitly computed: hU .x; z/ D and GU .x; y/ D
1 jxj2 2jx zj2
jx yj 1 log j1 x yj N
for x 2 U; y 2 U, and z 2 @U.
2.2 Conformal Invariance Conformal invariance of planar Brownian motion, first observed by Paul Lévy [8], can be described as follows: if one considers a planar Brownian motion Z started
On the Occupation Times of Brownian Excursions and Brownian Loops
153
from x and stopped at its first exit time of a simply connected domain D, and if ˚ denotes a conformal map from D onto some other domain D 0 , then the law of ˚.Z/ is that of a Brownian motion started from ˚.x/ and stopped at its first exit time of D 0 . Actually, for this statement to be fully true, one has to reparametrize time of ˚.Z/ in a proper way. The rigorous statement is that for all t < , 0 ˚.Zt / D ZH with Ht D t
Z
t
dsj˚ 0 .Zs /j2 ;
0
where Z 0 is a Brownian motion started from ˚.x/, stopped at 0 D H , which is its exit time of D 0 . Conformal invariance of Brownian motion is closely related to the conformal invariance of the Green’s function and of the Poisson kernel. Let us give a rather convoluted explanation of the conformal invariance of Green’s functions using Brownian motion (a direct proof using the analytic characterization of the Green’s function is much more straightforward) that will be helpful for what follows. Suppose that x and y are in D and that is very small. We have seen that the expected time spent in the ball U.y; /, centered at y and of radius , by the Brownian motion Z started at x behaves like 2 GD .x; y/ when ! 0. Equivalently, the expected time spent in the ball U.˚.y/; j˚ 0 .y/j / by the Brownian motion ˇ started at ˚.x/, behaves like j˚ 0 .y/j2 2 GD 0 .˚.x/; ˚.y// as ! 0. The process ˚.Z/ can be viewed as a time-changed Brownian motion, and the time-change when Z is close to y is described via Ht . It follows easily that this expected time of ˚.Z/ spent in the ball U.˚.y/; j˚ 0 .y/j / behaves like j˚ 0 .y/j2 2 GD 0 .˚.x/; ˚.y// D 2 GD 0 .˚.x/; ˚.y//: j˚ 0 .y/j2 As a result, we have indeed that G˚.D/ .˚.x/; ˚.y// D GD .x; y/:
(3)
For a more rigorous derivation along the same lines, we can use the integral representation of occupation times of domains: on the one hand, Z
D 0
E˚.x/ . 0
dtf .Zt0 //
Z D Z
D0
dyGD 0 .˚.x/; y/f .y/ j˚ 0 .y/j2 dyGD 0 .˚.x/; ˚.y//f .˚.y//
D D
154
H. Wu
for indicator functions f D 1A , and on the other hand, Z
D 0
E˚.x/ . 0
dtf .Zt0 // D Ex . Z
Z
D
j˚ 0 .Zt /j2 dtf .˚.Zt ///
0
dyGD .x; y/f .˚.y//j˚ 0 .y/j2 :
D D
Conformal invariance of planar Brownian motion can also be used in a similar way to see that j˚ 0 .z/j h˚.D/ .˚.x/; ˚.z// D hD .x; z/ (4) for all x 2 D; z 2 @D when @D is smooth. Let us stress again that these conformal invariance properties of the Green’s functions and of the Poisson kernel can be derived much more directly without any reference to Brownian paths. Note that GU .0; y0 / D 1 log jy0 j for all y0 ¤ 0. The formula for GU .x; y/ then follows immediately, using the Möbius transformation x of U onto itself that maps x onto 0 and vice-versa (this is the map z 7! .z x/=.1 xz/) N because then GU .x; y/ D GU .0; x .y//. Note also that this conformal invariance also provides one possible explanation of the symmetry of the Green’s function GU .x; y/ D GU .y; x/ (because for any x and y, there exists a conformal map from D onto itself that maps x onto y and y onto x). Similarly, since clearly hU .0; z/ D 1=.2/ for all z 2 @U, the formula for hU .x; z/ recalled at the end of the previous subsection follows using conformal invariance.
2.3 The Gaussian Free Field In the present text, we will briefly relate our Brownian excursions to the Gaussian Free Field, which is a classical and basic building block in Field theory, see for instance [2, 9]. So we recall its definition, in the Gaussian Hilbert space framework (as in [12] for instance): Consider the space Hs .D/ of smooth, real-valued functions on R2 that are supported on a compact subset of a domain D Rd (so that, in particular, their first derivatives are in L2 .D/). This space can be endowed with the Dirichlet inner product defined by Z .f1 ; f2 /r D
dx.rf1 rf2 / D
It is immediate to see that this Dirichlet inner product is invariant under conformal transformation. Denote by H.D/ the Hilbert space completion of Hs .D/. The quantity .f; f /r is called the Dirichlet energy of f . A Gaussian Free Field is any Gaussian Hilbert space G .D/ of random variables denoted by “.h; f /r ”—one variable for each f 2 H.D/—that inherits the Dirichlet inner product structure of H.D/, i.e.,
On the Occupation Times of Brownian Excursions and Brownian Loops
155
EŒ.h; a/r .h; b/r D .a; b/r : In other words, the map from f to the random variable .h; f /r is an inner product preserving map from H.D/ to G .D/. The reason for this notation is that it is possible to view h as a random linear operator, but we will not need this approach. We also view .h; / as being well defined for all 2 .4/H.D/ (if D 4f for some f 2 H.D/, then we denote .h; / D .h; f /r ). When 1 and 2 are in Hs .D/, the covariance of .h; 1 / and .h; 2 / can be written as .41 1 ; 41 2 /r D . 1 ; 1 2 / D .1 1 ; 2 /. From the Poisson problem that we discussed before, 1 can be written using the Green’s function as Z 1 1 Œ .x/ D dy GD .x; y/ .y/; 2 D we may also write: CovŒ.h; 1 /; .h; 2 / D
1 2
Z dxdy GD .x; y/ 1 .x/ 2 .y/
(5)
Both the Dirichlet inner product and the Gaussian Free Field inherit naturally conformal invariance properties from the conformal invariance of the Green’s function. The 2-dimensional Gaussian free field (GFF) is a particular rich object, in which a number of geometric features can be detected, and that has been shown to play a central role in the theory of random surfaces and conformally invariant geometric structures, see [13] and the references therein.
3 Occupation Times of Brownian Excursions Brownian excursion measure. Let us first very briefly recall the construction of Brownian excursion measures. For the unit disc U, for each > 0, let denote the measure of total mass 1= defined as 1= times the law of a Brownian motion started uniformly on the circle of radius .1 /, and stopped at its first hitting time of the unit circle. In some appropriate topology, the measures converge when ! 0 to an infinite measure on two-dimensional paths that start and end on the unit circle. For a general simply connected domain D, the excursion measure D can either be defined as the image of by the conformal map ˚ that maps U onto D, or alternatively in an analogous way as in the disc, by integrating over the choice of the starting point of the excursion on @D. The fact that these two definitions are equivalent is the conformal invariance property of the Brownian excursion measures. See e.g. [16] for details and references. Note that is a measure on paths .Bt ; 0 < t < / that start and end on @D (i.e., B0 2 @D and B 2 @D) that are “oriented”, i.e. B0 and B do a priori not play the same role. However, it turns out that the Brownian excursions are reversible i.e., that .Bt ; 0 < t < / and .B t ; 0 < t < / are defined under the same measure (this
156
H. Wu
can for instance be easily seen using the definition in the case where D is the upper half-plane). Brownian Excursion Occupation Times and the Dirichlet Problem. Let us first make a comment on the relation between the Brownian excursion measure and the Dirichlet problem. Let u be the solution to the Dirichlet problem, i.e. u D 0 in U and u D f on @U. For all z 2 @U and all positive , we have that Z
Z
1
dt1A .t /f . // D E.1 /z .
E.1 /z . 0
dt1A .t /1t f . // Z
0
Z
0
Z
0
1
D E.1 /z .
dt1A .t /1t E.f . /jFt //
D E.1 /z .
dt1A .t /Et .f . ///
D E.1 /z .
dt1A .t /u.t // 0
Z D
dyGU ..1 /z; y/u.y/ A
And for the Brownian excursion measure D U , we have that Z Z Z 2 d . E dt1A .t /f . // D lim dt1A .t /f . // i . !0 0 .1 /e 0 0 Z Z 2 d D lim dyGU ..1 /e i ; y/u.y/ !0 0 A Z 2 Z D 2d dy hU .y; e i /u.y/ 0
Z
D2
A
Z
A
0
Z D2
2
d hU .y; e i /
dy u.y/ dy u.y/ A
That is to say, we can represent the solution to the Dirichlet problem via the Brownian excursion measure by the formula Z Z . dt1A .t /f . // D 2 dy u.y/ A
0
Since the Brownian excursion is reversible, we also have that Z Z .f .0 / dt1A .t // D 2 dy u.y/ 0
A
(6)
On the Occupation Times of Brownian Excursions and Brownian Loops
157
Hence, if we put a weight f on starting point of the excursion, then the mean occupation time spent in A is measured by the integral of u on A, where u is the solution to the corresponding Dirichlet problem. By conformal invariance, (6) also holds for any simply connected domain. R We would like to note that, if we set f D 1 in (6), we get that D . 0 dt1A .t // is equal to twice the area of A. In particular, D ./ is therefore just twice the area of D. The Covariance Structure. We now turn our attention towards the proof of (1). This formula can be understood as follows: we can cut A B into very small pieces, calculate on each small piece and then add all these pieces together. On each small piece dx dy, the Brownian excursion starts from the boundary, firstly it hits the small piece dx (with a small probability), after this time, it is a true Brownian motion starting nearby x, which is (almost) independent of the past and then the expected time of this new Brownian motion spent in the neighborhood of y before exiting D is close to GD .x; y/dy. When we add up all these small pieces together and we obtain the right-hand side of the formula. For a precise calculation, we first consider the case where D D U as the general case will then follow from conformal invariance. We also use the notation that D U . Let denote a Brownian excursion in U. For all z 2 @U and all positive , Z
Z
E.1 /z .
ds1A .s / 0
Z dt1B .t // D E.1 /z .
s
Z
ds1A .s /E. Z
0
dt1B .t /jFs // s
Z
D E.1 /z .
ds1A .s /Es . Z
0
dt1B .t /// 0
D E.1 /z .
ds1A .s /GU .s ; B// 0
Z D
dyGU ..1 /z; y/GU .y; B/: A
And for the Brownian excursion measure, we have that Z
Z
.
ds1A .s / 0
s
Z Z d E dt1B .t // D lim ds1A .s / dt1B .t // i . !0 0 .1 /e 0 s Z 2 Z d D lim dyGU ..1 /e i ; y/GU .y; B/ !0 0 A Z 2 Z D 2d dyhU .y; e i /GU .y; B/ Z
2
A
0
Z
D2
Z
2
d hU .y; e i /
dyGU .y; B/ Z
A
D2
0
dyGU .y; B/ A
158
H. Wu
By symmetry of the Green’s function (GU .x; y/ D GU .y; x/), we have that Z
Z
0
Z
ds1B .s // D 4
ds1A .s /
.
0
dxdyGU .x; y/: AB
This concludes the proof of the (1), since we can use to conformal invariance to derive the formula for general simply connected domain D. More generally, we have that Z Z Z D . dsf .s / dsg.s // D 4 dx dyGD .x; y/f .x/g.y/ (7) 0
0
for all measurable bounded functions f and g. Large Intensity Clouds of Excursions and GFF. Let us now use this formula to make a link between Brownian excursions and the GFF. For this we are going to use Poissonian cloud of excursions in D, as in [15]. Recall that a Poisson cloud of excursions with intensity cD is a random countable family of Brownian excursions in D, that is defined as a Poisson point process with intensity cD . In particular, the union of two independent Poissonian clouds of Brownian excursions in D with intensity c1 D and c2 D is a Poissonian cloud of excursions in D with intensity .c1 C c2 /D . Let us now consider an i.i.d. sequence M j ; j 1 of Poissonian clouds of excursions in D with the common intensity D . For each j 1, and each f 2 ./H.D/, define the “cumulative occupation” time of M j by X Z . / j Xf D dsf .s /: 2M j
0
j
The fact that ./ is finite (as soon as the area of D is finite) ensures that Xf is almost surely finite (as soon as f is bounded) because its expectation is bounded. We then define j j j XQ f D Xf E.Xf /: On an enlarged probability space, we can also define an i.i.d. family of random variable indexed by the set of excursions in [j M j such that P . D 1/ D P . D 1/ D 1=2. We can then define j Yf
D
X 2M j
Z
. /
dsf .s /: 0
It is easy to see that Yf1 ; Yf2 ; Yf3 ; : : : are i.i.d. centered random variables with common variance Z Z Z 2 dsf .s / dsf .s // D 4 dxdyGD .x; y/f .x/f .y/: f D D . 0
0
On the Occupation Times of Brownian Excursions and Brownian Loops
159
The same is true for XQ f1 ; XQ f2 ; XQ f3 ; : : :. By the Central Limit Theorem, we have that 1 p .Yf1 C ::: C YfN / N converges in law as N ! 1 to a centered Gaussian random p variable Yf with 2 1 N Q Q variance f . The same holds for the sequence .Xf C ::: C Xf /= N . Hence, we see that the GFF can be viewed as the limit (in law, and in the sense of finite-dimensional distributions) of the occupation times fluctuations of a Poisson cloud of Brownian excursions, when the intensity tends to infinity. Higher-Order “Moments”. We just mention that our proof can be adapted directly in order to show that for all p 2: Z D . .0; /p
dt1 : : : dtp 1t1 <:::<tp 1A1 .t1 / 1Ap .tp //
Z D2
A1 Ap
dx1 dxp GD .x1 ; x2 / GD .xp1 ; xp /
which gives R for instance (when one sums over all possible order of visits) a formula for D .. 0 f .s /ds/p /. We have chosen to focus on the case p D 2 because of the above-mentioned link with Gaussian fields. Non-Simply-Connected Domains. Suppose that D is a finitely connected open domain in the plane. Then, by Koebe’s uniformization Theorem (see [3]), it is possible to map it conformally onto a circular domain i.e., the unit disk U punctured by a finite number of disjoint closed disks. It is very easy to generalize the definition of the Brownian excursion measure in circular domains (adding the contributions corresponding to starting points in the neighborhood of each of the boundary disks), and to see that all our proofs go through without any real difficulty, so that all our statements are in fact valid also in circular domains. One can then define the excursion measure in D via conformal invariance starting from circular domains, and then, by conformal invariance of all the quantities involved, we easily see that all our statements are also valid in D.
4 Occupation Times of Brownian Loops Brownian Loop Measure. We now briefly recall the construction of the Brownian loop measure [5]. As for the Brownian excursion measure, we can first define it in the unit disc, and then define it in any other simply connected domain using conformal invariance (and one then checks that this is indeed consistent with other possible constructions).
160
H. Wu
For any r 2 .0; 1, define Ur D rU. For any x 2 Ur and any z 2 @Ur , one can define the Brownian motion started at x and conditioned to exit Ur at z (this can be rigorously defined as the limit when ! 0 of the law of the Brownian motion conditioned to exit Ur in an -neighborhood of z). Let us denote this probability r measure by Px!z . Then, as for the excursion measure, one can let x ! z, and renormalize it in order to get a measure on macroscopic sets i.e. define r mrz ./ D lim 1 hUr .z C n; z/PzC n!z ./ !0
where n D nz;Ur is the inwards pointing normal vector at z 2 @Ur : Then, one can define the loop measure in U by integrating z on @Ur , and then integrating r from 0 to 1: Z Z 1
U ./ D
2
rdr 0
0
d mrrei ./:
In fact, the above definition is not quite the loop measure because it defines a measure on parametrized loops. We will forget about the precise parametrization of the loop and view U as a measure on loops defined modulo monotone reparametrization (where the time-parameter should be viewed as an element of the circle, because the end-point of the loop is the same as the starting point, this is possible). It turns out that this definition of U is then invariant under the Moebius transformations that map the unit disc onto itself. Hence, it is possible to define, for a general simply connected domain D, the loop measure D as the image of U by any conformal map ˚ that maps U onto D. And we usually denote D U : Before going on, we would like to say a word on the value of ./. In fact, by direct computation we have that ./ D 1 which is very different from ./ mentioned before. A direct way to check that ./ D 1 goes as follows. Consider D to be the square Œ0; 12 . For any dyadic square d in D with sidelength 2n , a direct scaling argument shows that the mass (for ) of the set of loops that stay in d and have a time-length in Œ4n ; 24n / does not dependPon d . Hence, if we sum this quantity over all dyadic squares d in D, and because n 4n 4n D 1, we readily see that ./ D 1. However, almost the same argument ensures that . 1C / is finite for > 0 (and bounded D). Indeed, in the case of the unit square, we can decompose the set of loops with time-length in Œ4n ; 41n / according to the dyadic square in which its lowest point lies. This leads readily to the bound .1 <1 1C / C
X
4n .41n /1C < 1
n1
and one can see by other means that . > t/ decays exponentially fast as t ! 1. In particular, we get that . 2 / is finite (as soon as D is bounded). Covariance Structure. Our goal is now to prove (2). As before, we are going to derive the result first in the case where D D U, and the general result will then follow using conformal invariance. Again, it will be convenient to (loosely speaking)
On the Occupation Times of Brownian Excursions and Brownian Loops
161
divide AB into infinitesimal pieces dx dy, make the computation on each piece, and then add all these pieces together. Clearly, this will give a formula of the type Z
Z
D .
Z
ds1B .s // D
ds1A .s / 0
0
dxdyFD .x; y/ AB
where FD .x; y/ is the “covariance” function between x and y determined by the Brownian loop measure. Just as what we have done to derive the conformal invariance of the Green’s function in the (3), we can also derive the conformal invariance of F : F˚.D/ .˚.x/; ˚.y// D FD .x; y/: To determine FD .x; y/, it is enough to describe FU .0; y0 / for y0 2 .0; 1/, because there exists a y0 and a conformal map ˚ from D onto U such that ˚.x/ D 0; ˚.y/ D y0 . And now begin our computation. For r 2 .0; 1/; z 2 @Ur , we can write Z r . EzC n!z
Z
ds1A .s / 0
dt1B .t // s
1 2 U.z; 0 / \ @Ur / Z Z r . ds1A .s / dt1B .t /1 2U.z; 0 /\@Ur / EzC n
D lim 0
r !0 PzC n .
0
s
1 hUr .z C n; U.z; 0 / \ @Ur / Z Z r . ds1A .s / dt1B .t /hUr .t ; U.z; 0 / \ @Ur // EzC n
D lim 0
!0
0
s
1 Er . D hUr .z C n; z/ zC n
Z
Z
ds1A .s / 0
dt1B .t /hUr .t ; z//: s
Hence, Z hUr .z C
r n; z/EzC n!z .
Z r D EzC n .
Z
Z
0
ds1A .s / 0
dt1B .t /hUr .t ; z// s
Z
dx GUr .z C n; x/
D A
dt1B .t // s
Z
ds1A .s /
dy GUr .x; y/hUr .y; z/ B
162
H. Wu
and letting ! 0, we get Z mrz .
Z
ds1A .s / 0
ds1B .s // 0
Z Z 2 D lim dx GUr .z C n; x/ dy GUr .x; y/hUr .y; z/ !0 A B Z D4 dxdyGUr .x; y/hUr .x; z/hUr .y; z/: AB
For simplicity, we define a new kernel Z
2
KUr .x; y/ D 4
d hUr .x; rei /hUr .y; rei / 0
and then we have that Z
Z
U . 0
Z
Z
1
ds1B .s // D
ds1A .s /
rdr
0
dxdyGUr .x; y/KUr .x; y/: AB
0
Note that KUr .rx; ry/ D
1 KU .x; y/ r2
and GUr .rx; ry/ D GU .x; y/: Furthermore, KU .0; y/ D 2=. Suppose that A D U.0; / and B D U.y0 ; ı/ where and ı are both small. In our decomposition of U , the loop can visit B only if it started on a circle of radius r > y0 . Hence, on the one hand, as and ı tend to 0, Z
Z
U .
ds1A .s / 0
Z
0
Z
1
ds1B .s // D
rdr .A
y0
Z
T
Ur /.B
T
dxdyGUr .x; y/KUr .x; y/ Ur /
1
rdr. 2 ı 2 /GUr .0; y0 /KUr .0; y0 / y0
Z
1
1 y0 y0 drGU .0; /KU .0; / r r r y0 Z 1 2 1 y0 D . 2 ı 2 / 2 dr. log. // y0 r r
D . ı / 2
2
D . 2 ı 2 /
1 .log y0 /2 2
On the Occupation Times of Brownian Excursions and Brownian Loops
163
On the other hand, this quantity is precisely behaving as . 2 ı 2 /FU .0; y0 / and as a result, we get that FU .0; y0 / D
1 .log y0 /2 D .GU .0; y0 //2 : 2
We can then conclude that (2) holds in U, and then also in D by conformal invariance. More generally, we have that Z
Z
D . 0
Z
dsg.s // D
dsf .s /
dxdy.GD .x; y//2 f .x/g.y/: AB
0
for all measurable bounded functions f and g. Brownian Loop-Soups and Fields. Just as in the case of Brownian excursion measure, we can use this formula to make a link between Brownian loops and some Gaussian Fields. Let M j ; j 1 be a sequence of i.i.d Poissonian clouds of loops in D with the common intensity D . We can try to give the same definitions of the j j quantities XQ f ; Yf ; j 1. However, things are a little more complicated, due to the fact that the same scaling argument that showed that ./ D 1 implies that X
./ D 1
2M j
almost surely, so that some care is needed. j The definition of Yf is however not a big problem. Recall that on an enlarged probability space, one associates to each loop a random variable with E. / D 0 and E.. /2 / D 1. But (2) precisely ensures that the sum X 2M j
Z
. /
f .s /ds 0
makes sense in L2 , and that its second moment is equal to Z f2
Z
D D .
dsf .s / 0
Z dsf .s // D
dxdy.GD .x; y//2 f .x/f .y/
0
which is finite. Then, just as in the case of the clouds of excursions, the sequence Yf1 , Yf2 ; : : : is made of i.i.d centered random variables with common variance f2 . By the Central Limit Theorem, 1 p .Yf1 C ::: C YfN / N
164
H. Wu
converges in law as N ! 1 to a centered Gaussian random variable with variance f2 . Hence, we obtain another Gaussian Field, characterized by this new covariance structure. j It is also still possible to make sense of XQf even though it is not possible to j
define Xf . It suffices to partition the set of loops (in D) into a countable set of loops Ak ; k 1 such that for each k, .1 2Ak / is finite (for instance, one can take Ak D f W ./ > 1=kg n .A1 [ : : : [ Ak1 /g. Then, one can define j XQf
D
X k1
0 @
X 2Ak
\M j
Z
. /
f .s /ds E. 0
X 2Ak
\M j
1
Z
. /
f .s /ds/A
0
and check that this sum with respect to k converges in L2 , and that its second moment is the same as that of Yf . The rest of the argument is again the same.
5 Intersections of Brownian Excursions In this section, we try to find the relation between intersections of Brownian excursion “occupations times” and Brownian loop occupation times, the former being defined via the intersection local time. Let us first recall some features of Brownian intersection local times. Let p 2 be an integer, and let Z 1 ; :::; Z p denote p independent Brownian motions in R2 , started at x 1 ; :::; x p respectively. The intersection local time of Z 1 ; :::; Z p is a p random measure ˛.ds1 :::dsp / on RC , supported on p
f.s1 ; :::; sp / 2 RC W Zs11 D ::: D Zspp g: The basic description concerning the intersection local time that we will use goes as follows (see [6] for details): Proposition 2. Almost surely, one can define a (random) measure ˛.ds1 :::dsp / on p RC such that, for any A1 ; :::; Ap bounded Borel subsets of RC , ˛.A1 ::: Ap / D lim ˛ .A1 ::: Ap / !0
in the Ln norm, for any n < 1, where Z ˛ .ds1 :::dsp / D ds1 :::dsp with ıy .z/ D
1 1 .z/. 2 U.y; /
R2
dyıy .Zs11 /:::ıy .Zspp /
On the Occupation Times of Brownian Excursions and Brownian Loops
165
Let us use this in the context of the Brownian excursion measure. This time we shall consider two Brownian excursions and 0 defined under the (infinite) measure D ˝ D , and study the behavior of their intersection local time that spent in two disjoint sets A and B, as before: Z
D ˝ D .
Z
0 0
0
Z
0
˛.dtdt /1.t D 00 2A/ t
Z
Z
0
D lim D ˝ D . !0
Z
0
Z
0
Z
!0
0
0
0
Z
D lim
Z
0
0
Z
D lim
!0
dsds 0
˛ .dtdt 0 /1.t 2A/ 1. 00 2A/
Z
Z
0
dtdt 0
t
s
Z dy
Z
0 0
dy.4
dtıx .t /1.t 2A/
0
Z
Z
Z
Z
D ˝ D .
AB
D 16
dxıx .t /ıx .t00 /1.t 2A/ 1. 00 2A/
dyıy .s /ıy .s00 /1.s 2B/ 1. 00 2B/ /
dt 0 ıx .t00 /1. 00 2A/ t dx
Z
0
Z
dx
!0
˛.dsds 0 /1.s D 00 2B/ / s
0
t
0
0
Z
0
˛ .dsds 0 /1.s 2B/ 1. 00 2B/ /
D lim D ˝ D . Z
0
Z
s
0
0
0
dsıy .s /1.s 2B/
ds 0 ıy .s00 /1. 00 2B/ / s
dadbıx .a/ıy .b/GD .a; b//2
dxdy.GD .x; y//2 AB
Hence, we see that pairs of Brownian excursions give rise to the same covariance structure as the Brownian loops. In a way, this is not too surprising, as for two points x and y that are both visited by and by 0 , one sees in a way a loop structure (the part of from x to y, and then the part of 0 back from y to x). Note that by a similar calculation, one gets that for any p 3, if one defines for any A, Z Tp .AI ; : : : ; / D 1
p
Z
1
0
˛.dt1 : : : dtp /1.t1 DDtp 2A/ ; 1
0
then ˝p
p
:::
D .Tp .A/Tp .B// D 4p
Z dxdy.GD .x; y//p : AB
p
166
H. Wu
Acknowledgements This paper is based on my Master’s thesis and was completed under the guidance of my supervisor Professor Wendelin Werner. The author acknowledges the support from a Fondation CFM-JP Aguilar grant.
References 1. R. Durrett, Brownian Motion and Martingales in Analysis (Wadsworth Mathematics Series, 1984) 2. K. Gawe¸dzki, Lectures on Conformal Field Theory, Quantum fields and strings: a course for mathematicians, Vol. 1, 2 (Princeton, NJ, 1996/1997), Amer. Math. Soc., (Providence, RI, 1999), pp. 727–805 3. P. Koebe, Abhandlungen zur Theorie der konformen Abbildung VI. Abbildung mehrfach zusammenhängender schlichter Bereiche auf Kreisbereiche, Math. Z. 7 235–301 (1920) 4. G.F. Lawler, Conformally Invariant Processes in the Plane, (Mathematical Surveys and Monographs, AMS, 2005), no. 114, xii+242 5. G.F. Lawler, W. Werner, The brownian loop soup, Probab. Theor. Relat. Field. 128(4), 565–588 (2004) 6. J.-F. Le Gall, Some properties of Planar Brownian Motion, École d’Été de Probabilités de Saint-Flour XX—1990, Lecture Notes in Mathematics, vol. 1527, (Springer, Berlin, 1992), pp. 111–235 7. Y. Le Jan, Markov Paths, Loops and Fields, École d’Été de Probabilités de Saint-Flour XXXVIII—2008, Lecture Notes in Mathematics, vol. 2026, (Springer, Berlin, 2011), p. 134 8. P. Lévy, Processus Stochastiques et Mouvement Brownien, Les Grands Classiques GauthierVillars. Éditions (Jacques Gabay, Sceaux, 1992); Followed by a note by M. Loève, Reprint of the second (1965) edition. 9. E. Nelson, The free markoff field, J. Funct. Anal. 12, 211–227 (1973) 10. S.C. Port, C.J. Stone, Brownian Motion and Classical Potential Theory, (Academic Press, New York, 1978), p. 236 11. M. Rao, Brownian Motion and Classical Potential Theory, Lecture Notes Series, No. 47. Matematisk Institut, (Aarhus University, Aarhus, 1977) 12. S. Sheffield, Gaussian free fields for mathematicians, Probab. Theor. Relat. Field. 139(3–4), 521–541 (2007) 13. S. Sheffield, Conformal Weldings of Random Surfaces: SLE and the Quantum Gravity Zipper, preprint, arXiv:1012.4797 (2010) 14. S. Sheffield, W. Werner, Conformal Loop Ensembles: The Markovian Characterization and the Construction Via Loop-Soups, Ann. Math. (to appear), arXiv:1006.2374v3 (2011) 15. W. Werner, Conformal restriction and related questions, Probab. Surv. 2, 145–190 (2005) 16. W. Werner, Some Recent Aspects of Random Conformally Invariant Systems, (Mathematical statistical physics, Elsevier B. V., Amsterdam, 2006), pp. 57–99
Discrete Approximations to Solution Flows of Tanaka’s SDE Related to Walsh Brownian Motion Hatem Hajri
Abstract In a previous work, we have defined a Tanaka’s SDE related to Walsh Brownian motion which depends on kernels. It was shown that there are only one Wiener solution and only one flow of mappings solving this equation. In the terminology of Le Jan and Raimond, these are respectively the stronger and the weaker among all solutions. In this paper, we obtain these solutions as limits of discrete models.
1 Introduction and Main Results Consider Tanaka’s equation: Z
t
's;t .x/ D x C
sgn.'s;u .x//d Wu ; s t; x 2 R;
(1)
s
where sgn.x/ D 1fx>0g 1fx0g ; Wt D W0;t 1ft >0g Wt;0 1ft 0g and .Ws;t ; s t/ is a real white noise on a probability space .˝; A ; P/ (see Definition 1.10 [6]). This is an example of a stochastic differential equation which admits a weak solution but has no strong solution. If K is a stochastic flow of kernels (see Sect. 2.1 [5]) and W is a real white noise, then by definition, .K; W / is a solution of Tanaka’s SDE if for all s t; x 2 R, f 2 Cb2 .R/ (f is C 2 on R and f 0 ; f 00 are bounded), Z
t
Ks;t f .x/ D f .x/ C s
1 Ks;u .f sgn/.x/W .d u/ C 2 0
Z
t
Ks;u f 00 .x/d u a:s:
(2)
s
H. Hajri () Département de Mathématiques, Université Paris-Sud, 91405 Orsay Cedex, France e-mail:
[email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__8, © Springer-Verlag Berlin Heidelberg 2012
167
168
H. Hajri
When K D ı' is a flow of mappings, K solves (2) if and only if ' solves (1) by Itô’s formula. In [7], Le Jan and Raimond have constructed the unique flow of mappings associated to (1). It was shown also that 1 W .x/ D ıxCsgn.x/Ws;t 1ft s;x g C .ıW C C ıW C /1ft >s;x g ; s t; x 2 R; Ks;t s;t 2 s;t is the unique F W adapted solution (Wiener flow) of (2) where s;x D inffr s W Ws;r D jxjg; Ws;tC WD Ws;t inf Ws;u : u2Œs;t
In [5], an extension of (2) in the case of Walsh Brownian motion was defined as follows Definition 1. Fix N 2 N , ˛1 ; ; ˛N > 0 such that
N X
˛i D 1 and consider the
i D1
graph G consisting of N half lines .Di /1i N emanating from 0 (see Fig. 1). Let ei be a vector of modulus 1 such that Di D fhei ; h > 0g and define for all z 2 G, e.z/ D ei if z 2 Di ; z ¤ 0 (convention e.0/ D eN ). Define the following distance on G: ( h C h0 if i ¤ j; .h; h0 / 2 R2C ; d.hei ; h0 ej / D jh h0 j if i D j; .h; h0 / 2 R2C :
G
ei O
Fig. 1 Graph G
(Di , αi)
Discrete Approximations to Solution Flows of Tanaka’s SDE
169
For x 2 G, we will use the simplified notation jxj WD d.x; 0/. We equip G with its Borel -field B.G/ and set G D G n f0g. Let Cb2 .G / be the space of all f W G ! R such that f is continuous on G and has bounded first and second derivatives (f 0 and f 00 ) on G (here f 0 .z/ is the derivative of f at z in the direction e.z/ for all z ¤ 0), both limz!0;z2Di ;z¤0 f 0 .z/ and limz!0;z2Di ;z¤0 f 00 .z/ exist for all i 2 Œ1; N . Define ( D.˛1 ; ; ˛N / D f 2
Cb2 .G /
W
N X i D1
) ˛i
0
lim
z!0;z2Di ;z¤0
f .z/ D 0 :
Now, Tanaka’s SDE on G extended to kernels is the following (see Remarks 3 (1) in [5] for a discussion of its origin). Tanaka’s equation on G. On a probability space .˝; A ; P/, let W be a real white noise and K be a stochastic flow of kernels on G. We say that .K; W / solves .T / if for all s t; f 2 D.˛1 ; ; ˛N /; x 2 G, Z
t
Ks;t f .x/ D f .x/ C
Ks;u f 0 .x/W .d u/ C
s
1 2
Z
t
Ks;u f 00 .x/d u a:s:
s
If K D ı' is a solution of .T /, we just say that .'; W / solves .T /. Equation .T / is a particular case of an equation .E/ studied in [5] (it corresponds to " D 1 with the notations of [5]). It was shown (see Corollary 2 [5]) that if .K; W / solves .T /, then .W / .K/ and therefore one can just say that K solves .T /. We also recall Theorem 1. [5] There exists a unique Wiener flow K W (resp. flow of mappings ') which solves .T /. As described in Theorem 1 [5], the unique Wiener solution of .T / is simply W Ks;t .x/
D ıxCe.x/Ws;t 1ft s;x g C
N X i D1
˛i ıei W C 1ft >s;x g : s;t
(3)
where s;x D inffr s W x C e.x/Ws;r D 0g D inffr s W Ws;r D jxjg:
(4)
However, the construction of the unique flow of mappings ' associated to .T / relies on flipping Brownian excursions and is more complicated. Another construction of ' using Kolmogorov extension theorem can be derived from Sect. 4.1 [7] similarly to Tanaka’s equation. Here, we restrict our attention to discrete models. The one point motion associated to any solution of .T / is the Walsh Brownian motion W .˛1 ; ; ˛N / on G (see Proposition 3 [5]) which we define as a strong Markov process with càdlàg paths, state space G and Feller semigroup .Pt /t 0 as
170
H. Hajri
given in Sect. 2.2 [5]. When N D 2, it corresponds to the famous skew Brownian motion [4]. Our first result is the following Donsker approximation of W .˛1 ; ; ˛N / Proposition 1. Let M D .Mn /n0 be a Markov chain on G started at 0 with stochastic matrix Q given by: 1 8i 2 Œ1; N ; n 2 N : 2 (5) Let t 7 ! M.t/ be the linear interpolation of .Mn /n0 and Mtn D p1n M.nt/; n 1. Then law .Mtn /t 0 ! .Zt /t 0
Q.0; ei / D ˛i ; Q.nei ; .n C 1/ei / D Q.nei ; .n 1/ei / D
n ! C1
in C.Œ0; C1Œ; G/ where Z is an W .˛1 ; ; ˛N / started at 0. This result extends that of [2] who treated the case ˛1 D D ˛N D N1 and of course the Donsker theorem for the skew Brownian motion (see [1] for example). We show in fact that Proposition 1 can be deduced immediately from the case N D 2. In this paper we study the approximation of flows associated to .T /. Among recent papers on the approximation of flows, let us mention [8] where the author construct an approximation for the Harris flow and the Arratia flow. Let GN D fx 2 GI jxj 2 Ng and P.G/ (resp. P.GN /) be the space of all probability measures on G (resp. GN ). We now come to the discrete description of .'; K W / and introduce Definition 2 (Discrete flows). We say that a process p;q .x/ (resp. Np;q .x/) indexed by fp q 2 Z; x 2 GN g with values in GN (resp. P.GN /) is a discrete flow of mappings (resp. kernels) on GN if: (i) The family f i;i C1 I i 2 Zg (resp. fNi;i C1 I i 2 Zg) is independent. (ii) 8p 2 Z; x 2 GN ; p;pC2 .x/ D pC1;pC2 . p;pC1 .x// .resp. Np;pC2 .x/ D Np;pC1 NpC1;pC2 .x// a.s. where Np;pC1 NpC1;pC2 .x; A/ WD
X
NpC1;pC2 .y; A/Np;pC1 .x; fyg/ for all A GN :
y2GN
We call (ii), the cocycle or flow property. The main difficulty in the construction of the flow ' associated to (1) [7] is that it has to keep the consistency of the flow. This problem does not arise in discrete time. Starting from the following two remarks: (1) 's;t .x/ D x C sgn.x/Ws;t if s t s;x , (2) j's;t .0/j D Ws;tC and sgn.'s;t .0// is independent of W for all s t, one can easily expect the discrete analogous of ' as follows: consider an original random walk S and a family of signs .i / which are independent. Then:
Discrete Approximations to Solution Flows of Tanaka’s SDE
171
(1) A particle at time k and position n ¤ 0, just follows what the SkC1 Sk tells him (goes to n C 1 if SkC1 Sk D 1 and to n 1 if SkC1 Sk D 1). (2) A particle at 0 at time k does not move if SkC1 Sk D 1, and moves according to k if SkC1 Sk D 1. The situation on a finite half-lines is very close. Let S D .Sn /n2Z be a simple random walk on Z, that is .Sn /n2N and .Sn /n2N are two independent simple random walks on Z and .i /i 2Z be a sequence of i.i.d random variables with law N X ˛i ıei which is independent of S . For p n, set i D1 C Sp;n D Sn Sp ; Sp;n D Sn min Sh D Sp;n min Sp;h : h2Œp;n
h2Œp;n
and for p 2 Z; x 2 GN , define C ‰p;pC1 .x/ D x C e.x/Sp;pC1 if x ¤ 0; ‰p;pC1 .0/ D p Sp;pC1 :
Kp;pC1 .x/ D ıxCe.x/Sp;pC1 if x ¤ 0; Kp;pC1 .0/ D
N X i D1
˛i ıS C
p;pC1 ei
:
In particular, we have Kp;pC1 .x/ D EŒı‰p;pC1.x/ j.S /. Now we extend this definition for all p n 2 Z; x 2 GN by setting ‰p;n .x/ D x1fpDng C ‰n1;n ı ‰n2;n1 ı ı ‰p;pC1 .x/1fp>ng ; Kp;n .x/ D ıx 1fpDng C Kp;pC1 Kn2;n1 Kn1;n .x/1fp>ng : We equip P.G/ with the following topology of weak convergence: ) ( Z Z jg.x/g.y/j 1; g.0/ D 0 : ˇ.P; Q/ D sup j gdP gdQj; kgk1 C sup jxyj x6 D y In this paper, starting from .‰; K/, we construct .'; K W / and in particular show the following Theorem 2. (1) ‰ (resp. K) is a discrete flow of mappings (resp. kernels) on GN . (2) There exists a joint realization . ; N; '; K W / on a common probability space .˝; A ; P/ such that law
(i) . ; N / D .‰; K/. (ii) .'; W / (resp. .K W ; W /) is the unique flow of mappings (resp. Wiener flow) which solves .T /. (iii) For all s 2 R; T > 0; x 2 G; xn 2 p1n GN such that limn!1 xn D x, we have
172
H. Hajri
lim
sup
n!1 st sCT
and lim
sup
n!1 st sCT
1 jp n
bnsc;bnt c .
p nxn / 's;t .x/j D 0 a:s:
p p W ˇ.Kbnsc;bnt c. nxn /. n:/; Ks;t .x// D 0 a:s:
(6)
This theorem implies also the following Corollary 1. For all s 2 R; xn 2 GN , let1 t 7! ‰.t/n be the linear interpolation of n .x/ j.S /; t s; ‰bnsc;k .x/; k bnsc and ‰s;t .x/ WD pn ‰.nt/; Ks;t .x/DEŒı‰s;t n 1: For all 1 p q, .xi /1i q G, let xin 2 xi . Define
p1 GN n
such that limn!1 xin D
p p p n p Y n D ‰sn1 ; . nx1n /; ; ‰snp ; . nxpn /; KsnpC1 ; . nxpC1 /; ; Ksnq ; . nxqn / : Then law
Y ! Y in n
n ! C1
p Y i D1
q Y
C.Œsi ; C1Œ; G/
C.Œsj ; C1Œ; P.G//
j DpC1
where Y D 's1 ; .x1 /; ; 'sp ; .xp /; KsWpC1 ; .xpC1 /; ; KsWq ; .xq / : Our proof of Theorem 2 is based on a remarkable transformation introduced by Csaki and Vincze [9] which is strongly linked with Tanaka’s SDE. Let S be a simple random walk on Z (SRW) and " be a Bernoulli random variable independent of S (just one!). Then there exists a SRW M such that .M / D ."; S / and moreover 1 1 law . p S.nt/; p M.nt//t 0 ! .Bt ; Wt /t 0 in C.Œ0; 1Œ; R2 /: n ! C1 n n where t 7! S.t/ (resp. M.t/) is the linear interpolation of S (resp. M ) and B; W are two Brownian motions satisfying Tanaka’s equation d Wt D sgn.Wt /dBt : We will study this transformation with more details in Sect. 2 and then extend the result of Csaki and Vincze to Walsh Brownian motion (Proposition 2); Let
Discrete Approximations to Solution Flows of Tanaka’s SDE
173
S D .Sn /n2N be a SRW and associate to S the process Yn WD Sn minSk , flip indekn
pendently every “excursion ” of Y to each ray Di with probability ˛i , then the resulting process is not far from a random walk on G whose law is given by (5). In Sect. 3, we prove Proposition 1 and study the scaling limits of ‰; K.
2 Csaki-Vincze Transformation and Consequences In this section, we review a relevant result of Csaki and Vincze and then derive some useful consequences offering a better understanding of Tanaka’s equation.
2.1 Csaki-Vincze Transformation Theorem 3. ([9] page 109) Let S D .Sn /n0 be a SRW. Then, there exists a SRW S D .S n /n0 such that: Y n WD maxS k S n ) jY n jSn jj 2 8n 2 N: kn
Sketch of the proof. Here, we just give the expression of S with some useful comments (see also the figures below). We insist that a careful reading of the pages 109 and 110 [9] is recommended for the sequel. Let Xi D Si Si 1 ; i 1 and define 1 D min fi > 0 W Si 1 Si C1 < 0g; lC1 D min fi > l W Si 1 Si C1 < 0g 8l 1: For j 1, set Xj D
X
.1/lC1 X1 Xj C1 1fl C1j lC1 g :
l0
Let S 0 D 0; S j D X 1 C C X j ; j 1. Then, the theorem holds for S . We call T .S / D S the Csaki-Vincze transformation of S (Fig. 2). Note that T is an even function, that is T .S / D T .S /. As a consequence of (iii) and (iv) [9] (page 110), we have l D min fn 0; S n D 2lg 8l 1: This entails the following Corollary 2. (1) Let S be a SRW and define S D T .S /. Then: (i) For all n 0, we have .S j ; j n/ _ .S1 / D .Sj ; j n C 1/.
(7)
174
H. Hajri
6 4 2
O
τ1
τ2
τ3
S
Fig. 2 S and S
(ii) S1 is independent of .S /. (2) Let S D .S k /k0 be a SRW. Then: (i) There exists a SRW S such that: Y n WD maxS k S n ) jY n jSn jj 2 8n 2 N: kn
(ii) T 1 fS g is reduced to exactly two elements S and S where S is obtained by adding information to S. Proof. (1) We retain the notations just before the corollary. (i) To prove the inclusion , we only need to check that fl C 1 j lC1 g 2 .Sh ; h n C 1/ for a fixed j n. This is clear since fl P D mg 2 .Sh ; h m C 1/ for all l; m 2 N. For all 1 j n, we have Xj C1 D l0 .1/lC1 X1 X j 1fl C1j lC1 g . By (7), fl C 1 j lC1 g 2 .S h ; h j 1/ and so the inclusion holds. (ii) We may write 1 D min fi > 1 W X1 Si 1 X1 Si C1 < 0g; lC1 D min fi > l W X1 Si 1 X1 Si C1 < 0g 8l 1: This shows that S is .X1 Xj C1 ; j 0/-measurable and (ii) is proved. (2) (i) Set X j D S j S j 1 ; j 1 and l D min fn 0; S n D 2lg for all l 1. Let " be a random variable independent of S such that: P." D 1/ D P." D 1/ D
1 : 2
Discrete Approximations to Solution Flows of Tanaka’s SDE
175
Define 0
Xj C1
1 X D "1fj D0g C @ .1/lC1 "X j 1fl C1j lC1 g A 1fj 1g: l0
Then set S0 D 0; Sj D X1 C Xj ; j 1. It is not hard to see that the sequence of the random times i .S /; i 1 defined from S as in Theorem 3 is exactly i ; i 1 and therefore T .S / D S . (ii) Let S such that T .S / D S . By (1), .S / _ .S1 / D .S / and S1 is independent of S which proves (ii). t u
2.2 The Link with Tanaka’s Equation Let S be a SRW, S D T .S / and t 7! S.t/ (resp. S .t/) be the linear interpolation .n/ .n/ of S (resp. S ) on R. Define for all n 1, St D p1n S.nt/; S t D p1n S.nt/. Then, it can be easily checked (see Proposition 2.4 in [3] page 107) that .n/
law
.n/
.S t ; St /t 0 ! .Bt ; Wt /t 0 in C.Œ0; 1Œ; R2 /: n ! C1
In particular B and W are two standard Brownian motions. On the other hand, jYnC jSn jj 2 8n 2 N with YnC WD S n minS k by Theorem 3 which implies kn
jWt j D Bt min Bu . Tanaka’s formula for local time gives 0ut
Z jWt j D
t
sgn.Wu /d Wu C Lt .W / D Bt min Bu ; 0ut
0
where Lt .W / is the local time at 0 of W and so d Wu D sgn.Wu /dBu :
(8)
We deduce that for each SRW S the couple .T .S /; S /, suitably normalized and time scaled converges in law towards .B; W / satisfying (8). Finally, remark that T .S / D S ) T .S / D S is the analogue of W solves (8) ) W solves (8). We have seen how to construct solutions to (8) by means of T . In the sequel, we will use this approach to construct a stochastic flow of mappings which solves equation .T / in general.
176
H. Hajri
2.3 Extensions Let S D .Sn /n0 be a SRW and set Yn WD maxSk Sn . For 0 p < q, we say that kn
E D Œp; q is an excursion for Y if the following conditions are satisfied (with the convention Y1 D 0): • Yp D Yp1 D Yq D YqC1 D 0. • 8 p j < q; Yj D 0 ) Yj C1 D 1. For example in Fig. 3, Œ2; 14; Œ16; 18 are excursions for Y . If E D Œp; q is an excursion for Y , define e.E/ WD p; f .E/ WD q. Let .Ei /i 1 be the random set of all excursions of Y ordered such that: e.Ei / < e.Ej / 8i < j . From now on, we call Ei the i th excursion of Y . Then, we have Proposition 2. On a probability space .˝; A ; P /, consider the following jointly independent processes: • D .i /i 1 , a sequence of i.i.d random variables distributed according to N X ˛i ıei . i D1
• .S n /n2N a SRW. Then, on an extension of .˝; A ; P /, there exists a Markov chain .Mn /n2N started at 0 with stochastic matrix given by (5) such that: Y n WD maxS k S n ) jMn i Y n j 2 kn
on the ith excursion of Y . Proof. Fix S 2 T 1 fSg. Then, by Corollary 2, we have jY n jSn jj 2 8n 2 N. Consider a sequence .ˇ i /i 1 of i.i.d random variables distributed according to N X ˛i ıei which is independent of .S ; /. Denote by .l /l1 the sequence of random i D1
times constructed in the proof of Theorem 3 from S . It is sufficient to look to what happens at each interval Œl ; lC1 (with the convention 0 D 0). Using (7), we see that in Œl ; lC1 there are two jumps of maxS k ; from 2l to 2l C 1 kn
(J1 ) and from 2l C 1 to 2l C 2 (J2 ). The last jump (J2 ) occurs always at lC1 by (7). Consequently there are only 3 possible cases: (i) There is no excursion of Y (J1 and J2 occur respectively at l C 1 and l C 2, see Œ0; 1 in Fig. 3). (ii) There is just one excursion of Y (see Œ1 ; 2 in Fig. 3). (iii) There are 2 excursions of Y (see Œ2 ; 3 in Fig. 3).
Discrete Approximations to Solution Flows of Tanaka’s SDE
6
177
|S|
4
Y
2
O
τ
τ
τ
Fig. 3 jSj and Y
El1 Possible values for | S|
tl
Y tl+1
Fig. 4 The case (ii2)
Note that: Y l D Y lC1 D Sl D SlC1 D 0. In the case (i), we have necessarily lC1 D l C 2. Set Mn D ˇ l :jSn j 8n 2 Œl ; lC1 . To treat other cases, the following remarks may be useful: from the expression of S , we have 8l 0: (a) If k 2 Œl C 2; lC1 , S k1 D 2l C 1 ” Sk D 0. (b) If k 2 Œl ; lC1 , Y k D 0 ) jSkC1 j 2 f0; 1g and SkC1 D 0 ) Y k D 0. In the case (ii), let El1 be the unique excursion of Y in the interval Œl ; lC1 . Then, we have two subcases: (ii1) f .El1 / D lC1 2 (J1 occurs at lC1 1). If l C 2 k 6 f .El1 / C 1, then k 1 6 f .El1 /, and so S k1 ¤ 2l C 1. Using (a), we get: Sk ¤ 0. Thus, in this case the first zero of S after l is lC1 . Set: Mn D N.E 1 / :jSn j, where N.E/ is the number of the excursion E. l
(ii2)f .El1 / D lC1 1 (J1 occurs at l C 1 and so Y l C1 D 0/). In this case, using (b) and Fig. 4, we see that the first zero l of S after l is e.El1 / C 1 D l C 2 (Fig. 4).
178
H. Hajri
Set
( Mn D
if n 2 Œl ; l 1
ˇ l :jSn j
N.E 1 / :jSn j if n 2 Œl ; lC1 l
In the case (iii), let El1
and El2
denote respectively the first and 2nd excursion of Y in Œl ; lC1 . We have, l C2 k e.El2 / ) k 1 e.El2 /1 D f .El1 / ) S k1 ¤ 2l C 1 ) Sk ¤ 0 by (a). Hence, the first zero of S after l is l WD e.El2 / C 1 using Y k D 0 ) jSkC1 j 2 f0; 1g in (b). Set: ( Mn D
N.E 1 / :jSn j l
if n 2 Œl ; l 1
N.E 2 /:jSn j if n 2 Œl ; lC1 l
Let (Mn /n2N be the process constructed above. Then clearly jMn i Y n j 2 on the i th excursion of Y . To complete the proof, it suffices to show that the law of .Mn /n2N is given by (5). The only point to verify is P.MnC1 D ei jMn D 0/ D ˛i . For this, consider on another probability space the jointly independent processes .S; ; / such that S is a SRW and ; have the same law as . Let .l /l1 be the sequence of random times defined from S as in Theorem 3. For all l 2 N, denote by l the first zero of S after l and set ( l :jSn j if n 2 Œl ; l 1 Vn D l :jSn j if n 2 Œl ; lC1 law
It is clear, by construction, that M D V . We can write: f0 ; 0 ; 1 ; 1 ; 2 ; g D fT0 ; T1 ; T2 ; g with T0 D 0 < T1 < T2 < : For all k 0, let k WD
N X j D0
law
and k D
N X
ej 1fV jŒTk ;TkC1 2Dj g . Obviously, S and k are independent
˛i ıei . Furthermore
i D1 C1
P.VnC1 D ei jVn D 0/ D
X 1 P.VnC1 D ei ; Sn D 0; n 2 ŒTk ; TkC1 Œ/ P.Sn D 0/ kD0
C1
D
X 1 P.k D ei ; Sn D 0; n 2 ŒTk ; TkC1 Œ/ P.Sn D 0/ kD0
D ˛i This completes the proof of the proposition.
t u
Discrete Approximations to Solution Flows of Tanaka’s SDE
179
Remark 1. With the notations of Proposition 2, let .:Y / be the Markov chain defined by .:Y /n D i Y n on the ith excursion of Y and .:Y /n D 0 if Y n D 0. Then the stochastic Matrix of .:Y / is given by 1 ˛i ; M.0; ei / D ; M.nei ; .n C 1/ei / 2 2 1 D M.nei ; .n 1/ei / D ; i 2 Œ1; N ; n 2 N : 2
M.0; 0/ D
(9)
3 Proof of Main Results 3.1 Proof of Proposition 1 Let .Zt /t 0 be a W .˛1 ; ; ˛N / on G started at 0. For all i 2 Œ1; N , define Zti D jZt j1fZt 2Di g jZt j1fZt …Di g . Then Zti D ˆi .Zt / where ˆi .x/ D jxj1fx2Di g jxj1fx…Di g . Let Qi be the semigroup of the skew Brownian motion of parameter ˛i .SBM.˛i // (see [10] page 87). Then the following relation is easy to check: Pt .f ı ˆi / D Qti f ı ˆi for all bounded measurable function f defined on R. This shows that Z i is a SBM.˛i / started at 0. For n 1, i 2 Œ1; N , define 1 n T0n D 0; TkC1 D inffr 0 W d.Zr ; ZTkn / D p g; k 0: n 1 n;i T0n;i D 0; TkC1 D inffr 0 W jZri ZTi n;i j D p g; k 0: n k n;i n D TkC1 D inffr 0 W jjZr j jZTkn jj D Remark that TkC1
Zt 2 Di , then obviously d.Zt ; Zs / D
jZti
d.Zt ; Zs /
N X
Zsi j
p1 g. n
Furthermore if
for all s 0 and consequently
jZti Zsi j:
(10)
i D1
Now define Zkn D
p p law nZTkn ; Zkn;i D nZ i n;i . Then .Zkn ; k 0/ D M (see the Tk
proof of Proposition 2 in [5]). For all T > 0, we have X 1 n 1 n;i sup jZti p Zbnt ! 0 in probability sup d.Zt ; p Zbnt c/ cj n ! C1 n n t 2Œ0;T t 2Œ0;T i D1 N
by Lemma 4.4 [1] which proves our result. Remark 2. (1) By (10), a.s. t 7! Zt is continuous. We will always suppose that Walsh Brownian motion is continuous.
180
H. Hajri
(2) By combining the two propositions 1 and 2, we deduce that .:Y / rescales as Walsh Brownian motion in the space of continuous functions. It is also possible to prove this result by showing that the family of laws is tight and that any limit process along a subsequence is the Walsh Brownian motion.
3.2 Scaling Limits of .‰; K / Set p;n D e.‰p;n / for all p n where ‰p;n D ‰p;n .0/: C . Proposition 3. (i) For all p n; j‰p;n j D Sp;n (ii) For all p < n < q,
P.p;q D n;q jminh2Œp;q Sh D minh2Œn;q Sh / D 1 and C > 0 8j 2 Œn; q/ D 1: P.p;n D p;q jminh2Œp;n Sh D minh2Œp;q Sh ; Sp;j
(iii) Set Tp;x D inffq p W Sq Sp D jxjg. Then for all p n, x 2 GN , ‰p;n .x/ D .x C e.x/Sp;n /1fn Tp;x g C ‰p;n 1fn>Tp;x g I Kp;n .x/ D EŒı‰p;n.x/ j.S / D ıxCe.x/Sp;n 1fnTp;x g C
N X i D1
˛i ıSp;n 1 : C ei fn>Tp;x g
Proof. (i) We take p D 0 and prove the result by induction on n. For n D 0, this C is clear. Suppose the result holds for n. If ‰0;n 2 G , then S0;n > 0 and so minh2Œ0;n Sh D minh2Œ0;nC1 Sh . Moreover ‰0;nC1 D ‰0;n C 0;n Sn;nC1 D .SnC1 C C C minh2Œ0;n Sh /0;n D S0;nC1 0;n . If ‰0;n D 0, then S0;n D 0 and j‰0;nC1 j D Sn;nC1 . C But minh2Œ0;nC1 Sh D min.minh2Œ0;n Sh ; SnC1 / D min.Sn ; SnC1 / since S0;n D 0 which proves (i). C C (ii) Let p < n < q. If minh2Œp;q Sh D minh2Œn;q Sh , then Sp;q D Sn;q . When C C Sp;q D 0, we have p;q D n;q D eN by convention. Suppose that Sp;q > 0, then clearly C C J WD supfj < q W Sp;j D 0g D supfj < q W Sn;j D 0g: By the flow property of ‰, we have ‰p;q D ‰n;q D ‰J;q . The second assertion of (ii) is also clear. (iii) By (i), we have ‰p;n D ‰p;n .x/ D 0 if n D Tp;x and so ‰p;: .x/ is given by ‰p;: after Tp;x using the cocycle property. The last claim is easy to establish. t u For all s 2 R, let ds (resp. d1 ) be the distance of uniform convergence on every compact subset of C.Œs; C1Œ; G/ (resp. C.R; R/). Denote by D D fsn ; n 2 Ng
Discrete Approximations to Solution Flows of Tanaka’s SDE
181
e D C.R; R/ the set of all dyadic numbers of R and define C
C1 Y
C.Œsn ; C1Œ; G/
nD0
equipped with the metric: d.x; y/ D d1 .x 0 ; y 0 / C
C1 X nD0
1 inf.1; dsn .xn ; yn // where x D .x 0 ; xs0 ; /; 2n y D .y 0 ; ys0 ; /: .n/
Let t 7! S.t/ be the linear interpolation of S on R and define St p1 S.nt/; n 1. If u 0, we define buc D buc. Then, we have n .n/
St
D
1 1 D Stn C o. p /; with Stn WD p Sbnt c: n n .n/
n n Let ‰s;t D ‰s;t .0/ (defined in Corollary 1). Then ‰s;t WD and we have the following
p1 ‰bnsc;bnt c n
C o. p1n /
e . Then .Pn ; n 1/ is Lemma 1. Let Pn be the law of Z n D .S: .n/ ; .‰si ;: /i 2N / in C tight. .n/
Proof. By Donsker theorem PS .n/ ! PW in C.R; R/ as n ! 1 where PW is the law of any Brownian motion on R. Let PZsi be the law of any W .˛1 ; ; ˛N / started at 0 at time si . Plainly, the law of ‰p;pC: is given by (9) and so by Propositions 1 and 2, for all i 2 N, P‰.n/ ! PZsi in C.Œsi ; C1Œ; G/ as n ! 1. Now the lemma si ;: holds using Proposition 2.4 [3] (page 107). t u law e . In the next paragraph, Fix a sequence .nk ; k 2 N/ such that Z nk ! Z in C k ! C1
we will describe the law of Z. Notice that .‰p;n /pn and S can be recovered from .Z nk /k2N . Using Skorokhod representation theorem, we may assume that Z is defined on the original probability space and the preceding convergence holds almost surely. Write Z D .W; s1 ;: ; s2 ;: ; /. Then, .Wt /t 2R is a Brownian motion on R and . s;t /t s is an W .˛1 ; ; ˛N / started at 0 for all s 2 D.
3.2.1 Description of the Limit Process Set s;t D e. we have:
s;t /; s
2 D; s < t and define minu;v D minr2Œu;v Wr , u v 2 R. Then,
Proposition 4. (i) For all s t; s 2 D, j (ii) For all s < t; u < v; s; u 2 D,
s;t j
D Ws;tC .
P.s;t D u;v jmins;t D minu;v / D 1 if P.mins;t D minu;v / > 0:
182
H. Hajri
Proof. (i) is immediate from the convergence of Z nk towards Z and Proposition 3 (i). (ii) We first prove that for all s < t < u, P.s;u D t;u jmins;u D mint;u / D 1 if s; t 2 D
(11)
P.s;t D s;u jmins;t D mins;u / D 1 if s 2 D:
(12)
and Fix s < t < u with s; t 2 D and let show that a.s. fmins;u D mint;u g f9k0 ; bnk sc;bnk uc D bnk t c;bnk uc for all k k0 g:
(13)
We have fmins;u D mint;u g D fmins;t < mint;u g a:s: By uniform convergence the last set is contained in f9k0 ;
min
Sj <
min
Sj D
bnk scj bnk t c
min
Sj for all k k0 g
min
Sj for all k k0 g:
bnk t cj bnk uc
which is a subset of f9k0 ;
bnk scj bnk uc
bnk t cj bnk uc
This gives (13) using Proposition 3 (ii). Since x ! e.x/ is continuous on G , on fmins;u D mint;u g, we have 1 1 s;u D lim e. p ‰bnk sc;bnk uc / D lim e. p ‰bnk t c;bnk uc / D t;u a:s: k!1 k!1 nk nk which proves (11). If s 2 D; t > s and mins;t D mins;u , then s and t are in the C > 0 for all r 2 Œt; u. As preceded, same excursion interval of Ws;C and so Ws;r fmins;t D mins;u g is a.s. included in f9k0 ;
min
bnk scj bnk t c
Sj D
min
bnk scj bnk uc
C Sj ; Sbn > 0 8j 2 Œbnk t c; bnk uc; k k0 g: k sc;j
Now it is easy to deduce (12) using Proposition 3 (ii). To prove (ii), suppose that s u; mins;t D minu;v . There are two cases to discuss, (a) s u v t, (b) s u t v (in any other case P.mins;t D minu;v / D 0). In case (a), we have mins;t D minu;v D minu;t and so s;t D u;t D u;v by (11) and (12). Similarly in case (b), we have s;t D u;t D u;v . t u Proposition 5. Fix s < t; s 2 D; n 1 and f.si ; ti /I 1 i ng with si < ti ; si 2 D. Then: (i) s;t is independent of .W /. (ii) For all i 2 Œ1; N , h 2 Œ1; n, we have
Discrete Approximations to Solution Flows of Tanaka’s SDE
183
EŒ1fs;t Dei g j.si ;ti /1i n ; W D 1fsh ;th Dei g on fmins;t D minsh ;th g: (iii) The law of s;t knowing .si ;ti /1i n and W is given by
N X
˛i ıei when mins; t …
i D1
fminsi ;ti I 1 i ng.
e independently of .nk ; k 2 N/ This entirely describes the law of .W; s; ; s 2 D/ in C law e. and consequently Z n ! Z in C n ! C1
Proof. (i) is clear. (ii) is a consequence of Proposition 4 (ii). (iii) Write fs; t; si ; ti ; 1 i ng D frk ; 1 k mg with rj < rj C1 for all 1 j m 1. Suppose that s D ri ; t D rh with i < h. Then a.s. fminrj ;rj C1 ; i j h 1g are distinct and it will be sufficient to show that s;t is independent of ..si ;ti /1i n ; W / conditionally to A D fmins;t D minrj ;rj C1 ; mins;t ¤ minsi ;ti for all 1 i ng for j 2 Œi; h 1. On A, we have s;t D rj ;rj C1 , fminsi ;ti ; 1 i ng fminrk ;rkC1 ; k ¤ j g and so fsi ;ti ; 1 i ng frk ;rkC1 ; k ¤ j g. Since r1 ;r2 ; ; rm1 ;rm ; W are independent, it is now easy to conclude. t u In the sequel, we still assume that all processes are defined on the same probability a.s. e . In particular 8s 2 D; T > 0, space and that Z n ! Z in C n ! C1
1 j p ‰bksc;bk t c k!C1 st sCT k lim
sup
s;t j
D 0 a:s:
(14)
3.2.2 Extension of the Limit Process For a fixed s < t, mins;t is attained in s; tŒ a.s. By Proposition 4 (ii), on a measurable set ˝s;t with probability 1, 0 lim 0 s 0 ;t exists. Define "s;t D 0 lim 0 s 0 ;t on s !sC;s 2D
s !sC;s 2D
c . Now, let 's;t D "s;t Ws;tC . Then for all ˝s;t and give an arbitrary value to "s;t on ˝s;t s 2 D; t > s, ."s;t ; 's;t / is a modification of .s;t ; s;t /. For all s 2 R, t > s; 's;t D n lim 's ;t a:s:, where sn D b2 2scC1 and therefore .'s;t /t s is an W .˛1 ; ; ˛N / n n!1 n started at 0. Again, Proposition 4 (ii) yields
8s < t; u < v; P."s;t D "u;v jmins;t D minu;v / D 1 if P.mins;t D minu;v / > 0: (15) Define: 's;t .x/ D .x C e.x/Ws;t /1ft s;x g C 's;t 1ft >s;x g ; s t; x 2 G; where Ws;t D Wt Ws and s;x is given by (4).
184
H. Hajri
Proposition 6. Let x 2 G; xn 2 p1n GN ; limn!1 xn D x, s 2 R; T > 0. Then, we have p 1 sup j p ‰bnsc;bnt c . nxn / 's;t .x/j D 0 a:s: lim n!C1 st sCT n Proof. Let s 0 be a dyadic number such that s < s 0 < s C T . By (15), for t > s 0 : fmins;t D mins 0 ;t g f's;t D 's 0 ;t g a:s: and so, a.s. 8t > s 0 ; t 2 DI fmins;t D mins 0 ;t g f's;t D 's 0 ;t g: If t > s 0 ; mins;t D mins 0 ;t and tn 2 D; tn # t as n ! 1, then mins;tn D mins 0 ;tn which entails that 's;tn D 's 0 ;tn and therefore 's;t D 's 0 ;t by letting n ! 1. This shows that a.s. 8t > s 0 I fmins;t D mins 0 ;t g f's;t D 's 0 ;t g : As a result a.s. 8s 0 2 D\s; s C T Œ; 8t > s 0 I fmins;t D mins 0 ;t g f's;t D 's 0 ;t g:
(16)
By standard properties of Brownian paths, a.s. mins;sCT … fWs ; WsCT g and 8p 2 N I mins;sC 1 < Ws ; mins;sC 1 ¤ WsC 1 ; 9Šup 2s; s C p
p
p
1 ŒW mins;sC 1 D Wup : p p
The reasoning below holds almost surely: Take p 1; mins;sC 1 > mins;sCT . Let p
Sp 2s; s C p1 Œ: mins;sC 1 D WSp and s 0 be a (random) dyadic number in s; Sp Œ. p Then mins;s 0 > mins 0 ;t for all t 2 ŒSp ; s C T . By uniform convergence: Sbnuc 9n0 2 N W 8n n0 ; 8Sp t s C T; min0 Sbnuc > min 0 u2Œs;s
u2Œs ;t
and so ‰bns 0 c;bnt c D ‰bnsc;bnt c : Therefore for n n0 , we have 1 1 j p ‰bnsc;bnt c 's;t j D sup j p ‰bns 0 c;bnt c 's 0 ;t j .using (16)) n n Sp t sCT Sp t sCT sup
and so
Discrete Approximations to Solution Flows of Tanaka’s SDE
185
1 j p ‰bnsc;bnt c 's;t j n st sCT sup
1 1 sup j p ‰bnsc;bnt c 's;t j C sup j p ‰bnsc;bnt c 's;t j n n st Sp Sp t sCT
1 C 1 C Ws;tC / C sup j p ‰bns 0 c;bnt c 's 0 ;t j sup . p Sbnsc;bnt c n n 1 Sp t sCT st sC p
1 C 1 C Ws;tC / C sup j p ‰bns 0 c;bnt c 's 0 ;t j: sup . p Sbnsc;bnt c 0 0 n n 1 s t s CT st sC p
1 j p ‰bnuc;bnt c 'u;t j D 0. By letting n go n!C1 ut uCT n to C1 and then p go to C1, we obtain From (14), a.s. 8u 2 D; lim
lim
sup
sup
n!1 st sCT
We now show that
1 j p ‰bnsc;bnt c 's;t j D 0 a:s: n
1 Tbnsc;pnxn D s;x a:s: n!C1 n
(18)
lim
We have
(17)
bnsc 1 p T W Srn Ssn D jxn jg: D inffr n bnsc; nxn n
For > 0, from lim
n!1 u2Œ
sup
s;x ;s;x C
j.Sun Ssn C jxn j/ .Ws;u C jxj/j D 0;
we get lim
inf
n!1 u2Œs;x ;s;x C
.Sun Ssn C jxn j/ D
inf
u2Œs;x ;s;x C
.Ws;u C jxj/ < 0
which implies n1 Tbnsc;pnxn < s;x C for n large. If x D 0, n1 Tbnsc;pnxn bnsc n entails obviously (18). If x ¤ 0, then working in Œs; s;x as before and using infu2Œs;s;x .Wu Ws C jxj/ > 0, we prove that n1 Tbnsc;pnxn s;x for n large which establishes (18). Now p 1 1;n 2;n j p ‰bnsc;bnt c . nxn / 's;t .x/j sup Qs;t C sup Qs;t (19) n st sCT st sCT st sCT sup
where
186
H. Hajri 1;n Qs;t D j.xn C e.xn /.Stn Ssn //1fbnt cTbnsc;pnxn g .x C e.x/Ws;t /1ft s;x g j;
1 2;n Qs;t D j p ‰bnsc;bnt c 1fbnt c>Tbnsc;pnxn g 's;t 1ft >s;x g j: n By (17), (18) and the convergence of p1n Sbn:c towards W on compact sets, the righthand side of (19) converges to 0 when n ! C1. t u Remark 3. From the definition of "s;t (or Proposition 6), it is obvious that "r1 ;r2 ; ; "rm1 ;rm ; W are independent for all r1 <
ug; 's;t .x/ D "s;t Ws;tC and C D 's;u .x/: 't;u ı 's;t .x/ D 't;u ."s;t Ws;tC / D "s;t .Ws;tC C Wt;u / D "s;u Ws;u
since in this case mins;u D mins;t which implies "s;u D "s;t and C Ws;u D Wu mins;u D Wu Ws C Ws mins;t D Ws;tC C Wt;u :
Discrete Approximations to Solution Flows of Tanaka’s SDE
187
Thus we have, a.s. 's;u .x/ D 't;u ı 's;t .x/ which proves the cocycle property for '. It is now easy to check that ' is a stochastic flow of mappings in the sense of Definition 4 [5]. Note that .'0;t ; t 0/ is an W .˛1 ; ; ˛N / started at 0 and therefore satisfies Freidlin-Sheu formula (Theorem 3 [5]). Let f 2 D.˛1 ; ; ˛N /, then for all t 0, Z
t
f .'0;t / D f .0/ C
f 0 .'0;u /dBu C
0
1 2
Z
t
f 00 .'0;u /d u a:s:
0
where Bt D j'0;t j LQ t .j'0;: j/ and LQ t .j'0;: j/ is the symmetric local time at 0 of j'0;: j. Since j'0;t j D Wt min0;t , we get Bt D Wt . Let x 2 Di n f0g and fi .r/ D f .rei /; r 0. Since limz!0;z2Di ;z¤0 f 0 .z/ and limz!0;z2Di ;z¤0 f 00 .z/ exist, we can construct g which is C 2 on R and coincides with fi on RC . By Itô’s formula Z
t
g.jxj C Wt / D g.jxj/ C 0
1 g .jxj C Wu /d Wu C 2 0
Z
t
g 00 .jxj C Wu /d u a:s:
0
and so for t 6 0 .x/, we have Z
t
f .'0;t .x// D f .x/ C
f 0 .'0;u .x//d Wu C
0
R 1 0 .x/
R 0 .x/
f 0 .'0;u /d Wu C 2 0 Set ˛ D f .0/ C 0 C since W0; D 0. Then for t > 0 .x/, write 0 .x/ Z
1 2
Z
t
f 00 .'0;u .x//d u a:s:
0
f 00 .'0;u /d u D f .'0;0 .x/ / D f .0/
Z 1 t f .'0;t .x// D f .'0;t / D ˛ C f .'0;u /d Wu C f 00 .'0;u /d u 2 0 .x/ 0 .x/ Z t Z 1 t 0 D f .0/ C f .'0;u .x//d Wu C f 00 .'0;u .x//d u: 2 0 .x/ 0 .x/ t
0
R .x/ R .x/ But f .x/ C 0 0 f 0 .'0;u .x//d Wu C 12 0 0 f 00 .'0;u .x//d u D f .'0;0 .x/ .x// D f .0/ and so, for all t 0; f 2 D.˛1 ; ; ˛N /; x 2 G, Z
t
f .'0;t .x// D f .x/ C 0
1 f .'0;u .x//d Wu C 2 0
Z
t
f 00 .'0;u .x//d u a:s:
(20)
0
Now, let . ; W / be a any flow of mappings solution of .T /. Lemma 6 [5] implies 0;t .x/
D x C e.x/W0;t for 0 t 0;x with 0;x given by (4):
(21)
By considering a sequence .xk /k0 converging to 1, this shows that .Wt / . 0;t .y/; y 2 G/. Therefore, we can define a Wiener stochastic flow K obtained by filtering ı with respect to .W / (Lemma 3-2 (ii) in [6]) satisfying: 8s t,
188
H. Hajri
x 2 G; Ks;t .x/ D EŒı s;t .x/ j.W / a:s. In particular K solves .T / and since W K given by (3) is the unique Wiener solution of .T /, we get: 8s t; x 2 G; W W Ks;t .x/ D EŒı s;t .x/ j.W / a:s: (see Proposition 8 [5]). As K0;t .0/ is supported on C C fW0;t ei ; 1 i N g, we deduce that j 0;t .0/j D W0;t . Combining this with (21), we see that inffr 0 W 0;r .x/ D 0;r .0/g D 0;x :
This implies
0;r .x/
D
0;r .0/
for all r 0;x by applying the following
Lemma 2. For all .x1 ; ; xn / 2 G n ; denote by Px1 ; ;xn the law of . 0;: .x1 /; ; n 0;: .xn // in C.RC ; G /. Let T be a finite .Ft / stopping time where Ft D . 0;u ; u t/; t 0. Then the law of . 0;T C: .x1 /; ; 0;T C: .xn // knowing FT is given by P 0;T .x1 /; ; 0;T .xn / . C Note that W0;: can be recovered out from W0;: and consequently 0;: .x/ is a measurable function of 0;: .0/ for all x 2 G. Therefore, for all .x1 ; ; xn / 2 G n , . 0; .x1 /; ; 0; .xn // is unique in law since 0; .0/ is a Walsh Brownian motion. This completes the proof. t u
3.2.3 The Wiener Flow W Remark that Ks;t .x/ D EŒı's;t .x/ j.W / which entails that K W is a stochastic flow of kernels. By conditioning with respect to .W / in (20), we easily see that .K W ; W / solves .T /. In order to finish the proof of Theorem 2 and Corollary 1, we need only check the following lemma (the proof of (6) is similar)
Lemma 3. Under the hypothesis of Proposition 6, we have p W n .x/; Ks;t . nxn // ! 0 a.s. sup ˇ.Ks;t n ! C1
t 2Œs;sCT
jg.x/ g.y/j 1; g.0/ D 0. Then, jx yj x6Dy
Proof. Let g W G ! R such that kgk1 C sup
ˇZ ˇ Z ˇ ˇ p n ˇ g.y/K W .x/.dy/ g.y/Ks;t . nxn /.dy/ˇˇ Vs;t1;n C Vs;t2;n s;t ˇ G
G
where ˇ ˇ ˇ ˇ n /1fbnt cTbnsc;pnxn g g.x C e.x/Ws;t /1ft s;x g ˇ ; Vs;t1;n D ˇg.xn C e.xn /Ss;t Vs;t2;n D
N X j D1
ˇ ˇ ˇ ˇ C ˛j ˇg.ej Ws;tC /1ft >s;x g g.ej Sn;s;t C on /1fbnt c>Tbnsc;pnxn g ˇ
and on 2 G is a .S / measurable random variable such that jon j C Stn Ssn ; Sn;s;t D
p1 S C . n bnsc;bnt c
p1 , n
n Ss;t D
As bxc 1 x bxc C 1 for all x 2 R, we get
Discrete Approximations to Solution Flows of Tanaka’s SDE
189
.n/
Vs;t1;n sup jxn C e.xn /Ss;t x e.x/Ws;t j t 2In;s;x
C
.n/
sup jg.xn C e.xn /Ss;t /j C sup jg.x C e.x/Ws;t /j
t 2Jn;s;x
t 2Kn;s;x
(22)
with 1 1 C Tbnsc;pnxn /; n n 1 1 D Œs;x ; . Tbnsc;pnxn C / _ s;x ; n n 1 1 D Œs;x ^ . Tbnsc;pnxn /; s;x : n n
In;s;x D Œs; s;x _ . Jn;s;x Kn;s;x
Using jg.y/j jyj, we obtain .n/
sup jg.xn C e.xn /Ss;t /j C sup jg.x C e.x/Ws;t /j
t 2Jn;s;x
t 2Kn;s;x
.n/
sup jjxn j C Ss;t j C sup jjxj C Ws;t j: t 2Jn;s;x
Since
lim 1 Tbnsc;pnxn n!C1 n
t 2Kn;s;x
D s;x a.s., the right-hand side converges to 0. By dis-
cussing the cases x D 0; x ¤ 0, we easily see that limn!1 sup jxn C e.xn / t 2In;s;x
.n/ Ss;t
x e.x/Ws;t j D 0 and therefore
manner, we arrive at limn!1
limn!1 sup Vs;t1;n t 2Œs;sCT
D 0. By the same
sup Vs;t2;n D 0 which proves the lemma.
t u
t 2Œs;sCT
Acknowledgements I sincerely thank Yves Le Jan, Olivier Raimond and Sophie Lemaire for very useful discussions. I am also grateful to the referee for his helpful comments.
References 1. A.S. Cherny, A.N. Shiryaev, M. Yor, Limit behaviour of the “horizontal-vertical” random walk and some extensions of the Donsker-Prokhorov invariance principle. Teor. Veroyatnost. i Primenen 47(3) 498–517 (2002) 2. N. Enriquez, Y. Kifer, Markov chains on graphs and Brownian motion. J. Theoret. Probab. 14(2) 495–510 (2001) 3. S. Ethier, T. Kurtz, Markov Processes: Characterization and Convergence, Wiley Series in Probability and Mathematical Statistics, (Wiley, New York, 1986) 4. J.M. Harrison, L.A. Shepp, On skew Brownian motion. Ann. Probab. 9(2) 309–313 (1981) 5. H. Hajri, Stochastic flows related to Walsh Brownian motion. Electron. J. Probab. 16, 1563– 1599 (2011)
190
H. Hajri
6. Y. Le Jan, O. Raimond, Flows, coalescence and noise. Ann. Probab. 32(2), 1247–1315 (2004) 7. Y. Le Jan, O. Raimond, Flows associated to Tanaka’s SDE. ALEA Lat. Am. J. Probab. Math. Stat. 1, 21–34 (2006) 8. I.I. Nishchenko, Discrete time approximation of coalescing stochastic flows on the real line. Theory Stoch. Proc. 17(33)(1), 70–78 (2011) 9. P. Révész, Random Walk in Random and Non-random Environments. 2nd ed. (World Scientific Publishing, 2005) 10. D. Revuz, M. Yor, Continuous Martingales and Brownian Motion, 3rd ed. (Grundlehren der Mathematischen Wissenschaften, 1999) 11. S. Watanabe, The stochastic flow and the noise associated to Tanaka stochastic differential equation. Ukraïn. Mat. Zh. 52(9), 1176–1193 (2000)
Spectral Distribution of the Free Unitary Brownian Motion: Another Approach Nizar Demni and Taoufik Hmidi
Abstract We revisit the description provided by Ph. Biane of the spectral measure of the free unitary Brownian motion. We actually construct for any t 2 .0; 4/ a Jordan curve t around the origin, not intersecting the semi-axis Œ1; 1Œ and whose image under some meromorphic function ht lies in the circle. Our construction is naturally suggested by a residue-type integral representation of the moments and ht is up to a Möbius transformation the main ingredient used in the original proof. Once we did, the spectral measure is described as the push-forward of a complex measure under a local diffeomorphism yielding its absolute-continuity and its support. Our approach has the merit to be an easy yet technical exercise from real analysis.
1 Reminder and Motivation In his pioneering paper [1], Ph. Biane defined and studied the so-called free unitary or multiplicative Brownian motion. It is a unitary operator-valued Lévy process with respect to the free multiplicative convolution of probability measures on the unit circle T (or equivalently the multiplication of unitary operators that are free in some non commutative probability space). Besides, the spectral distribution t at any time t 0 is characterized by its moments Z mn .t/ WD
z dt .z/ D e n
T
nt =2
n1 X .t/k kD0
kŠ
k1
n
! n ; n 1; kC1
N. Demni () T. Hmidi IRMAR, Université de Rennes 1, Campus de Beaulieu, 35042 Rennes cedex, France e-mail: [email protected]; [email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__9, © Springer-Verlag Berlin Heidelberg 2012
191
192
N. Demni and T. Hmidi
and mn .t/ D mn .t/; n 1 since Y 1 defines a free unitary Brownian motion too. This alternate sum is not easy to handle analytically since for instance if we try to work out the moments generating function of .mn .t//n1 Mt .w/ WD
X
mn .t/wn ; jwj < 1;
n1
then we are led to Mt .w/ D u
X .ut/k k0
where Sk .u/ WD
X
.n C k C 1/
n0
k1
kŠ
Sk .u/
! nCk C1 n u ; kC1
k 0;
and u D we t =2 . Nonetheless, the explicit inverse function of Z t .w/ WD
T
zCw t .d z/ D 1 C 2Mt .w/ zw
in the open unit disc played a key role in the description of t [2]. More precisely, it was proved there that t is an absolutely continuous probability measure with respect to the normalized Haar measure on T and that its density is a real analytic function inside its support. The latter coincides with T when t > 4 while it is given by the angle p jj ˇ.t/ WD .1=2/ t.4 t/ C arccos.1 .t=2// when t 4. When proving these important results, the author relied on free stochastic integration (Lemma 11 p. 266), Caratheodory’s extension Theorem for Riemann maps (Lemma 12 p. 270) and a Poisson-type integral representation for this kind of maps (see the proof of Proposition 10 p. 270). In the present paper, we shall recover Biane’s results from simpler considerations than the ones used in the original proof. Indeed, for t 2 .0; 4/, there exists a unique piecewise smooth Jordan curve t around the origin, not intersecting the semi-axis Œ1; 1Œ and whose image under some function ht lies in T. Our construction is naturally suggested by a residue-type integral representation of mn .t/ and fails when t 4. Note that the same phenomenon happens here and in Biane’s proof: t is constructed upon two curves that have a non empty intersection if and only if t < 4, while the inverse function of t is defined on the interior s of two Jordan domains whose boundaries have the same phase transition [2, p. 267]. Moreover, the function ht appears in the integrand of our residue-type representation and coincides up to the Möbius transformation 2 z 7! 1 z
Spectral Distribution of the Free Unitary Brownian Motion: Another Approach
193
with the inverse function of t used in the original proof. Once the curve t is constructed, we consider a piecewise smooth parametrization zt of t and prove p that the derivative of 7! argŒht .zt .// vanishes if and only if D ˙ arccos. t =2/ p (note that similarly the derivative of t1 vanishes if and only if z D ˙i .4=t/ 1, [2] p. 269). As a matter of fact, 7! argŒht .zt .// defines far from the critical points a local diffeomorphism so that performing local changes of variables in our integral representation yields both the absolute-continuity of t with respect to the Haar measure on T and the description of its support.
2 A Residue-Type Representation of mn .t/ A residue-type integral representation of mn .t/; n 1 is not new in its own. Indeed, the one we derive below is an elaborated version of the one used in order to determine the decay order of mn .t/ (see [3] p. 566): mn .t/ D
e nt =2 2i n
Z
1 n e nt z 1 C dz z
where is a circle around the origin. More precisely, we need to integrate by parts since then the meromorphic integrand we obtain determines in a unique way the required curve t we mentioned above. To proceed, we first apply Cauchy’s Residues Theorem to z 7! zk e n=z in order to get 1 nkC1 D .k C 1/Š 2i
Z zk e n=z d z;
then use the combinatorial identity for binomial numbers: ! ! n n1 n D ; kC1 kC1 k
n 1:
As a result, for any n 1 mn .t/ D e nt =2
n1 X .t/k kD0
kŠ
nk1
! n kC1
! Z X n1 n1 .tz/k e n=z d z k kD0 Z 1 dz D .1 tz/n e n.1=zt =2/ 2i n 1 tz e nt =2 D 2i n
194
N. Demni and T. Hmidi
1 D 2i n 1 2i n
D WD
Z Z
1 2i n
n
dz t.1 z/
n .1 z/e t .1=z1=2/
dz t.1 z/
.1 z/e t .1=z1=2/
t
Z
Œht .z/n
dz : t.1 z/
Now, choose further such that it does not meet the semi-axis Œ1; 1Œ then z 7! log.1 z/ is well defined for z 2 and is holomorphic there. Hence, setting z D re i ; 0 < r < 1 and integrating by parts yield mn .t/ D
1 2i t
Z Œht .z/n
h0t .z/ log.1 z/d z: ht .z/
As a matter of fact, mn .t/ is the residue of z 7!
h0 .z/ 1 Œht .z/n t log.1 z/ t ht .z/
at z D 0 so that one may integrate along any piecewise smooth Jordan curve t (possibly depending on t) around zero provided that the integrand is well defined. Assume further that we can choose t such that jht .t /j 2 T and let 2 Œ; 7! zt ./ be a piecewise smooth parametrization of t , then Z h0 .zt .// 1 log.1 zt .//z0t ./d e i n arg ht .zt . // t 2i t ht .zt .// Z D e i n t .d/
mn .t/ D
where t is the push-forward of d h0t .zt .// 0 zt ./ log.1 zt .//1Œ; ./ ht .zt .// 2i under the map 7! arg ht .zt .//. Heuristically, we are attempted to conclude that t D t however it is not clear at all that t is a real measure and there is no guarantee even for t to exist. Below, we shall prove that t exists if and only if t 2 .0; 4/ and is unique. Then t splits into two curves t1 [t2 where 7! argŒht .zt .// is a diffeomorphism from ti to its image, for i D 1; 2: As a result, a change of variables shows that t and t coincide since their trigonometric moments do, therefore t is absolutely continuous and its support is easily recovered as ht .t /.
Spectral Distribution of the Free Unitary Brownian Motion: Another Approach
195
3 Construction of the Curve t Our main result is stated as Proposition 1. Let t 2 .0; 4/; then there exists a unique (piecewise smooth) Jordan curve t such that: • ht .t / 2 T. • t encircles z D 0 and t \ Œ1; 1Œ D ;. Proof. Before coming into computations, let us point to the fact that t has to be invariant under complex conjugation: this fact follows from jht .z/j D jht .z/j D jht .z/j. It is also coherent with the fact that t shares the same invariance property since Y 1 D Y ? is also a free unitary Brownian motion, therefore we only consider 2 Œ0; . We also inform the patient reader that both polar and cartesian coordinates are used in the sequel depending on how behaves the curve defined by jht .z/j D 1 when t runs over 0; 4Œ.
3.1 Polar Coordinates Let z D re i ; 2 Œ0; then jht .z/j D 1 is equivalent to gt; .r/ WD .1 C r 2 2r cos /e .2t cos /=r D e t : We distinguish two regions: f; cos < 0g: In this region the function gt; is increasing and satisfies lim gt; .r/ D 0;
r!0C
lim gt; .r/ D 1:
r!1
The monotonicity of gt; follows obviously from t cos 0 gt; .r/ D 2e 2t cos =r r cos .1 C r 2 2r cos / 2 : r
(1)
Then gt; .r/ D e t has a unique solution r D rt ./ > 0. Note that the implicit 0 function Theorem together with the fact that gt; .r/ > 0 show that 7! rt ./ is 1 at least C on =2; Œ. Now, it is obvious that rt does not vanish on fcos < 0g and we shall p check that it remains so on fcos D 0g. More precisely, we claim that rt ./ > t which may be proved as follows: use 1 C t < et to get
p p and 1 2 t cos < e 2 t cos
196
N. Demni and T. Hmidi p p p 1 C t 2 t cos .1 C t/.1 2 t cos / < e t 2 t cos :
This in turn yields
p gt; . t / < e t
and the monotonicity of gt; proves the claim. As a matter of fact rt extends continuously to =2 and one obviously has rt .=2/ D
p e t 1:
f; cos > 0}: For these values of , observe that gt; .2 cos / D e t for all t. However, rt ./ D 2 cos does not fulfill our requirements since on the one hand it vanishes at D ˙=2 and on the other hand it meets Œ1; 1Œ at D 0. Fortunately, there exists another radius rt p satisfying gt; .rt / D e t , rt .=2/ ¤ 0 and rt .0/ 20; 1Œ. Indeed, letting t WD arccos. t =2/ 20; =2, then 0 gt; .2 cos / D
4 cos2 t t e 2 cos
is negative on t ; =2Œ, positive on Œ0; t Œ and lim gt; .r/ D C1
r!1
for any p such that cos 0. Besides, this radius is unique except possibly for 2 C 3 < t < 4 and for close to zero, and we keep using polar coordinates. For exceptional values of .t; /, cartesian coordinates are more adequate and doing so we recover t as the p graph of some function. N f0 < t 2 C 3}: When t 20; 1, we shall prove that gt; is a convex function. To this end, we compute the second derivative of gt;
4t cos 4t 2 cos2 C 3 r r4 8t cos C2 .r cos / r2
00 gt; .r/ D e 2t cos =r
D
.1 C r 2 2r cos /
e 2t cos =r 4 2r 4t r 3 cos C 4t 2 r 2 cos2 r4 C4rt cos .1 2t cos2 / C 4t 2 cos2 :
00 The first equality shows that if r cos then gt; .r/ > 0. For r cos , define kt by e 2t cos =r 00 gt; .r/ D kt; .r/: r4 Then
Spectral Distribution of the Free Unitary Brownian Motion: Another Approach
197
0 kt; .r/ D 8r 3 12t r 2 cos C 8t 2 r cos2 C 4t cos .1 2t cos2 / 00 kt; .r/ D 8.3r 2 3t r cos C t 2 cos2 / 0
even for all r 0. Hence 0 0 .r/ kt; .t cos / D 4t cos .t 2 cos2 2t cos2 C 1/ kt;
D .t cos 1/2 C 2t cos .1 cos / 0: 0 Since t cos cos when t 1, then kt; .r/ 0 for all r 0 therefore
kt; .r/ kt; .0/ D 4t 2 cos2 > 0 which yields the strict convexity of gt; for t 2 Œ0; 1. Now since lim gt; .r/ D lim gt; .r/ D C1;
r!0
r!C1
then the equation gt; .r/ D e t admits exactly two solutions among them the trivial one r D r./ D 2 cos which has to be discarded. The required curve t is then constructed upon the non trivial solution and defines even a C 1 -piecewise curve. The last claim is obvious for the regular points of gt; by the virtue of the implicit function Theorem again. So we need to focus on the critical points of the 0 curve: gt; .r.// D 0. But, the strict convexity of gt; forces then rt ./ D 2 cos 0 which gives after substituting in gt; the unique critical point D t for which p p rt .t / D t: Before considering the range t 2 Œ1; 2 C 3, we point out that t ; t 20; 4Œ crosses the positive real semi-axis at p some point 0 < xt < 1 that is described in Lemma 1 below. Now, let t 2 Œ1; 2 C 3 and define vt; .r/ WD r 3 .t C 1/ cos r 2 C 2t cos2 r t cos so that 0 .r/ D gt;
e 2t cos =r vt; .r/: r2
Then the first derivative of vt; reads v0t; .r/ D 3r 2 2.t C 1/ cos r C 2t cos2 : The discriminant of this second degree polynomial is easily computed as 4.t 2 p 2 4t C 1/ cos and is easily seen to be negative on t 2 Œ1; 2 C 3. Consequently v0t; 0 thereby vt; .r/ vt; .0/ D t cos for any r 0. Finally, there exists 0 r0 D r0 .t; / such that gt; .r0 / D 0 so that the variations of gt; are described by
198
N. Demni and T. Hmidi
r vt; .r/
0
r0 C1 0 C
gt; .r/ C1 &
% C1
The conclusion p follows in similar way to the previous range of times t 20; 1: N 2 C 3 < t < 4: This part of the proof needs more care since for small amplitudes of , 7! rt ./ may be a multi-valued function (this is seen from computer-assisted pictures). This multivalence happens precisely in the interior of the region bounded by the curve 7! 2 cos or 0 < t , and cartesian coordinates are more adequate for our purposes. Nonetheless, we shall keep use of polar coordinates outside the latter curve where the existence of a required radius is ensured by the same arguments evoked above. It then remains to prove uniqueness on 2 cos ; C1Œ; 2t ; =2. To this end, one easily sees that the largest root of v0t; .r/ is given by RtC ./
WD
t C1C
p t 2 4t C 1 cos 3
and that t 7! RtC ./ is increasing. Thus, RtC ./ R4 ./ D 2 cos yielding v0t; .r/ > 0 for r > 2 cos : This entails vt; .r/ vt; .2 cos / D 4 cos .cos2 t=4/ which is negative on t ; =2. Therefore the variations of gt; are summarized below r 2 cos r0 C1 vt; .r/ 0 C gt; .r/
et
&
% C1
for some r0 D r0 .t; / > 2 cos , whence the uniqueness follows. Hence,p the p p obtained branch of t is smooth and its endpoints are i e t 1 and t e i arccos. t =2/ . ˚ Remark 1. The equation gt; .r/ D e t has no solution in the region r > 2 cos ; 0 < t : Indeed, vt; .r/ vt; .2 cos / > 0 so that gt; is increasing.
3.2 Cartesian Coordinates The curve 7! 2 cos ; 2 Œ0; =2 is the graph of the function x 7! y D p x.2 x/; x 2 Œ0; 2. Hence, we shall restrict ourselves to the region
Spectral Distribution of the Free Unitary Brownian Motion: Another Approach
n
0y<
199
o p x.2 x/; 0 < x < 2 :
Besides, the branch of t , if it exists, would meet the axis fy D 0g at a solution of kt .x/ WD .x 1/2 e t Œ.2=x/1 D 1; 0 < x < 2; whose needed properties are collected in the following Lemma. Lemma 1. For every t 20; 4Œ, the above equation admits a unique solution xt 2 p 0; 1Œ and the map t 7! xt is increasing. In particular xt > 3 5. Proof. Let t 20; 4Œ then kt0 .x/ D
2.x 1/.x 2 tx C t/ t Œ.2=x/1 e x2
and the polynomial x 2 tx C t is obviously positive since its discriminant is negative. As a result, we get the variations of kt x kt0 .x/
0
1
kt .x/ C1 & 0
2 C %1
This asserts the existence of a unique value xt 20; 1Œ solving the equation kt .x/ D 1. For the variations of t 7! xt , we write xt0 D
xt .1 xt /.2 xt / @t kt .xt / D >0 @x kt .xt / 2.xt2 txt C t/
so that xpt > x3 and the Lemma is proved by noting that k3 .3 1/2 e 3.1C 5/=2 1.
p p 5/ D . 5 t u
Now, we rewrite gt; .r/ D e t using cartesian coordinates as .1 C x 2 C y 2 2x/e 2tx=.x
2 Cy 2 /
D et
and denote gt;x .y/ the LHS. In this way, gt;x .0/ D kt .x/e t e t for x 2 Œxt ; 2 since kt .x/ < 1 for x 2 Œxt ; 2, while gt;x .0/ > e t for x … Œ0; xt Œ. We shall prove that for each x 2 Œxt ; t=2 oŒxt ; 2Œ, the equation gt;x .y/ D e t admits a unique n p solution 0 y < x.2 x/ while it has none when x …xt ; t=2Œ. This in turn finishes the proof of our main result. To proceed, we first compute
200
N. Demni and T. Hmidi
0 gt;x .y/ D
WD
2y 2 2 4 2 2 4 3 2 y C .2x 2xt/y C x 2x t C 4x t2xt e 2tx=.x Cy / .x 2 C y 2 /2 .x 2
2y 2 2 e 2tx=.x Cy / wt;x .y 2 /: 2 2 Cy /
The discriminant of wt;x .y/ is given by 4tx.2 C .t 4/x/ and is positive since t 3 and x 2. Thus wt;x .y/ has two roots yt˙ WD x.t x/ ˙
p tx.2 C .t 4/x/
satisfying the following properties: p Lemma 2. Let 2 C 3 < t < 4, then: 1. For x 2 Œxt ; 2; then 0 < yt .t/ ytC : 2. For x 2 Œ0; t=2 then yt .t/ x.2 x/ ytC . 3. For x 2 Œt=2; 2, then x.2 x/ yt ytC Proof. Since x 20; 2Œ, then ytC 0 and the first property .1/ is equivalent to ytC yt D x 4 2x.x 1/2 t > 0: But for any x > 0, the function t 7! x 4 2x.x 1/2 t is decreasing for positive t therefore p p x 4 2x.x 1/2 t > x 4 8x.x 1/2 D x.x 2/ x .3 5/ x .3 C 5/ p 4 2x.x 1/2 t > 0 for x 2 Œ3 5; 2 in particular for any t < 4. Consequently xp for x 2 Œxt ; 2 since xt > 3 5 by the virtue of Lemma 1. Since ytC > x.2 x/, then the second property .2/ is equivalent to x 2 .t 2/2 tx.2 C .t 4/x/ which is in turn equivalent to t 2x 0. We are done.
t u
It remains to discuss the variations of gt;x according to: ? x 2 Œxt ; t=2: Using properties .1/ and .2/ stated in Lemma 2 we get y 0 gt;x .y/
0
1
.yt / 2 .ytC / 2 C1 C C
gt;x .y/ gt;x .0/ %
&
1
%
C1
Spectral Distribution of the Free Unitary Brownian Motion: Another Approach
201
t Since gt;x .0/ e t ; then p the equation gt;x .y/ D e has a unique solution yt .x/ lying in the interval Œ0; x.2 x/Œ. This allows to construct a curve x 7! yt .x/ for x 2 Œxt ; t=2 which is continuous since the function gt;x depends continuously on the parameter x. By the virtue of the implicit function Theorem, it is even at least 0 C 1 -piecewise curve since the derivative gt;x .y/ vanishes only in a finite set. ? x 2 Œt=2; 2: Using properties .1/ and .3/ p pof Lemma 2, we conclude that gt;x is increasing on Œ0; x.2 x/. p Since gt;x . x.2 x// D e t then the equation gt;x .y/ D e t has no solution in Œ0; x.2 x/Œ: p ? x 20; xt Œ: Since xt 20; 1Œ and t=2 > 1 C . 3=2/ > xt , then we make use of property .2/ of Lemma 2. But the issue depends on whether p or not yt is positive. Assume yt is negative then gt;x is decreasing on Œ0; x.2 x/Œ and thus the equation gt;x .y/ D e t has no solution in this interval. Otherwise yt > 0 and gt;x keeps the same variations as in the range x 2 Œxt ; t=2:
y 0 gt;x .y/
p x.2 x/
1
0 C
gt;x .y/ gt;x .0/ %
.yt / 2
&
et
We remark that gt;x .0/ D e t k.t; x/ and according to the variations of x 7! k.t; x/ we get k.t; x/ > 1 for 0 < x < xt : Consequently the equation gt;x .y/ D e t has no solution in 0; xt Œ: ˚ Finally, the above discussion shows the set .x; y/; gt;x .y/ D e t ; x 20; 2Œ; 0 p y x.2 x/ is described by a unique C 1 -piecewise graph joining the points p p .xt ; 0/ and z D te i arccos. t=2/ : t u 0 .2 cos / D 0 if and only if D t D 0. Thus both curves Remark 2. For t D 4, gt; whose radii solve gt; .r/ D e t meet at D 0 thereby satisfy rt .0/ D 2 > 1. When t > 4, they even become disconnected.
4 Critical Points of ht Let zt be a piecewise smooth parametrization of t and consider 7! argŒht .zt .//; 2 Œ; . Using the invariance of t under complex conjugation, we restrict our attention to 2 Œ0; . If 2 f0; g then arg.1 z/ D 0; z 2 t since rt .0/ 20; 1Œ therefore argŒht .zt .0// D 0. Thus we discard these two values and consider 2 .0; /. Then arg.1 z/ 2 .; 0/ and we need to look for critical points of t r cos 1 argŒht .z/ D cot1 sin r sin r
202
N. Demni and T. Hmidi
under the constraint z D zt D rt ./e i 2 t \ CC . For ease of notations, we shall omit the dependence on t of the radius of t ; t 20; 4Œ and write simply r . Hence r cos r0 sin r 2 r cos r0 sin d argŒht .r e i / D 2 t d r 2r cos C 1 r2 which vanishes if and only if i h i h r r3 r2 cos t cos .r2 2r cos C 1/ Dr0 sin .r2 t .r2 2r cos C 1/ :
By the virtue of (1), the LHS may be written as 1 3 r @r .gt; /.r /e .2t cos /=r 2 while @ .g /.r / D
i 2 sin h 2 r t.r2 2r cos C 1/ e .2t cos /=r : r
It follows that the following equality holds at any critical point r2 @r .gt; /.r / D r 0 ./@ .gt; /.r /: Now comes the constraint gt; .r / D e t that we shall differentiate with respect to to get r 0 ./@r .gt; /.r / C @ .gt; /.r / D 0: Both identities yield r2 Œ@r .gt; /2 .r / D Œ@ .gt; /2 .r /: Since r ¤ 0 then a critical value must satisfy @r .gt; /.r / D @ .gt; /.r / D 0 which can not occur unless r D 2 cos : As a result D t and one easily derives using 2 cos t D
p t
1p t.4 t/ 2 p 1p t D 2 arccos t.4 t/: 2 2
argŒht .2 cos t e it / D 2t
Finally
Spectral Distribution of the Free Unitary Brownian Motion: Another Approach
203
# " p # p t t t D cos 2 arccos D1 cos 2 arccos 2 2 2 "
whence we deduce that 1p t t.4 t/ D ˇ.t/: argŒht .2 cos t e / D arccos 1 2 2 it
A similar analysis shows that 7! argŒht .zt .// has a unique critical point when zt 2 t \ C . It is precisely given by D t and argŒht .2 cos t e it / D ˇ.t/: We now proceed to the description of t ; t 2 .0; 4/.
5 Description of t ; t 2 .0; 4/ We have already seen that there are exactly two critical points of argŒht .zt .//; 2 Œ; . This fact leads easily to:
7!
Proposition 2. There exists a partition t D t1 [ t2 with t1 fz; jz 1j 1g; t2 fz; jz 1j 1g and such that the maps ht;1 ht W t1 ! ht .t / and ht;2 ht W t2 ! ht .t / are diffeomorphisms. Moreover, let e i 2 ht .t / then the equation ht .z/ D e i ; z 2 t has exactly two solutions given by z 2 t1 and z 2 t2 : z1 Proof. The curves t1 and t2 are given by ˚ t1 D z 2 t ; jz 1j 1 ˚ t2 D z 2 t ; jz 1j 1 : It is clear that the critical points are located in the circle fz; jz1j D 1g and therefore they are the end points of the curves t1 and t2 . Therefore by the previous analysis of 7! argŒht .zt .// we deduce that ht;1 ; ht;2 are diffeomorphisms. Finally, for any z 2 t 1 z D D ht .z/: ht z1 ht .z/
204
N. Demni and T. Hmidi
We have used in the last identity the fact that ht .z/ 2 T: We point out that the möbius z transform z 7! z1 is a bijective map from t;1 to t;2 : t u Thus we obviously have1 mn .t / D
1 2i t
Z t1
h0t .z/ 1 log.1 z/d z C ht .z/ 2i t
Œht .z/n
Z t2
Œht .z/n
h0t .z/ log.1 z/d z: ht .z/
We perform the change of variables in the first integral: ht;1 .z/ D ht .z/ D e i then i.d/ D
h0t .z/ d z: ht .z/
Since argŒht .zt .// reaches its minimum at D t , then 1 2i t
Z t1
Œht .z/n
1 h0t .z/ log.1 z/d z D ht .z/ 2 t
Z
ˇ.t / ˇ.t /
i e i n log.1 h1 t;1 .e //d:
Similarly we get 1 2i t
Z
0 n ht .z/
1 log.1 z/d z D Œht .z/ 2 h .z/ 2 t t t
Z
ˇ.t / ˇ.t /
i e i n log.1 h1 t;2 .e //d;
consequently mn .t/ D
1 2 t
Z
ˇ.t /
e i n log ˇ.t /
h 1 h1 .e i / i t;2
i 1 h1 t;1 .e /
d:
The last part of Proposition 2 shows that: i h1 t;1 .e / D
implying that
i 1 h1 t;2 .e / i 1 h1 t;1 .e /
i h1 t;2 .e /
!
i h1 t;2 .e / 1
i 2 D jh1 t;2 .e / 1j :
i Together with jh1 t;2 .e / 1j 1 yield
1 mn .t/ D t
1
Z
ˇ.t / ˇ.t /
i e i n log jh1 t;2 .e / 1j d:
t is parametrized from t to 2 counter-clockwise.
Spectral Distribution of the Free Unitary Brownian Motion: Another Approach
205
Thus mn .t/ D mn .t/ D
1 t
Z
ˇ.t / ˇ.t /
i e i n log jh1 t;2 .e / 1j d; n 1
whence we deduce that spectral measure t is given by dt ./ D
2 d i 1Œˇ.t /;ˇ.t / ./ log jh1 WD t ./d: t;2 .e / 1j t 2
Moreover, t is continuous since t .ˇ.t// D t .ˇ.t// D 0 which follows from ˙iˇt h1 / D 1 C e 2it . t;2 .e Remark 3. Set ZD
then ht .z/ D ht
2 ZC1
2 1 z
D
Z 1 t Z=2 e D t1 .Z/ ZC1
where the last equality holds for t WD f<.Z/ > 0; jt1 .Z/j < 1g D fjz 1j < 1; jht .z/j < 1g and extends to t (see [2]). It follows that h1 t;1 D
2 t C 1
on the closed unit disc and one derives 2 2 i 1 i log jh1 t;2 .e / 1j D log j1 ht;1 .e /j t t ˇ ˇ ˇ t .e i / 1 ˇ 2 ˇ D log ˇˇ t .e i / C 1 ˇ t
ˇ ˇ 2 i ˇ ˇ D log ˇe t t .e /=2 ˇ D <Œ.e i / t as stated in [2] p. 270. Acknowledgements This work was partially supported by Agence Nationale de la recherche grant ANR-09-BLAN-0084-01.
206
N. Demni and T. Hmidi
References 1. P. Biane, Free Brownian Motion, Free Stochastic Calculus and Random Matrices. Fields Institute Communications, 12, (American Mathematical Society Providence, RI, 1997), pp. 1–19 2. P. Biane, Segal-Bargmann transform, functional calculus on matrix spaces and the theory of semi-circular and circular systems. J. Funct. Anal. 144(1), 232–286 (1997) 3. T. Lévy, Schur-Weyl duality and the heat kernel measure on the unitary group. Adv. Math. 218(2), 537–575 (2008)
Another Failure in the Analogy Between Gaussian and Semicircle Laws Nathalie Eisenbaum
Abstract We establish a new characterization of the Gaussian law that is not transferable to the non commutative probability spaces. Keywords Gaussian law • Semicircle law • Free poisson distribution • Free probability theory • Free convolution • R-transform • Dynkin’s isomorphism theorem AMS 2000 Classification: 60J25, 60J55, 60G15, 60E07
1 Introduction and Main Results Dynkin’s isomorphism theorem and its variants relate the law of the local time process of a given transient symmetric Markov process with the law of a centered Gaussian process with covariance the Green function of this Markov process. We remind in particular the following version established in [4] for the local time process .Lxt ; x 2 E; t 0/ of a symmetric recurrent Markov process Z with state space E. Denote by o an element of E, by To its first hitting time by Z and set: .r/ D inffs 0 W Los > rg. The Green function of Z killed at To is a positive definite function on E E. Let .x ; x 2 E/ be a centered Gaussian process with covariance this Green function, independent of Z. We then have the following identity in law for every real r 1 .law/ 1 . .x C r/2 ; x 2 E/ D . 2x C Lx r 2 ; x 2 E/ . 2 / 2 2 N. Eisenbaum () CNRS: UMR7599, Université Paris 6, 4 Place Jussieu 75252, cedex Paris 05, France e-mail: [email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__10, © Springer-Verlag Berlin Heidelberg 2012
207
208
N. Eisenbaum
which can easily be reformulated as follows: for every a, b in R such that jbj jaj 1 .law/ 1 . .x C b/2 ; x 2 E/ D . .x C a/2 C Lx b2 a2 ; x 2 E/: . 2 / 2 2
(1)
One immediate consequence of this isomorphism theorem is the existence of a real random variable X such that for every a, b in R such that jbj jaj there exists a random variable ` depending on .a; b/ only via .b 2 a2 /, independent of X such that .law/ (A) .X C b/2 D .X C a/2 C ` : The following theorem states that the above property actually characterizes the centered Gaussian variables. Theorem 1. A real centered random variable X satisfies Property A if and only if X has a centered Gaussian law. Seeking for a free analogue of the isomorphism theorem, one asks first about one of its consequences: the free analogue of Property A. To enunciate the answer (for a complete exposure on free probability, one can consult [7]), we consider a non commutative probability space .A ; / such that A is a C*-algebra and W A ! C is a positive linear functional with .1/ D 1. An element x of A is called a random variable. A family of C*- subalgebras fAi ; i 2 I g of .A ; / is said to be free if .x1 x2 :::xn / D 0 whenever xj 2 Aij with i1 6D i2 6D ::: 6D in and .xj / D 0 for 1 j n. For fxi ; i 2 I g a family of random variables, fxi ; i 2 I g is said to be free if the family of the subalgebras generated respectively by f1; xi g is free. One associates to each random variable x a linear functional x on CŒX the algebra of polynomials, defined by x .P .X // D .P .x/// for any P 2 CŒX . A random variable x is said to be centered if x .X / D 0. If x is self-adjoint then x extends to a probability measure on R. In particular the semicircle law centered p at m and of radius r > 0 defined by its density 2r 2 1.mr;mCr/ r 2 .t m/2 dt, corresponds to an important non commutative random variable. Indeed the central limit theorem for free random variables holds with limit distribution equal to a semicircle law. Hence, in non commutative probability the part of the Gaussian law is played by the semicircle law. A priori, one would expect from the semicircle law to satisfy the free analogue of Property A. But we have the following proposition. Proposition 1. There is no non null centered non commutative variable x such that for every a, b in R such that jbj jaj, there exists a variable y depending of .a; b/ only via .b 2 a2 / such that fx; yg is free and .law/
.x C b/2 D .x C a/2 C y:
(2)
Proposition 1 shows that Theorem 1, minimal consequence of the isomorphism theorem, does not have a free analogue. Moreover Theorem 1 provides a characterization of the Gaussian law that is not working for the semi-circle law. This is
Another Failure in the Analogy Between Gaussian and Semicircle Laws
209
not the only example of such failure. In [2], Bercovici and Voiculescu show that the characterization of the Gaussian law given by Cramér Theorem is not transferable to the semi-circle law. In [1], Benaych-Georges shows similarly that Raikov’s characterization of the Poisson law is neither transferable to the free Poisson law. One consequence of Property A for a centered Gaussian variable X , is that for every real m, .X C m/2 is always infinitely divisible. In view of Proposition 1, one can then ask whether the free analogue of this weaker consequence of the isomorphism theorem holds for the semi-circle law. The answer is presented by the proposition below. Proposition 2. Let x be a non commutative variable with a centered semicircular law. Let m be a real number. Then .x Cm/2 is infinitely divisible for the free additive convolution if and only if m D 0. The original isomorphism theorem due to Dynkin holds for a symmetric transient Markov process Z with a state space E. Denote o an element of E, o its last visit by Z, Po the probability under which Z starts from o a.s. and .Lxt ; x 2 E; t 0/ the local times process of Z. Let .x ; x 2 E/ a centered Gaussian process with covariance the Green function of Z, independent of Z. Then under Po , we have: .Lxo C has the law of . 12
2 x; x
1 2
2 x; x
2 E/ 2
2 E/ under the probability E. E. o2 / ; :/. o
As for the previous isomorphism theorem, one deduces the existence of a nonnegative integrable real random variable X such that there exists an independent random variable ` verifying: .X C `/ has the law of X under E.
X ; :/: E.X /
(B)
But Property B does not characterize the squared centered Gaussian variables. Indeed, as an immediate consequence of Lemma 3.1 in [3], we have: A nonnegative random variable has Property B if and only if it is infinitely divisible. For Ra given non commutative variable x with distribution on RC such that / 0< RC t.dt/<1, define the variable T .x/ with distribution R t.dt . A natural s.ds/ RC free analogue of Property B for the Gaussian law would be the existence of a variable y such that fx; yg is free and .law/
T .x 2 / D x 2 C y
(3)
210
N. Eisenbaum
for x with a standard semicircular law. The distribution of the free variable x 2 is called the free Poisson distribution (or the Marchenko-Pastur distribution) with parameter 1. We show in Sect. 2 that such a variable does not exist. To do so, we will first note that .law/ T .x 2 / D x C 2 :
2 The Proofs We adopt the following notation. For x non commutative variable, its law is denoted by x . For X classical random variable, its law is denoted by L .X /. Proof of Theorem 1. Any centered Gaussian variable X satisfies Property A thanks to the identity (1). Conversely, let X be a random variable satisfying (A). One writes ` D `.b 2 a2 /. Since (A) gives: .law/
.X C a/2 D X 2 C `.a2 /; with `.a2 / independent of X , we obtain L ..X C b/2 / D L .X 2 C `.a2 // L .`.b 2 a2 //: Consequently for every a, c in R we have: L .`.a2 C c 2 // D L .`.a2 // L .`.c 2 //;
(4)
Now denote by X1 ; X2 ; :::; Xn , n independent copies of X . For every .ai /1i n element of Rn , we have thanks to (A) and (4) L.
n n X X .Xi C ai /2 / D L . Xi2 / L .`.a12 // ::: L .`.an2 //: i D1
Hence L.
i D1 n X i D1
.Xi C ai /2 / D L . Pn
n X
Xi2 / L .`.
i D1
n X
ai2 //;
i D1
which shows that the law of i D1 .Xi C ai / depends of .ai /1i n only via P . niD1 ai2 /. This is precisely the characterization of the Gaussian law established by Kagan and Shalayaevski [6]. t u 2
Proof of Proposition 1. Let x be a centered variable satisfying (2). We denote by y.b 2 a2 / the variable y of (2). One easily obtain for any real numbers a and b
Another Failure in the Analogy Between Gaussian and Semicircle Laws
211
y.a2 / y.b 2 / D y.a2 Cb 2 /
(5)
Hence y.a2 / is infinitely divisible for the free additive convolution. Now let fx1 ; x2 ; :::; xn g be a free family of self-adjoints elements identically distributed as x. Then, thanks to (2), we have for every .ak /1kn element of Rn PnkD1 .xk Cak /2 D PnkD1 x 2 y.a2 / ::: y.an2 / D PnkD1 x 2 y.PnkD1 a2 / : 1
k
k
k
P P Consequently the law of nkD1 .xk C ak /2 depends of .ak /1kn only via nkD1 ak2 . This is the characterization of the centered semicircular law established by Hiwatashi et al. (Theorem 3.2 in [5]). Hence x must have a centered semicircular law. Consider now x a variable with a centered semicircle law of radius r > 0. Assume that x satisfies (2). For every real m, one has then: .law/
.x C m/2 D x 2 C y.m2 /; hence .x C m/2 has to be infinitely divisible. Proposition 2 shows that this is not true. Consequently since the only possible distribution does not have the requested property (2), no distribution has this property. t u Proof of Proposition 2. It is well known that x 2 is infinitely divisible. Fix m 6D 0. It is clearly sufficient to consider the case where x has a semicircular law of radius 1. Hiwatashi et al. [5] have computed the R-transform of .x C m/2 : R.xCm/2 .z/ D
4m2 1 C : .4 z/ .2 z/2
(6)
For z 2 C with z D u C i v such that v > 0, Im.R.xCm/2 .z// has the same sign as P .v2 / D v4 C 2.2 u/.4m2 C 2 u/v2 C .2 u/.8m2 .4 u/2 C .2 u/3 /: The discriminant of P is equal to .2 u/Œ4m2 .2 u/.4m2 C 2.2 u// 8m2 .4 u/2 which is positive for 2 < u < 2 C 2m2 . Hence under this condition, P has two distinct roots. Their sum is equal to S D 2.u 2/.4m2 C 2 u/, which is positive. This implies that at least one root is positive and hence that one can choose v > 0 such that P .v2 / < 0. For this choice of .u; v/ we have Im.z/ > 0 and Im.R.xCm/2 .z/ < 0. Hence thanks to [7] (Theorem 3.7.2 (1)), we know that .xCm/2 can not be infinitely divisible.
212
N. Eisenbaum
With the notation of the introduction, we have the following proposition under the assumption that the state is faithful. Proposition 3. Let x be a standard semicircular variable. There is no random variable y, such that fx; yg is free and .law/
T .x 2 / D x 2 C y: Proof of Proposition 3. The density of x is given by : w.t/ D
1 p 4 t 2 1Œ2;2 .t/: 2
The density of x 2 is given by: u.t/ D
1 p 4t t 2 1.0;4 .t/: 2 t
One easily checks that for every f complex polynomial in one variable, we have EŒf .x C 2/ D EŒx 2 f .x 2 /; .law/
or equivalently: x C 2 D T .x 2 /. Assume the existence of y such that fx; yg is free and: .law/
x 2 C y D x C 2:
(7)
For any non commutative variable z, denote by var.z/ the quantity .z2 / Œ.z/2 . We have: var.x C 2/ D var.x/ D 1. Besides, since fx; yg is free, we have: var.x 2 C y/ D var.x 2 / C var.y/ D 1 C var.y/: Consequently, thanks to (7), we obtain: var.y/ D 0, which leads to the existence of .law/
a real constant c such that: x 2 D x C c. This last identity is obviously false.
t u
Acknowledgements We thank the referee for improvements on the original version.
References 1. F. Benaych-Georges, Failure of the Raikov theorem for free random variables. Séminaire de Probabilités XXXVIII, (Lecture Notes in Mathematics 1857, Springer, 2004), pp. 313–320 2. H. Bercovici, D. Voiculescu, Superconvergence to the central limit and failure of the Cramér theorem for free random variables. Probab. Theor. Relat. Field. 103, 215–222 (1995)
Another Failure in the Analogy Between Gaussian and Semicircle Laws
213
3. N. Eisenbaum, A Cox process involved in the Bose-Eistein condensation. Annales Henri Poincaré 9, 1123–1140 (2008) 4. N. Eisenbaum, H. Kaspi, M.B. Marcus, J. Rosen, Z. Shi, A Ray-Knight theorem for symmetric Markov processes. Ann. Probab. 28(4), 1781–1796 (2000) 5. O. Hiwatashi, T. Kuroda, M. Nagisa, H. Yoshida, The free analogue of noncentral chi-square distributions and symmetric quadratic forms in free random variables. Math. Z. 230(1), 63–77 (1999) 6. A.M. Kagan, O.V. Chalaevskii, A characterization of the normal law by a property of the noncentral chi-square distribution. (Russian) Litovsk. Mat. Sb. 7, 57–58 (1967) 7. D.V. Voiculescu, K.J. Dykema, A. Nica, Free Random Variables. A Non Commutative Probability Approach to Free Products with Applications to Random Matrices, Operator Algebras and Harmonic Analysis on Free Groups. CRM Monograph Series, 1. (American Mathematical Society, Providence, RI, 1992)
Global Solutions to Rough Differential Equations with Unbounded Vector Fields Antoine Lejay
Abstract We give a sufficient condition to ensure the global existence of a solution to a rough differential equation whose vector field has a linear growth. This condition is weaker than the ones already given and may be used for geometric as well as non-geometric rough paths with values in any suitable (finite or infinite dimensional) space. For this, we study the properties the Euler scheme as done in the work of A.M. Davie. Keywords Controlled differential equations • Rough paths • Euler scheme • Global solution to differential equation • Rough differential equation
AMS Classification: 60H10, 65C30
1 Introduction Initiated a decade ago, the theory of rough paths imposed itself as a convenient tool to define stochastic calculus with respect to a large class of stochastic processes out of the range of semi-martingales (fractional Brownian motion, ...) and also allows one to define pathwise stochastic differential equations [6, 8, 10, 11, 13, 16, 18]. The idea to define integrals of differential forms along irregular paths or solutions of differential equations driven by irregular paths is to extend properly such paths in a suitable non-commutative space. These extensions encode in some sense the “iterated integrals” of the paths. Let us denote by x the driving path which lives in a
A. Lejay () Project-team TOSCA, Institut Elie Cartan Nancy (Nancy-Université, CNRS, INRIA), Boulevard des Aiguillettes B.P. 239 F-54506 Vandœuvre-lès-Nancy, France e-mail: [email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__11, © Springer-Verlag Berlin Heidelberg 2012
215
216
A. Lejay
tensor space and Lie group T .Rd / WD R ˚ Rd ˚ .Rd ˝ Rd / and that projects onto some continuous path of finite p-variations on Rd with p 2 Œ2; 3/. Such a path x is called a p-rough path. The goal of this article is to study global existence of solutions to the rough differential equation (RDE) Z
t
yt D y0 C
f .ys / dxs
(1)
0
when f is not bounded. What is called a solution to (1) needs to be properly defined. Indeed, there exist two notions of solutions we will deal with (See Definitions 1 and 2 below). Let us recall here the already existing results. The end of the introduction contains a short discussion about the differences between our result and the ones presented here. Linear case: The special case of a linear vector field f was studied for example in the articles [1, 6, 9, 12, 15]. The original approach of T. Lyons: In the original approach, uniqueness and continuity of the Itô map x 7! y, where y is a solution to (1), is proved for a bounded function f which is twice differentiable with bounded derivatives and such that r 2 f is -Hölder continuous with 2 C > p [6, 10, 11, 13, 16, 18]. One knows from [2] that these conditions on f are essentially sharp. As pointed out soon after by A.M. Davie [2], existence of a solution to (1) is granted when f is bounded with a bounded derivative which is -Hölder continuous, 2 C > p (See also [6, 17] for example). However, in this case, one has to restrict to finite-dimensional spaces and several solutions may exist [2]. The approach of A.M. Davie: Providing an alternative approach to the one of T. Lyons based on a fixed point theorem, A.M. Davie studied in [2] the Euler scheme defined by yinC1 D yin C f .yin /xt1i ;ti C1 C f rf .yin /xt2i ;ti C1 where xs;t D xs1 ˝ xt and x 1 (resp. x 2 ) is the projection of x onto Rd (resp. Rd ˝ Rd ). With conditions only on the regularity of the vector field, local existence is shown. Global existence is granted if there exist two positive increasing functions A and D on Œ1; C1/ with D.R/ R1C A.R/, 2 p < C 2 < 3 such that jf .y/j D.R/; jrf .y/ rf .z/j A.R/jy zj ; 8y; z s.t. jyj; jzj R; and
Z 1
C1
1 p1 A.R/ D.R/1Cp
1 1C
dR D C1:
(2)
Besides, if (2) is not satisfied for some functions A and D as above, it is possible to construct a vector field f and a driver x such that the solution to (1) explodes in a finite time.
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
217
Applied to functions f with a -Hölder continuous derivative rf with 2C >p, one may take A.R/ equal to a constant. Then global existence is granted for example if D.R/ D Rı with ı < .1C /=.1Cp/, but an explosion may occur for a function f is in class of functions if ı > .1 C /=.1 C p/, which is the case for ı D 1. The approach by P. Friz and N. Victoir: In [5, 6], P. Friz and N. Victoir provide an alternative construction of the solutions of RDE that relies on sub-Riemannian geodesics and hence of geometric rough paths. The case of geometric rough paths shall be considered using .p; p=2/-rough paths [17]. In [6]*Exercise 10.61, they show that it is not necessary that the vector field is bounded, provided that f is Lipschitz continuous, rf is -Hölder continuous and f rf is also -Hölder continuous. Using a fixed point approach: In [12], we have also studied the existence of a global solution, by using the approach on fixed point theorem. With this approach, the RDE (1) is solved up to a finite “short time” horizon. Global existence follows from the convergence of a series related to the sum of the horizons. The complete conditions are cumbersome to write, but if h.R/ D supjyjR jf .y/j, then global existence is granted provided that f has a bounded derivative rf which is -Hölder continuous with 2 C > p and h.R/ R!1 Rı , 0 < ı 1=p or h.R/ R!1 log.R/. The case of a “regular enough” driver (also called Young case): If 1 p < 2, then global existence is granted if the vector field f is -Hölder continuous with 1 C > p [14]. The one-dimensional approach of H. Doss and H. Sussmann: If d D 1, H. Doss [4] and H. Sussmann [20] have shown that a solution to an equation of type (1) may be defined by considering the solution to the ODE dzt D f .zt / dt and setting yt D zxt . Local existence to dzt D f .zt / dt is granted provided that f is continuous. Global existence holds under a linear growth condition on f as it follows easily from an application of the Gronwall inequality [19] which provides a global bound on z: If jf .z/j A C Bjzj, then jzt j .jz0 j C At/ exp.Bt/ for any t 0. This approach may be generalized for a multi-dimensional driver x when f has vanishing Lie brackets. The situation is more intricate when x lives in a space of dimension bigger than 2. In particular, the asymptotic behavior of rf plays a fundamental role. If x is a p-rough path (a rough path of finite p-variation) with p 2 Œ2; 3/, then for any function of finite p=2-variations with values in Rd ˝ Rd , z D x C is also a p-roughR path. In addition, as shown in [17] for the general case, the solution t to yt D y0 C 0 f .ys / dzs is also solution to Z
t
yt D y0 C 0
Z
t
f .ys / dxs C
f rf .ys / ds :
(3)
0
Consider any rough path z living above the path 0 on Rd with .t/ D ct for a matrix c (if c is anti-symmetric, then such a rough path is the limit of a sequence of smooth paths lifted in the tensor space by their iterated integrals). Then y is solution to
218
A. Lejay
Z
t
yt D y0 C
f rf .ys /c ds: 0
Thus, explosion may occurs according to the behavior of f rf . In particular, f may grow linearly, but f rf may grow faster that linearly, and an explosion may occur. Of course, if rf is bounded and f grows linearly, then f rf also grows linearly. 1 (M. Gubinelli). Consider the solution y of the RDE yt D a C RExample t 2 0 f .ys / dxs living in R and driven by the rough path xt D .1; 0; .1 ˝ 1/t/ with values in 1 ˚ R ˚ .R ˝ R/. This rough path lies above the constant path at 0 2 R and has only a pure area part which proportional to t. Note that this rough path can only beRseen as a p-rough path with p > 2 [13]. Then y is also a solution to t yt D a C 0 .f rf /.ys / ds (See [17]). The vector field f 2 R2 ! L.R; R2 / given by f ./ D .sin.2 /1 ; 1 /; D .1 ; 2 / 2 R2 has a linear growth but .f rf /./ D .sin2 .2 /1 C 12 cos.2 /; sin.2 /1 / has a quadratic growth. R t Take the initial point a D .a1 ; 0/ with a1 > 0. Then .yt /2 D 0 and .yt /1 D a1 C 0 .ys /21 ds so that .yt /1 ! C1 in finite time equal to 1=a1 . This proves that explosion may occur in a finite time. It proves that in the context of RDE, the growth of f rf is important. The previous example involves a pure non-geometric rough path. Yet a slight modification of this example show that this may happen also for geometric rough paths. Example 2. Let us consider ytn with
Z t n 1 Z t .y /s sin..y n /2s / 0 n dns DaC ds C .y n /1s 0 0 0
(4)
. n ; n /t D n1=2 Œcos.2nt/; sin.2nt/:
The smooth rough path x n living above . n ; n / converges in p-variation with p > 2 to the rough path in 1 ˚ R2 ˚ .R2 ˝ R2 / defined by 0 2 .t s/ : xs;t D 1; 0; 2 0 This is the standard example which show the discontinuity of the Lévy area [18]. Let us assume that the solutions y n to (4) is uniformly bounded on Œ0; T for some T > 0. By the continuity theorem (one may assume that f , rf and r 2 f are
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
219
bounded), the solution y n to (4) converges to the solution to Z t yt D a C 2 0
.ys1 /2 cos.ys2 / ds: ys1 sin.ys2 /
Hence, we are in the same situation as Example 1 if a2 D 0 and then y 2 0. If f ./ D .f1 ./; f2 .// with f1 ./ D . 1 sin. 2 /; 0/ and f2 ./ D .0; 1 /, then rf is not bounded although f grows linearly. Since y n converges to y which explodes in a finite time 1=a1 , y n cannot be uniformly bounded on any time interval Œ0; T for T bigger than 1=a1 . Note however that the p-variation norm of the rough path .x n / remains bounded. Unlike the situations where the Doss-Sussmann approach may be used, bounds on the solution to RDE driven by a p-rough path with p > 2 cannot be derived from the sole information on the linear growth of f . Although we use some general ideas already used in the context of rough paths theory (See for example [2, 5, 6, 8, 12, 14], ...), we should note that: • Our conditions on the vector field are weaker than the ones already given in the literature. In particular, we show that for existence, f and rf need to be continuous and rf needs to be bounded. Yet the Hölder regularity of rf plays no role here, while f rf shall be Hölder continuous. Even if the computations are very close to the one of [2, 5], we then give a simple natural condition on the vector field as well as a simple bound on the solution. The class of functions f with a bounded derivative rf which is -Hölder continuous is different from the class of functions f with a bounded derivative rf and such that f rf is -Hölder continuous. Example 3. Let us consider f .x/ D sin.x 2 /=x for jxj 1. The first and second order derivatives of f are rf .x/ D 2 cos.x 2 /
sin.x/2 x2
and r 2 f .x/ D 4x sin.x 2 /
sin.x 2 / 2 cos.x/ sin.x/ C : x x3
Hence, rf is not uniformly -Hölder continuous on fxI jxj 1g whatever 2 .0; 1. Yet 2 cos.x 2 / sin.x 2 / sin2 .x 2 / f .x/rf .x/ D x x3 and it is easily checked that f rf has a bounded derivative on fxI jxj 1g and is then uniformly Lipschitz continuous (and thus -Hölder continuous) on fxI jxj 1g. The class of vector fields f with a bounded derivatives rf and such that f rf is -Hölder continuous enjoys the property to be stable under some change of variables. This is not necessarily the case for vector fields f with bounded first and second-order derivatives rf and r 2 f .
220
A. Lejay
For a toy example, let us consider the simple case of a function f W R ! R such that rf is bounded and RL D f rf has a bounded derivative. Assume that for a smooth t path x, yt D a C 0 f .ys / dxs has a solution y which remains in .1; C1/. Set Rt zt D log.yt /. Then z is solution to zt D log.a/C 0 g.zs / dxs with g.z/ D e z f .e z /. The vector field g has the interesting property to remain bounded on .0; C1/ and one may then hope to deduce some bounds on the solution y from some bound on the solution z. Since L D f rf has a bounded derivative, g has a bounded derivative on .0; C1/ and grg has also a bounded derivative, which means that grg is globally Lipschitz. Yet g 00 .z/ D e z f .e z / f 0 .e z / C e z f 00 .e z /: This means that g 0 is uniformly Lipschitz on .0; C1/—a condition required to deal with RDE with a regularity index D 1—only if yf 00 .y/ remains bounded, which is a stronger condition than assuming that f 00 remains bounded. This kind of computations may be carried to the multi-dimensional case using polar coordinates and an exponential change of variable of the radial component. This explains the failure of the attempt carried by some persons, including the author of the present article, to get a global bound by such a change of variable under the sole assumption that rf is uniformly -Hölder and bounded. • The condition on the vector field appears naturally when one uses the approach by the Euler scheme proposed by A.M. Davie in [2]. This is not the case when one uses the notion of solution proposed by T. Lyons because the term f rf is somewhat hidden in the cross iterated integral between the solution y and the driver x. The idea used in this article to get a global bound in closely related to the one in [12] where a fixed point approach was used. Yet in this article [12], we did not succeed article to obtain a global bound in a general. In addition, the superfluous assumption on the Hölder regularity of rf was used and necessary. • The use of sub-Riemaniann geodesics and the reduction to smooth drivers was the core ideas of [5,6]. Here, there is no need to restrict to geometric rough paths and the results may be used for any finite-dimensional Banach spaces and even infinitedimensional in some cases. Hence, the structure of the underlying spaces plays no role here. Example 4. The most natural example of a non-geometric rough path is the one 2;i;j B living above a d -dimensional Brownian path W with B 1 D W and Bs;t D Rt j i i s .Wr Ws / dBr . Here the integral has to be understood in the Itô sense. e may be constructed above the Brownian path W with A geometric rough pathRB t 2;i;j j 1 e e B D W and B s;t D s .Wri Wsi / ı dWr , where the stochastic integral is the 2;i;j e and B are linked by Bs;t e2;i;j Stratonovich one. Hence, B D B C 12 .t s/ıi;j . s;t e RDE driven by B correspond to Itô SDE while the ones driven by B corresponds to Stratonovich SDE. When one use the Brownian rough path B, the Euler scheme presented here corresponds indeed to the Milstein scheme. In the very beginning of rough paths
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
221
theory, the rate of convergence in this case has been studied by J. Gaines and T. Lyons [7] with aim at developing simulation algorithms with adaptative stepsize. In view of (3), the regularity condition on f rf is also the one which is necessary to deal with a non-geometric p-rough path seen as a .p; p=2/-geometric rough path using the decomposition of the space introduced in [17]. • The notion of solution to a RDE introduced by T. Lyons [18] cannot be used here (See Definition 1), so that we use the notion of solution introduced by A.M. Davie in [2] (see Definition 2) which is similar to the one proposed by M. Gubinelli [8]. We show in Sect. 10 that the two notions of solutions coincide when rf is -Hölder continuous. • A general conclusion to draw from the cited works on global existence for non bounded vector fields f is that a variety of results could be given according to the behaviour at infinity of f and its derivatives. Hence, the growth of f is not the only factor to look at.
2 Notations and Hypotheses Let ! be a control. By this, we mean a function defined from 2 WD f0 s t T g to RC which is continuous close to its diagonal and super-additive !.s; r/ C !.r; t/ !.s; t/; 0 s < r < t T: For the sake of simplicity, we assume here that ! is continuous and that !.s; t/ > 0 as soon as t > s. Let x be a path with values in T .Rd / D R ˚ Rd ˚ .Rd ˝ Rd /. We set xs;t D 1 x ˝ xt . The part of x in Rd is denoted by x 1 and the part in Rd ˝ Rd is denoted by x 2 . Then x is a rough path of finite p-variation controlled by ! when the quantity (
1 2 j j jxs;t jxs;t kxk WD sup max ; 1=p !.s; t/ !.s; t/2=p 0s
)
is finite for a fixed p. If !.s; t/ D t s, then we work indeed with paths that are 1=p-Hölder continuous. Throughout all this article, we consider only the case where p 2 Œ2; 3/. The case p 2 is covered for example by [14]. For a path y with values in Rm , we set kyk WD
sup
0s
1 jxs;t j : !.s; t/1=p
222
A. Lejay
A vector field is an application f which is linear from Rm to L.Rd ; Rm /, the space of linear applications from Rd to Rm . With indices, we set f .x/ D ei fji .x/b e j , m where fei gi D1;:::;m is the canonical basis of R , and fb ej gj D1;:::;d is the dual of the d canonical basis of R . We set f rf .x/ WD
m X i D1
ei
d X @fji kD1
@xk
.x/f`k .x/b e ` ˝b e ?j ;
which means that f rf is an application from Rm to L.Rd ˝ Rd ; Rm /. Hypothesis 1. The function f is continuously differentiable from Rm to L.Rd ; Rm / and is such that rf is bounded and F .x/ WD f .x/ rf .x/ is -Hölder continuous with norm H .F / from Rm to L.Rm ˝ Rd ; Rm /. Note that this hypothesis is slightly different from the one given usually to prove existence to solutions of RDE where it is assumed that rf is -Hölder continuous. Hypothesis 2. We assume that p 2 Œ2; 3/ and that WD .2C /=p is greater than 1. Definition 1 (Solution in the sense of Lyons). We call by a solution to (1) in the sense of Lyons the projection onto Rm of a rough path z of finite p-variations controlled by ! with values in T .Rm ˚ Rd / which solves the following equation Z
t
zt D z0 C
g.zs / dzs ;
(5)
0
where z projects onto T .Rd / as x, onto Rm as y, and g is the differential form g.y; x/ D dx C f .y/ dx. The integral in (5) shall be understood as the “rough integral”, that is as an integral in the sense of rough path. Note that the definitions of z involves the “cross-iterated integrals” between x and y, and requires that the derivative of f is -Hölder continuous. Under Hypothesis 1, it is not compulsory that rf is -Hölder continuous, so that we shall use another notion of solution, since we cannot use a fixed point theorem that relies on the definition of a rough integral.
3 Solution in the Sense of Davie The notion of solution of (1) we use is the notion of solution in the sense of Davie, introduced in [2]. Definition 2 (Solution in the sense of Davie). A solution of (1) in the sense of Davie is a continuous path y from Œ0; T to Rm of finite p-variation controlled by ! such that for some constant L, jyt ys D.s; t/j L!.s; t/ ; 8.s; t/ 2 2
(6)
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
223
with 1 2 D.s; t/ D f .ys /xs;t C F .ys /xs;t :
The next propositions, which assume the existence of a solution in the sense of Davie, will be proved below in Sect. 4 Proposition 1. Let y be a solution of (1) in the sense of Davie under Hypotheses 1 and 2. Then D.s; t/ is an˚almost rough path whose associated rough path is y and for a family of partitions ftin gi D0;:::;n n2N of Œ0; t whose meshes decrease to 0, yt D y0 C lim
n!1
n X
D.tin ; tinC1 /:
(7)
i D0
Proposition 2 (Boundedness of the solution). Let y be a solution to (1) in the sense of Davie under Hypotheses 1 and 2. Then kyk and kyk1 are bounded by some constants that depend only on krf k1 , H .F /, jf .y0 /j, jy0 j, kxk, !.0; T /, and p. More precisely, there exist some constants C1 depending only on krf k1 , H .F /, kxk, p and and C2 depending only on jf .0/j, krf k1 , H .F /, kxk, p, such that sup jyt j R.T /jy0 j C C2 .R.T / 1/ t 2Œ0;T
with R.T / exp.C1 maxf!.0; T /1=p ; !.0; T /g/:
4 Proofs of Propositions 1 and 2 Let us set for 0 s r t T , D.s; r; t/ WD D.s; t/ D.s; r/ D.r; t/: Then the main idea of the proofs (as well as the proofs of the other theorems) are the following: First, we find a function C.kyk; L; T / such that jD.s; r; t/j C.kyk; L; T /!.s; t/ for all .s; t/ 2 2 . From the sewing Lemma (See for example [13]*Theorem 5, p. 89), after having shown that y is the rough path corresponding to the almost rough path .D.s; t//.s;t /2 2 , we get that for some universal constant M , jyt ys D.s; t/j M C.kyk; L; T /!.s; t/ : Then, after having estimated jD.s; t/k C 0 .kyk; L; T /!.s; t/1=p , we get an inequality of type kyk M C.kyk; L; T /!.0; T / 1=p C C 0 .kyk; L; T /:
224
A. Lejay
A careful examination of the functions C.kyk; L; T / and C 0 .kyk; L; T / shows that indeed L itself depends on y0 , kyk and T and that kyk A.y0 ; T / C B.T /kyk with B.T / decreasing to 0 as T decreases to 0. Then, choosing T small enough implies that kyk is bounded in small time and then for any time T using the arguments presented in appendix. This idea is the central one used for example in [12, 14]. Let us set 1 B1 WD sup jF .yt /j; B2 .s; t/ WD yt ys f .ys /xs;t ; t 2Œ0;T
B3 .a; b/ WD f .b/ f .a/; B4 .a; b/ WD F .b/ F .a/; 1 and B5 .s; t/ WD f .yt / f .ys / F .ys /xs;t
as well as WD !.0; T /1=p : Remark 1. The quantity will be used to denote the “short time” horizon and is the central quantity for getting our estimates. Since F is -Hölder continuous, B1 jF .y0 /j C H .F /kyk : If y is a solution in the sense of Davie with a constant L, jB2 .s; t/j L!.s; t/ C B1 kxk!.s; t/2=p : Since f is Lipschitz continuous, jB3 .a; b/j krf k1 jb aj: In addition, jB4 .a; b/j H .F /jb aj : Finally, Z
1
f .yt / f .ys / D
rf .ys C ys;t /ys;t d 0
Z D
0
Z
1 1 rf .ys C ys;t /f .ys /xs;t d C
Z
D 0
0
1
rf .ys C ys;t /B2 .s; t / d
1 1 rf .ys C ys;t /.f .ys C ys;t / f .ys //xs;t d
Global Solutions to Rough Differential Equations with Unbounded Vector Fields Z C Z D
0
Z
1 1 F .ys C ys;t /xs;t d C
0
Z
1 1 F .ys C ys;t /xs;t d C
225
1 0
rf .ys C ys;t /B2 .s; t / d
1 0
rf .ys C ys;t /.B2 .s; t /
1 / d B3 .ys ; ys C ys;t /xs;t
Z 1 D F .ys /xs;t C
Z C
0
0
1 1 B4 .ys ; ys C ys;t /xs;t d
1 1 rf .ys C ys;t /.B2 .s; t / B3 .ys ; ys C ys;t /xs;t / d :
(8)
This proves that jB5 .s; t/j H .F /kyk kxk!.s; t/.1C /=p C krf k21 kxkkyk!.s; t/2=p Ckrf k1 L!.s; t/ C krf k1 jF .y0 /jkxk!.s; t/2=p CH .F /kyk !.0; T /=p kxk!.s; t/2=p : Lemma 1. For any 0 s r t T , jD.s; r; t/j .C3 ./kyk C C4 ./kyk C C5 ./L C C6 .; y0 //!.s; t/
with
C3 ./ WD H .F / kxk2 .1 C 1C / C kxk/; C4 ./ WD krf k21 kxk2 ; C5 ./ WD krf k1 kxk
and C6 .; y0 / WD krf k1 jF .y0 /jkxk2 1 jf .y0 /jkrf k21 kxk2 1 : 2 2 2 1 1 D xs;r C xr;t C xs;r ˝ xr;t , Proof. With xs;t 1 1 1 2 C F .ys /xs;r ˝ xr;t C .F .ys / F .yr //xr;t : D.s; r; t/ D .f .ys / f .yr //xr;t
Hence jD.s; r; t/j jB5 .s; r/jkxk!.s; t/1=p C jB4 .ys ; yr /jkxk!.s; t/2=p : t u
This proves the result. Lemma 2. For any .s; t/ 2 , 2
jD.s; t/j .C7 .; y0 / C C8 ./kyk C C9 ./kyk/!.s; t/1=p
226
A. Lejay
with C7 .; y0 / WD .jf .y0 /j C jF .y0 /j/kxk jf .y0 /j.1 C krf k1 /kxk; C8 ./ WD H .F /kxk ; and C9 ./ WD krf k1 kxk: Proof. This follows from jD.s; t/j jf .y0 /jkxk!.s; t/1=p C kykkxkkrf k1 !.s; t/2=p CjF .y0 /jkxk!.s; t/2=p C kyk kxkH .F /!.s; t/.1C /=p : t u
This proves the Lemma. We have now all the required estimates to prove Propositions 1 and 2.
Proof (Proof of Proposition 1). It follows from Lemma 1 that fD.s; t/g.s;t /2 2 is an almost rough path. From the sewing lemma (See for example [13]*Theorem 5, p. 89), there exists a path fzt gt 2Œ0;T as well as a constant M depending only on
such that jzt zs D.s; t/j M!.s; t/ ; 8.s; t/ 2 2 : (9) This function is unique in the class of functions satisfying (9) and with (6), z is equal to y. Equality (7) follows from the very construction of z. t u Proof (Proof of Proposition 2). We assume first that y is a solution in the sense of Davie with constant L. Let us note that jyt ys j jyt ys D.s; t/j C jD.s; t/j L!.s; t/ C jD.s; t/j: With Lemma 2, if D !.0; T /1=p is small enough such that C8 ./
1 1 and C9 ./ ; 4 4
(10)
then
1 1 kyk L1C C C7 .; y0 / C kyk C kyk : 4 4 Since 1, if kyk 1, then kyk kyk and then for small enough (note that the choice of does not depend on y0 ), kyk 2 maxf1; L1C C C7 .; y0 /g:
(11)
Since C7 .; y0 / decreases with , the boundedness of kyk and kyk1 also hold, with a different constant, for any time T by applying Proposition 6 in appendix. Now, if y is a solution in the sense of Davie with a constant L, it is also a solution in the sense of Davie with the constant
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
L0 WD
sup .s;t /2 2 ; s6Dt
227
jyt ys D.s; t/j : !.s; t/
From the sewing Lemma, there exists a universal constant M depending only on
such that jD.s; r; t/j L0 M sup :
2 .s;t /2 ; s6Dt; r2.s;t / !.s; t/ From Lemma 1 and the inequalities a 1 C a for a 0 and 2 Œ0; 1 as well as jf .y0 /j jf .0/j C krf k1 jy0 j, L0 M.C3 ./kyk C C4 ./kyk C C5 ./L0 C C6 .; y0 // C10 ./ C C13 ./kyk C M C5 ./L0 C C12 ./jy0 j; with C10 ./ WD M C3 ./ C jf .0/jC11 ./; C11 ./ WD M krf k21 kxk2 1 ; C12 ./ WD krf k1 C11 ./; and C13 ./ WD M.C3 ./ C C4 .//: If M C5 ./ 1=2; then
(12)
L0 2C10 ./ C 2C13 ./kyk C 2C12 ./jy0 j:
With (11), under conditions (10) and (12) on , kyk 2 C 4C10 ./1C C 4C13 ./1C kyk C 4C12 ./jy0 j1C C C7 .; y0 /; Under the additional condition that 4C13 ./1C
1 ; 2
we get that kyk C14 ./ C C15 ./jy0 j with C14 ./ WD 4 C 8C10 ./1C C 2jf .0/j.1 C krf k1 /kxk; C15 ./ WD 8C12 ./1C C 2krf k1 .1 C krf k1 /kxk:
(13)
228
A. Lejay
Due to the dependence of the constants with respect to , conditions (10), (12) and (13) hold true if K for a constant K depending only on krf k1 , H .F /, kxk, p and . Indeed, (11) also holds for yjŒS;S 0 on any time interval !.S; S 0 / provided that !.S; S 0 /1=p is small enough. The result follows from Proposition 7 in appendix. t u
5 Existence of a Solution The existence of a solution is proved thanks to the Euler scheme, which allows one to study to define a family of paths that is uniformly bounded with the uniform norm. Let us fix a partition fti gi D0;:::;n of Œ0; T and let us set xi;j WD xt1 ˝ xtj and i !i;j WD !.ti ; tj /. Let us consider the Euler scheme 1 2 yi C1 D yi C f .yi /xi;i C1 C F .yi /xi;i C1 ; i D 0; : : : ; n 1
as well as the family fyi;j g0i <j n defined by 1 2 C F .yi /xi;j : yi;j WD f .yi /xi;j
We set kyk?;0;n WD
sup
jyj yi j
0i <j n
1=p
:
!i;j
(14)
1=p
Lemma 3. If WD !0;n is small enough so that C5 ./ D krf k1 kxk and
1K with K WD 21 < 1 4
(15)
L WD 4
C6 .; y0 / C C3 ./kyk?;0;n C C4 ./kyk?;0;n 1K
then
jyi;k yk yi j L!i;k
(16)
for all 0 i k n. Proof. The proof of this Lemma follows along the lines the one of Lemma 2.4 in [2] and relies on an induction on k i . Clearly, (16) is true for k D i C 1. Fix m > 1 and let us assume that (16) holds for any i < k such that k i < m.
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
229
Let us choose i < k such that k i D m, m 2. Let j be the index such that !i;j !i;k =2 and !i;j C1 > !i;k =2. Since ! is super-additive, !j C1;k !i;k =2 and then
!i;j C !j C1;k 21 !i;k D K!i;k : (17) We set yi;j;k WD yi;k yi;j yj;k : For j as above, since yj;j C1 yj yj C1 D 0, and using (17), jyi;k yk C yi j jyi;j;k j C jyi;j yj C yi j C jyj;k yk C yj j jyi;j;k j C jyj;j C1;k j C jyj C1;k yk C yj C1 j C jyi;j yi C yj j Cjyj;j C1 yj C1 C yj j
: jyi;j;k j C jyj;j C1;k j C LK!i;k 2 2 2 1 1 Since xi;k D xi;j C xj;k C xi;j ˝ xj;k , 1 2 1 1 yi;j;k D .f .yj / f .yi //xj;k C .F .yj / F .yi //xj;k C F .yi /xi;j ˝ xj;k :
Using the same computations as (8), 1 1 1 2 yi;j;k D .f .yi / f .yj //xi;j F .yi /xi;j ˝ xj;k C .F .yi / F .yj //xj;k Z 1 1 1 D .F .yi C .yj yi // F .yi //xi;j ˝ xj;k d 0
Z
1 2 1 rf .yi C .yj yi //.yj yi yi;j C F .yi /xi;j /xi;j d
C Z
0
0
1
Z
1 0
1 1 rf .yi C .yj yi //rf .yi C .yj yi //.yj yi / d d xi;j ˝ xj;k
2 C.F .yi / F .yj //xj;k :
(18)
We then face the same estimates as the one in the proof of Lemma 1, where we replace the fact that y is a solution in the sense of Davie with a constant L by our induction hypothesis on jyi;j yj C yi j. Then
jyi;j;k j .C3 ./kyk?;0;n C C4 ./kyk?;0;n C C5 ./L C C6 .; y0 //!i;k :
The results follows from our choice of and L.
t u
The next lemma is the equivalent of Proposition 2 for the Euler scheme. Lemma 4. For n such that tn D T , kyk?;0;n is bounded by a constant that depends only on !.0; T /, kxk, krf k1 , H .F /, , p and jf .0/j.
230
A. Lejay
Proof. Using Lemma 2,
1=p
jyi;j j .C7 .; y0 / C C8 ./kyk?;0;n C C9 ./kyk?;0;n /!i;j and then with Lemma 3,
kyk?;0;n .C7 .; y0 / C C8 ./kyk?;0;n C C9 ./kyk?;0;n /
C41C
C6 .; y0 / C C3 ./kyk?;0;n C C4 ./kyk?;0;n 1K
:
In addition to (15), we choose small enough so that C8 ./ C 4.1 K/1 1C C3 ./
1 4
(19)
C9 ./ C 4.1 K/1 1C C3 ./
1 : 4
(20)
and
Since 1, kyk?;0;n kyk?;0;n when kyk?;0;n 1. Hence, kyk?;0;n maxf1; 2C7 .; y0 / C 81C C6 .; y0 //g: This proves that for a choice of (or equivalently T or n) small enough depending only on kxk, krf k1 , H .F /, and p, then kyk?;0;n is bounded by a constant that depends only on kxk, krf k1 , H .F /, , p, jf .y0 /j and jF .y0 /j. However, jF .y0 /j and jf .y0 /j are bounded by some constants that depends only on jf .0/j and krf k1 . The result is proved by finding a sequence n0 D 0 < n1 < < nN such that !n1 ;ni C1 p with tnN D T and satisfying (15), (19) and (20). Since ! is continuous close to its diagonal, there exists such a finite number N of intervals, and this number depends only on the choice of , and then on kxk, krf k1 , H .F /, and p (and not on y0 nor f .0/). Finally, it is easily shown that kyk?;0;nN N
p1
N 1 X
kyk?;ni ;ni C1 ;
i D1
which proves the result by applying the result on the successive time intervals Œtni ; tni C1 and replacing y0 by yti . t u Finally, is order to interpolate the Euler scheme and to get a good control, we shall add an hypothesis on !, which is trivially satisfied in the case of !.s; t/ D t s, that is for Hölder continuous rough paths. Hypothesis 3. We assume that there exists a continuous, increasing function such that
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
231
.t/ .s/ !.s; t/ and .t/ .s/ !.s; t/ are bounded for 0 s < t T . Proposition 3. Under Hypotheses 1, 2 and 3, there exists a least a solution in the sense of Davie to (1). Proof. From the family fyi g, we construct a path from Œ0; T to Rm by yt D yi C
.t/ .ti / yi;i C1 ; t 2 Œti ; ti C1 : .ti C1 / .ti /
(21)
From standard computations, there exists a constant C16 which depends only on p and the lower and upper bounds of !.s; t/=..t/ .s// such that kyk C16 kyk?;0;n : With Lemma 4, we have a uniform bound in kyk?;0;n , and then on the constant L when (or n) is small enough. Hence, for any partition satisfying Hypothesis 3, the path y has a p-variation which is bounded by a constant that does not depend on the choice of the partition. Now, let y n be a family of paths constructed along an increasing family of partitions ˘ n whose meshes decrease to 0. Then there exists a subsequence fy nk gk1 of fy n gn2N which converges in q-variation for q > p to some path y of finite p-variation. For any .s; t/ in \n0 ˘ n , 1 2 F .ys /xs;t j L!.s; t/ : jyt ys f .ys /xs;t
Since \n0 ˘ n is dense in Œ0; T and y is continuous, this proves that y is the solution to (16) in the sense of Davie, at least when T is small enough. The passage from a solution on Œ0; T with T small enough to a global solution is done by using the arguments of Lemma 7, Lemma 8 and Proposition 6. t u
6 Distance Between Two Solutions and Uniqueness We now consider a more stringent assumption than Hypothesis 1. Hypothesis 4. The function f is twice continuously differentiable from Rm to L.Rd ; Rm / and is such that rf , r 2 f are bounded and F .x/ WD f .x/ rf .x/ is such that rF is -Hölder continuous with constant H .rF / from Rm to L.Rm ˝ Rd ; Rm /.
232
A. Lejay
We consider two rough paths u and v of finite p-variation controlled by !, p 2 Œ2; 3/, as well as two vector fields f and gR satisfying Hypothesis 4. Let R t y and z be t respectively some solutions to yt D y0 C 0 f .ys / dus and zt D z0 C 0 g.zs / dvs . We have seen that y and z remain is a ball of radius R that depends only on krf k1 , krgk1 , krF k1 , krGk1 (since F and G D g rg are Lipschitz continuous), kuk, kvk, y0 , z0 , jf .0/j, jg.0/j, !.0; T /, and p. Definition 3. We say that a constant C satisfies Condition .S/ if it depends only on the above quantities, as well as H .rF /, H .rG/, kr 2 f k1 and kr 2 gk1 . We then set for some functions h, h0 , ıR .h; h0 / D
sup jh.z/ h0 .z/j: z2B.0;R/
We also set ı.u; v/ D ku vk and ı.y; z/ D
sup
0s
jys;t zs;t j : !.s; t/1=p
Theorem 1. Under Hypotheses 2 and 4, there exists some constant C17 satisfying Condition .S/ such that for all .s; t/ 2 2 , jys;t zs;t j C17 !.s; t/1=p .jy0 z0 j C ı.u; v/ C ıR .f; g/ C ıR .F; G/ C ıR .rf; rg/ C ıR .rF; rG//:
(22)
The proof of the following corollary is then immediate from the previous estimate. Corollary 1. Under Hypotheses 2 and 4, there exists a unique solution in the sense of Davie to (1). Theorem 1 proves that the Itô map which sends x to the unique solution to (1) is locally Lipschitz continuous in y0 , f and x. The next lemma is the main estimate of the proof and show why extra regularity shall be assumed on F . This lemma is already well-known (See Lemma 3.5 in [2]) but we recall its proof which is straightforward. Lemma 5. Let h be a function of class C 1 from Rm to R such that rh is -Hölder and bounded. Then for all a; b; c; d in Rm , jh.a/ h.b/ C h.c/ h.d /j ja b c C d jkrhk1 C H .rh/.ja bj C jc d j /jb d j: Proof. The result follows from
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
jh.a/ h.b/ C h.c/ h.d /j ˇZ 1 Z ˇ rh.c C .a c//.a c/ d D ˇˇ
233
ˇ ˇ rh.d C .b d //.b d / d ˇˇ 0 0 ˇ ˇZ 1 ˇ ˇ ˇ rh.c C .a c//.a c b C d / d ˇˇ Dˇ 0
ˇZ ˇ C ˇˇ
1 0
1
ˇ ˇ rh.c C .a c// rh.d C .b d // .b d / d ˇˇ Z
1
ja b c C d jkrhk1 C H .rh/
j.1 /.cd / C .ab/j jb d j d ; 0
since .x C y/ x C y for x; y 0 and 2 Œ0; 1.
t u
Let us denote by d the operator which, applied to an expression involving y, f and u, takes the difference between this expression and the similar expression with y replaced by z, f replaced by g and u replaced by v. For example, 1 d.f .ys /xs;t / D f .ys /u1s;t g.zs /v1s;t :
If ˛.y; f; u/ and ˇ.y; f; u/ are two expressions, then d.˛.y; f; u/ˇ.y; f; u// D d.˛.y; f; u//ˇ.y; f; u/ C ˛.z; g; v/d.ˇ.y; f; u//: (23) Proof (Proof of Theorem 1). In the proof, we assume without loss of generalities that < 1. Let us choose a constant A such that jd.D.s; r; t//j A!.s; t/
(24)
jd.D.s; t//j A!.s; t/1=p :
(25)
and From the sewing lemma on the difference of two almost rough paths (See for example [13]*Theorem 6, p. 95), there exists some universal constant M (depending only on ) such that jd.ys;t D.s; t//j MA!.s; t/ :
(26)
Our aim is to obtain some estimate on A. Let us note first that jdF .ys /u2s;t j .jF .ys / F .zs /j C jF .zs / G.zs /jkuk!.s; t/2=p C .jG.z0 /j C krGk1 kzk/ı.u; v/!.s; t/2=p :
234
A. Lejay
Since jys zs j ı.y; z/ C jy0 z0 j, with Lemma 5, jF .ys / F .zs /j krF k1 .ı.y; z/ C jy0 z0 j/; then for some constant C18 satisfying Condition .S/, jdF .ys /u2s;t j C18 .ıR .F; G/ C jy0 z0 j C ı.y; z/ C ı.u; v//!.s; t/2=p : With (26), jdB2 .s; t /j D jd.ys;t f .ys /u1s;t /j D jd.ys;t D.s; t / C F .ys /u2s;t /j MA!.s; t / C jdF .ys /u2s;t j MA!.s; t / C C18 .ıR .F; G/ C jy0 z0 j C ı.y; z/ C ı.u; v//!.s; t /2=p :
Since f is differentiable, for 2 Œ0; 1, Z
1
f . yt C .1 /ys / f .ys / D
rf .ys C . yt C .1 /ys // ys;t d : (27)
0
Hence jf . yt C .1 /ys / f .ys / g. yt C .1 /ys / C g.ys /j ıR .rf; rg/kyk!.s; t/1=p : (28) With (28) and (27) for 2 Œ0; 1, d.f . yt C .1 /ys / f .ys // ıR .rf; rg/kyk!.s; t/1=p C 2kr 2 f k1 .ı.y; z/ C jy0 z0 j/kyk!.s; t/1=p C ı.y; z/krf k1 !.s; t/1=p : With Lemma 5 and (27)–(28) applied to F and G, for 2 Œ0; 1, jdB4 .ys ; ys C ys;t /j D d.F . yt C .1 /ys / F .ys // ıR .rF; rG/kyk!.s; t/1=p C 2krF k1 ı.y; z/!.s; t/1=p C2H .rF /.kyk Ckzk /!.s; t/=p .ı.y; z/Cjy0 z0 j/: Besides, d.rf .ys C ys;t // ıR .rf; rg/ C kr 2 f k1 .ı.y; z/ C jy0 z0 j/: In addition, jd.F .y0 //j ıR .F; G/ C krF k1 jy0 z0 j:
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
235
Combining all these estimates using (23) with the ones given in Sect. 4 in some lengthy computations, jd.D.s; r; t//j C19 !.s; t/ MA C ı.y; z/ C ı.y; z/1
C jz0 y0 j C ı.u; v/ C ıR .F; G/ C ıR .rF; rG/ C ıR .rf; rg/ ; (29)
where C19 satisfies Condition .S/ (Note that C19 decreases with T ). We have also jd.D.s; t//j ıR .f; g/.kyk C jy0 j/kuk!.s; t/1=p C.krf k1 C krF k1 /.ı.y; z/ C jy0 z0 j/kuk.1 C /!.s; t/1=p C.kgk1 C kGk1 /ı.u; v/!.s; t/2=p C ıR .F; G/kuk!.s; t/2=p C20 .ıR .F; G/ C ıR .f; g/ C ı.u; v/ C jy0 z0 j C ı.y; z//!.s; t/1=p ; where C20 satisfies Condition .S/ and decreases when T decreases. With (26), jdys;t j jd.ys;t D.s; t//j C jdD.s; t/j MA!.s; t/
C C20 .ıR .F; G/ C ıR .f; g/ C ı.u; v/ C jy0 z0 j C ı.y; z//!.s; t/1=p : (30) Let us choose A such that an equality holds in either (24) or (25). If jd.ys;t /j D A, then from (30), A MA1C C C20 .ıR .F; G/ C ıR .f; g/ C ı.u; v/ C jy0 z0 j C ı.y; z//: If jd.D.s; r; t//j D A, then from (29), A C19 MA C ı.y; z/ C ı.y; z/1
C jz0 y0 j C ı.u; v/ C ıR .F; G/ C ıR .rF; rG/ C ıR .rf; rg/ :
In any case, we may choose small enough in function of C19 or of M (which depends only on ) such that A 2C19 ı.y; z/ C ı.y; z/1 C jz0 y0 j C ı.u; v/ C ıR .F; G/ C ıR .rF; rG/ C ıR .rf; rg/
C 2C20 .ıR .F; G/ C ıR .f; g/ C ı.u; v/ C jy0 z0 j C ı.y; z//: Injecting this inequality in (30), we get that ı.y; z/ C21 .B C . C 2 C 2.1 / /ı.y; z/ C jy0 z0 j/
236
A. Lejay
with B WD ıR .F; G/ C ıR .f; g/ C ı.u; v/ C ıR .rf; rg/ C ıR .rF; rG/ and C21 depends only on C19 , C20 , M and !.0; T /. Again, choosing small enough in function of C19 , C20 , M and !.0; T / gives the required bound in ı.y; z/ in short time. As the choice of does not depend on jy0 z0 j and satisfies Condition .S/, Proposition 7 may be applied to y z. u t
7 Distance Between Two Euler Schemes Let us give two rough paths u and v of finite p-variations, as well as a partition fti gniD0 of Œ0; T . We use the same notations and conventions as in Sect. 5. Again, we set 1=p WD !0;n . For 0 i < j n, we set zi C1 D zi C g.zi /v1i;i C1 C G.zi /v2i;i C1 ; zj D zi C g.zi /v1i;j C G.zi /v2i;j ;
is finite, we set For a family fi;j gi;j D1;:::;n such that sup0i j n ji;j j=!i;j
yi C1 D yi C f .yi /u1i;i C1 C F .yi /u2i;i C1 C i;i C1 ; i D 0; : : : ; n 1
(31)
yj D yi C f .yi /u1i;j C F .yi /u2i;j C i;j ; 0 i < j n: In addition, define ˛ WD
ji;i C1 j :
i D0;:::;n1 !i;i C1 sup
Here, there are three cases of interest: (a) Both z and y are given by some Euler schemes and then i;j D 0 for all 0 i < j n. (b) The path y is a solution to Rt yt D y0 C 0 f .ys / dus and then i;j D yj yi f .yi /u1i;j F .yi /u2i;j , while v D u and g D f . (c) While v D u and g D f , the family fyi gi D0;:::;n is given by the Euler scheme with respect to a partition fti0 gi D0;:::;m fti gi D0;:::;n . We assume that z and y belong to the ball of radius R and that kzk?;0;n and kyk?;0;n (defined by (14)) are bounded by R0 . In any cases, R and R0 depend only on krf k1 , krF k1 , krgk1 , krGk1 , kuk, kvk, jf .0/j, jg.0/j, jy0 j, jz0 j, !.0; T /, and p. Definition 4. We say that a constant C satisfies condition .Se / if it depends only on the quantities listed above as well as H .rF /, H .rG/, kr 2 f k1 , kr 2 gk1
and sup0i j n ji;j j=!i;j .
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
237
Theorem 2. If f and g satisfies Hypothesis 4, then for some constant C22 satisfying Condition .Se /, kz yk?;0;n C22 .˛ C jz0 y0 j C ıR .f; g/ C ıR .rf; rg/ C ıR .rF; rG/ C ıR .F; G/ C ı.u; v//: Proof. We set jyj yi yi;j zj zi zi;j j :
0i <j n !i;j
A WD max
(32)
We have jF .yi / F .zi /j krF k1 .jy0 z0 j C kz yk?;0;n / and jrf .yi C .yj yi // rf .zi C .zj zi //j 2kr 2 f k1 .jy0 z0 j C kz yk?;0;n /: Besides, with Lemma 5, 1=p
jF .yj / F .yi / F .zj / C F .zi /j krF k1 kz yk?;0;n !i;j
=p
C H .rF /.kyk C kzk /.jy0 z0 j C kz yk?;0;n /!i;j : With (18) and considering the difference between yi;j;k and zi;j;k for i < j < k,
jyi;j;k zi;j;k j C23 !i;k .jy0 z0 j C A C . C 1 /kz yk?;0;n C B1 /:
with C23 satisfying Condition .Se / and B1 WD ı.u; v/ C ıR .F; G/ C ıR .rf; rg/ C ıR .rF; rG/: On the other hand, for i D 0; : : : ; n 1, zi C1 zi zi;i C1 D 0 and
jyi C1 yi yi;i C1 j ˛!i;i C1 :
For j given as in the proof of Lemma 3, we have for K WD 21 ,
238
A. Lejay
jyk yi yi;k zk zi zi;k j
!i;k KA C ˛ C
jyi;j;k zi;j;k j jyj;j C1;k zj;j C1;k j C
!i;k !i;k
KA C ˛ C C24 .jy0 z0 j C A C . C 1 /kz yk?;0;n C B1 / with C24 D 2C23 . From the definition of A (see (32)), A 2˛ C KA C C24 .jy0 z0 j C A C . C 1 /kz yk?;0;n C B1 /:
(33)
On the other hand, since both f and F are Lipschitz continuous, for some constant C25 satisfying Condition .Se /, kz yk?;0;n
sup
jyj yi yi;j zj C zi C zi;j j
0i <j n
1=p !i;j
C
jyi;j zi;j j 1=p
!i;j
A1C C C25 .jy0 z0 j C kz yk?;0;n / C B2 with B2 WD C26 .ıR .F; G/ C ıR .f; g/ C ı.u; v//; for some constant C26 satisfying Condition .Se /. If is small enough so that C25 1=2, then kz yk?;0;n 2C25 jy0 z0 j C 2A1C C 2B2 : (34) Injecting this in (33), A 2˛ C KA C C27 .jy0 z0 j C AC28 .// C B3 with a C28 ./ satisfying Condition .Se / for fixed and that decreases to 0 as decreases to 0, C27 satisfying Condition .Se /, and B3 D C29 .B1 C B2 / for some constant C29 satisfying Condition .Se /. For K C C27 C28 ./ < .1 C K/=2 < 1, then A
2 .2˛ C C27 jy0 z0 j C B3 /: 1K
Using the inequality on (34), this leads to the required inequality for a value of small enough. Thus usual arguments proves now that this is true for any time horizon T up to changing the constants. t u
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
239
8 Rate of Convergence of the Euler Scheme Let us consider now the solution to yt D y0 C Euler scheme
Rt 0
f .ys / dxs as well as the associated
1 2 ei C1 D ei C f .ei /xi;i C1 C F .ei /xi;i C1 ; e0 D y0
when f satisfies Hypothesis 4. We are willing to estimate the difference between y and e. Our key argument is provided by Theorem 2 above. Proposition 4. Assume Hypothesis 4 on f . For ı WD sup0i
where C30 depends only on y0 , jf .0/j, kxk, krf k1 , kr 2 f k1 , krF k1 , H .rF /, !.0; T /, p and . It follows that the rate of convergence is smaller than .3 p/=p and belongs to .0; 1=2/. In addition, when p increases to 3, the rate of convergence decreases to 0. This rate is similar to the one given by A.M. Davie [2]. (See also [3] for the convergence of the Milstein scheme for the fractional Brownian motion). Proof. Let us note first that since y and e remains bounded, if rF is -Hölder continuous, then it is locally 0 -Hölder continuous for any 0 < . Since the constraint 2 C 0 > p is in force, we set 0 D p 2 C p for some 0 < < .3 p/=p < 1=2 and 0 WD .2 C 0 /=p D 1 C . Since F is Lipschitz continuous, then y is a solution in the sense of Davie with
D 3=p. This way, 1 2 jyt ys f .ys /xs;t F .ys /xs;t j L!.s; t/3=p :
It follows that fyi gniD0 is solution to (31) with 1 2 i;i C1 D yi C1 yi f .yi /xi;i C1 F .yi /xi;i C1 :
Then
ji;i C1 j 0 Lı 3=p D Lı .3p/=p :
0 0i
˛ WD sup
Applying Theorem 2 with our choice of 0 leads to ke yk C31 ı .3p/=p ; where C31 depends only on y0 , jf .y0 /j, kxk, krf k1 , kr 2 f k1 , krF k1 , H .rF /, !.0; T /, p and . Then (35) is immediate. t u
240
A. Lejay
Let fftin gi D0;:::;n gn2N be an increasing family of partitions such that the mesh supi D0;:::;n !.tin ; tinC1 / converges to 0. Let ffeingi D0;:::;n gn2N be the corresponding Euler schemes for a rough path x and a vector field f satisfies Hypothesis 4. With Lemmas 3, 4 and 8, sup
n2N
sup
jein ejn f .ein /xt1n ;tj F .ein /xt2n ;t n j i
i
!.tin ; tjnC1 /
0i <j n
j
< C1
and with Proposition 3 under Hypothesis 3, the interpolation of the Euler scheme e n constructed as in (21) has a p-variation norm ke n k which is bounded. Using the same proof as above, we get the following corollary. Corollary 2. Under Hypotheses 2, 3 and 4, the family of interpolated Euler schemes fe n gn2N for a Cauchy sequence of the uniform norm and the q-variation norm for any q > p. Remark 2. Here the dimension d of the space plays no role so that this argument may be used for an infinite dimensional rough path. This Corollary allows one to define the solution to (1) as the limit of the sequence fe n gn2N .
9 Case of Geometric Rough Paths We now consider the case where x is a geometric rough path of finite p-variation controlled by ! and the vector field f satisfies Hypothesis 4. This means that there exists a sequence of rough paths x n converging to x in p-variation that x n lives R t such n d n n n above a piecewise smooth path z in R and xt D 1 C zt C 0 .zs zn0 / ˝ zns ds. Such a path x n is called a smooth rough path. In order to simplify the analysis, we assume that !.s; t/ D t s and then we are dealing with ˛-Hölder continuous paths with ˛ D 1=p. The core idea of P. Friz and N. Victoir was to consider a family of smooth rough paths .e x n /n1 such that, given a family of partitions fftin gi D0;:::;n gn2N , e x n converges n to x in the ˇ-Hölder norm for any ˇ < ˛ and e x t n D xtin for i D 0; : : : ; n. i For this, they used sub-Riemannian geodesics. In [13], we have provided an alternative construction using some segments and some loops. Let e zn beRthe projection of e x n onto Rd , and let y n be the solution of the ODE t n n n yt D y0 C 0 f .ys / de zs . As e zn is piecewise smooth, one knows that this solution Rt x ns in the corresponds to the one of the solution of the RDE ytn D y0 C 0 f .ysn / de sense of Davie or in the sense of Lyons (See Sect. 10 below). Let e e n be the Euler scheme associated to y n with the partition ftin gi D0;:::;n : e ni C f .e e ni /e x 1;n e e niC1 D e t n ;t n i
and e n the Euler scheme
i C1
C F .e e ni /e x 2;n t n ;t n ; i D 0; : : : ; n; i
i C1
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
einC1 D ein C f .ein /xt1;n n ;t n i
i C1
241
C F .ein /xt2;n n ;t n ; i D 0; : : : ; n: i
i C1
The key observation from P. Friz and N. Victoir is that from the very construction of e xn, e x ntn ;t n D e x ntn ;t n and then e en D en . i i C1 i i C1 With Proposition 4, for some 2 .0; 3˛ 1/, ke n y n k? WD
sup
jejn ein yjn C yin j
0i <j n
1=p !i;j
D ke e n y n k C32 ı 3˛1
where C32 depends only on kxk, f , T , ˛ and , and ı WD supi D0;:::;nC1 !.tin ; tinC1 /. On the other hand with Theorem 1, ky n y m k? ky n y m k C33 ke xm e x n k; x n is such that ke x n k Akxk where C33 depends only on kxk, f , T , ˛ and (when e which is the case with the constructions mentioned above). Since .e x n /n2N is a Cauchy sequence in the space of ˇ-Hölder continuous functions, ˇ < ˛, it follows that .y n / is a Cauchy sequence and converges to some element y. We then obtain the convergence of e n to y in the sense that ke n yk? decreases to 0 as n tends to infinity.
10 Solution in the Sense of Davie and Solution in the Sense of Lyons The notion of solution in the sense of Davie (Definition 2) is different of the solution in the sense of Lyons (Definition 1), as the iterated integrals of y and the crossiterated integrals between y and x are not constructed, while they are part of the solution in the sense of Lyons. However, once a solution in the sense of Davie is constructed, it is easy to construct a rough paths with values in T1 .Rd ˚ Rm /. Lemma 6. A solution y in the sense of Davie — which is a path with values in Rm — may be lifted to a rough path with values in T1 .Rd ˚ Rm /, as the rough path e y associated to the almost rough path 2 2 2 hs;t WD 1 Cxs;t Cys;t Cf .ys /˝f .ys / xs;t Cf .ys /˝1 xs;t C1 ˝f .ys / xs;t : (36)
Besides, the map y 7! e y is locally Lipschitz continuous. Proof. It is easily checked that jhs;r;t hs;r ˝ hr;t j C34 !.s; t/
242
A. Lejay
with > 1 and then that it is an almost rough path. With Proposition 1, the associated rough path e y projects onto .x; y/ in Rd ˚ Rm . The local Lipschitz continuity of y 7! e y follows from the same kind of computation as the one of the proof of Theorem 1. t u Proposition 5. Let us assume that f is a vector field with a bounded derivative which is -Hölder continuous, such that a solution to (1) exists in the sense of Lyons. Let us assume also that f rf is -Hölder continuous, so that a solution to (1) exists in the sense of Davie, lifted as a rough path with values in T1 .Rd ˚ Rm / as above. Then the two solutions coincide. Proof. Let y be a solution in the sense of Davie of (1), which is lifted as a rough path e y using Lemma 6. Let us consider 1 us;t WD f .ys /xs;t C rf .ys /e y s;t ; m d where e y y s;t (or roughly speaking, it is the s;t is the projection onto R ˝ R of e cross-iterated integral between y and x). 1 Since e y s;t D ys;r ˝ xr;t , 1 1 jus;t us;r ur;t j j.f .yr / f .ys //xr;t rf .ys /ys;r ˝ xr;t /j C jrf .yr / rf .ys /jje y s;r j ˇ ˇZ 1 ˇ ˇ 1
d ˇˇ C H .rf /kyk kzk!s;t ˇˇ .rf .ys C ys;r / rf .ys //ys;r ˝ xr;t 0
H .rf /.kyk1C kxk C kyk kzk/!.s; t / :
Then fus;t g.s;t /2 2 is an almost rough path in Rm whose associated rough path v satisfies jvt vs us;t j C35 !.s; t/ : Rt From the very construction, v is the rough integral vt D v0 C 0 g.zs / dzs with z is a rough path lying above .x; y; y / 2 .R ˚ Rd ˚ Rd ; Rm ; Rm ˝ Rd / and g.y; x/ D dx C f .y/ dx. Thus y D v when v0 D y0 . With Lemma 6, this is also true for the iterated integrals. This proves that the a solution in the sense of Davie is also a solution in the sense of Lyons. If z is a solution in the sense of Lyons, by construction, for a constant L and all .s; t/ 2 2 , 1
jzt zs f .zs /xs;t rf .zs /z s;t j L!.s; t/ 2
and jz s;t 1 ˝ f .zs /xs;t j L!.s; t/ ;
where z lives in Rm ˝ Rd . Hence, it is immediate the a solution in the sense of Lyons is also a solution in the sense of Davie. u t
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
243
11 From Local to Global Theorems We present here some results which allows one to pass from local estimates to global estimates and then show the existence for any horizon T provided some uniform estimates. Lemma 7. Let > 0 be fixed. Then there exists a finite number of times 0 D T0 < T1 < < TN C1 with TN T < TN C1 and !.Ti ; Ti C1 / D . Proof. Let us extend ! on DC WD f.s; t/ j 0 s tg by ( !.s; t/ D
!.s; T / C T s
if s T t;
t s
if T s t:
Let us note that ! is continuous on DC , and that !.s; t/ ! C1. Then for any t !1
> 0, for any s 0, there exists a value .s/ such that !.s; .s// D , since !.s; s/ D 0. For T > 0 fixed, let us set T0 D 0 and Ti C1 D .Ti /. Then there exists a finite number N such that !.0; TN 1 / !.0; T / !.T0 ; TN C1 / and !.Ti ; Ti C1 / D . This follows from the fact that N D
N 1 X
!.Ti ; Ti C1 / !.0; TN /
i D0
and then !.0; TN / ! C1. This proves the lemma. N !1
t u
For a path y and times 0 S < S 0 T , let us set kykŒS;S 0 WD
sup
S s
jyt ys j : !.s; t/1=p
Lemma 8. Let us assume that a continuous path y on ŒS; S 00 is a solution in the sense of Davie on time interval ŒS; S 0 and ŒS 0 ; S 00 respectively with constants L1 and L2 . Then y is a solution in the sense of Davie on ŒS; S 00 with a new constant L3 that depends only on L1 , L2 , H .F /, krf k1 , kxk, kykjŒS;S 0 , !.0; T /, and p. Proof. First, it is classical that kykŒS;S 00 211=p maxfkykŒS;S 0 ; kykŒS 0 ;S 00 g; so that y is of finite p-variation over ŒS; S 00 . For s 2 ŒS; S 0 and t 2 ŒS 0 ; S 00 , jyt ys D.s; t/j jyt yS 0 D.S 0 ; t/j C jyS 0 ys D.s; S 0 /j C jD.s; S 0 ; t/j:
244
A. Lejay
Let us note that
jD.s; S 0 ; t/j C36 !.s; t/
where C36 depends only on L1 , kykjŒS;S 0 , H .F /, krf k1 , jF .ys /j, kxk, !.0; T /, and p. Hence, since !.s; S 0 / C !.S 0 ; t/ !.s; t/, jyt ys D.s; t/j .maxfL1 ; L2 g C C36 /!.s; t/ : t u
This proves the result.
Proposition 6. Let us assume that a solution to (1) exists on any time interval ŒS; S 0 with a condition that !.S; S 0 / C37 where C37 does not depend on S . Then a solution exists for any time T . Proof. Using the sequence fTi gi D0;:::;N of times given by Lemma 7, it is sufficient to solve (1) successively on ŒTi ; Ti C1 with initial condition yTi and to invoke Lemma 8 to prove the existence of a solution in the sense of Davie in Œ0; T . t u Proposition 7. Let y be a path of finite p-variations such that for some constants A, B and K, when !.S; S 0 / K; kykŒS;S 0 BjyS j C A: Then sup jyt j R.T /jy0 j C .R.T / 1/
t 2Œ0;T
with
A B
(37)
R.T / D exp.B.1 C K 1 /11=p maxf!.0; T /; !.0; T /1=p g/;
and the p-variation norm of kykŒ0;T depends only on A, B, !.0; T /, K and p. Proof. Remark first that sup jyt j jyS j.1 C B!.S; S 0 /1=p / C A!.S; S 0 /1=p :
t 2ŒS;S 0
Let us choose an integer N such that WD !.0; T /=N K and construct recursively Ti with T0 D 0 and !.Ti ; Ti C1 / D . Then jyTi C1 j jyTi j.1 C 1=p B/ C A1=p : From a classical result easily proved by an induction, jyT j exp.N1=p B/jy0 j C As this is true for any T , surely (37) holds.
A .exp.N1=p B/ 1/: B
Global Solutions to Rough Differential Equations with Unbounded Vector Fields
245
Choosing N such that !.0; T /=N K and !.0; T /=.N 1/ > K, N
!.0; T / C 1: K
This way 1=p
N
!.0; T / C1 K
1 p1
( !.0; T /
e WD .1 C K 1 /11=p . with K Finally, kykŒ0;T N 11=p which proves the last statement
1 p
max
e K!.0; T /1=p e K!.0; T/
i D0;:::;N 1
if !.0; T / 1; if !.0; T / > 1:
kykŒTi ;Ti C1 ; t u
Acknowledgements This work has been supported by the ANR ECRU founded by the French Agence Nationale pour la Recherche. The author is also indebted to Massimiliano Gubinelli for interesting discussions about the topic. The author also wish to thank the anonymous referee for having suggested some improvements in the introduction.
References 1. S. Aida, Notes on proofs of continuity theorem in rough path analysis, 2006. Preprint of Osaka University 2. A.M. Davie, Differential equations driven by rough paths: an approach via discrete approximation. Appl. Math. Res. Express. AMRX 2(abm009), 40 (2007) 3. A. Deya, A. Neuenkirch, S. Tindel, A Milstein-type scheme without Levy area terms for sdes driven by fractional brownian motion, 2010. arxiv:1001.3344 4. H. Doss, Liens entre équations différentielles stochastiques et ordinaires. Ann. Inst. H. Poincaré Sect. B (N.S.) 13(2), 99–125 (1977) 5. P. Friz, N. Victoir, Euler estimates of rough differential equations. J. Differ. Eqn. 244(2), 388–412 (2008) 6. P. Friz, N. Victoir, Multidimensional Stochastic Processes as Rough Paths. Theory and Applications (Cambridge University Press, Cambridge, 2010) 7. J.G. Gaines, T.J. Lyons, Variable step size control in the numerical solution of stochastic differential equations. SIAM J. Appl. Math. 57(5), 1455–1484 (1997) 8. M. Gubinelli, Controlling rough paths. J. Funct. Anal. 216(1), 86–140 (2004) 9. Y. Inahama, H. Kawabi, Asymptotic expansions for the Laplace approximations for Itô functionals of Brownian rough paths. J. Funct. Anal. 243(1), 270–322 (2007) 10. T. Lyons, M. Caruana, T. Lévy, Differential equations driven by rough paths. In École d’été des probabilités de Saint-Flour XXXIV — 2004, ed. by J. Picard, vol. 1908, Lecture Notes in Mathematics (Springer, New York, 2007) 11. A. Lejay, An introduction to rough paths. In Séminaire de probabilités XXXVII, vol. 1832 of Lecture Notes in Mathematics (Springer, New York, 2003), pp. 1–59 12. A. Lejay. On rough differential equations. Electron. J. Probab. 14(12), 341–364 (2009)
246
A. Lejay
13. A. Lejay, Yet another introduction to rough paths. In Séminaire de probabilités XLII, vol. 1979 of Lecture Notes in Mathematics (Springer, New York, 2009), pp. 1–101 14. A. Lejay, Controlled differential equations as Young integrals: a simple approach. J. Differ. Eqn. 249, 1777–1798 (2010) 15. T. Lyons, Z. Qian, Flow of diffeomorphisms induced by a geometric multiplicative functional. Probab. Theor. Relat. Field 112(1), 91–119 (1998) 16. T. Lyons, Z. Qian, System Control and Rough Paths, Oxford Mathematical Monographs (Oxford University Press, Oxford, 2002) 17. A. Lejay, N. Victoir, On .p; q/-rough paths. J. Differ. Eqn. 225(1), 103–133 (2006) 18. T.J. Lyons, Differential equations driven by rough signals. Rev. Mat. Iberoamericana 14(2), 215–310 (1998) 19. D.S. Mitrinovi´c, J.E. Peˇcari´c, A.M. Fink, Inequalities involving functions and their integrals and derivatives, vol. 53 of Mathematics and its Applications (East European Series). (Kluwer, Dordrecht, 1991) 20. H.J. Sussmann, On the gap between deterministic and stochastic ordinary differential equations. Ann. Probability 6(1), 19–41 (1978)
Asymptotic Behavior of Oscillatory Fractional Processes Renaud Marty and Knut Sølna
Abstract In this paper we consider the antiderivative of the product of a fractional random process and a periodic function. We establish that the rescaled process constructed in this way converges to a Brownian motion whose variance depends on the frequency of the periodic function and the Hurst parameter. We also prove that for two different frequencies the limits are independent. Finally, we discuss applications to wave propagation in random media. Keywords Fractional processes • Brownian motion • Waves in random media
AMS Classification: 60F17, 60G10, 60G15
1 Introduction Fractional processes have attracted a lot of attention in recent years. They provide relevant models for long-range dependence or antipersistent random behavior [11]. Fractional Brownian motion is the most famous fractional process. It satisfies many relevant properties such as self-similarity, Gaussianity and stationarity of the increments.
R. Marty () Institut Elie Cartan de Nancy, Nancy-Université, CNRS, INRIA, B.P. 239, F-54506 Vandoeuvre-lès-Nancy Cedex, France e-mail: [email protected] K. Sølna University of California at Irvine, Irvine, CA 92697-3875, USA e-mail: [email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__12, © Springer-Verlag Berlin Heidelberg 2012
247
248
R. Marty and K. Sølna
Another crucial property for applications is the invariance principle [11]. Let H 2 .0; 1/, BH D fBH .z/gz be a fractional Brownian motion with Hurst index H . Depending on the value of H , the fractional white noise ıBH D fBH .z C 1/ BH .z/gz (which is stationary) satisfies various “memory-properties”. For H 2 .1=2; 1/ the process ıBH is long-range dependent, for H D 1=2 it is mixing, and for H 2 .0; 1=2/ it is antipersistent. Let consider a stationary and Gaussian process D f.z/gz that satisfies roughly the same range property as the fractional white noise ıBH , that is, it has the same decay rate of the integrated correlation. Then, the invariance principle says that when " goes to 0, the rescaled process (
Z "
)
z="2
2H
D
.y/dy 0
z
1 "22H
Z
z
0
y "2
dy z
converges in distribution to the fractional Brownian motion BH . Notice that the invariance principle is of great importance. In particular it makes the increments of BH serve as universal stationary Gaussian models with various memory properties depending on the index H [11]. Applications of fractional processes to wave propagation in random media have recently been studied in [5, 9, 10]. Here they are used to model very complex and long range media. In addition to random phenomena, due to the nature of waves, periodic modulation arises naturally in harmonic descriptions associated with wave phenomena [2]. This then leads to the study of evolution problems driven by periodic and random processes. A simple example is the asymptotic study of the process defined by v"p
D
1 "
Z
z 0
y y 2 p dy " " z
(1)
where and are two positive constants and p is a periodic function. Such a question have been investigated in particular cases in [8]. It was proven that if p D cos.!/ (or sin.!/), 2 .0; 2/, D 1 C .2 /.1=2 H / and is a Gaussian and stationary process that is essentially a fractional white noise, then v"p converges in distribution to a Brownian motion. The result with H D 1=2 had been known for many years and applied to wave propagation in mixing media, see [2], while the results with H 2 .0; 1/ allowed for the study of waves in fractional media. In the present paper we aim to generalize the result presented above. Our framework is based on a Gaussian process defined in terms of a general harmonizable representation. We first consider functions p D cos.!/ (or sin.!/) and 2 .0; 2/. The proof of the convergence of v"p in [8] was based on the particular form of the covariance of and the cases H < 1=2, H D 1=2 or H > 1=2 had to
Asymptotic Behavior of Oscillatory Fractional Processes
249
be considered separately. Here we give a new proof that unifies the three cases for H and that is based on the harmonizable representation and thus more general. We also consider general period functions p. We prove that under general assumptions on p the process converges to a Brownian motion whose covariance depends on the Fourier coefficients of p. A second point we address is the correlation of the limits for different frequencies. Let !1 and !2 be two real numbers satisfying !1 6D !2 and !1 6D !2 . It is well known in the mixing case (H D 1=2), that if we denote respectively by B !1 and B !2 the limits of v"p with p D cos.!1 / and p D cos.!2 / respectively, then the limits B !1 and B !2 are independent [2]. This asymptotic decorrelation is a crucial property in the applications to waves. Here we generalize this result to every H 2 .0; 1/. The assumption 2 .0; 2/ corresponds to models for wave propagations where the phases are slowly varying with respect to the random fluctuations [2]. Another problem of interest is when phases oscillate at the same speed as the random medium fluctuations. Physically this corresponds to a strong “resonance” between the waves and the fluctuations. This motivates studying the case D 2 and then, in the mixing case, one obtains different results from those of the case R2 .0; 2/. In particular, 1 the effective second R 1 moment coefficient can be written as 0 EŒ.0/.y/dy if 2 .0; 2/ and 0 cos.!y/EŒ.0/.y/dy if D 2 [2]. In the present paper we analyze the specific behavior when D 2 for the full range of H 2 .0; 1/. In conclusion, we discuss some possible extensions of our work for applications to wave propagation. In particular we give a conjecture that would generalize the study of wave propagation in mixing [1, 2] or long-range [9] media to a general fractional media. Our paper is organized as follows. In Sect. 2 we establish the main results and we prove them in Sect. 3. In Sect. 4 we extend our results to general periodic functions. In Sect. 5 we discuss applications to the study of wave propagation in random media.
2 Assumptions and Main Result 2.1 Assumptions Here we describe the process that we will be working with throughout the paper. We define for every z 0 by Z
b .d / exp.i z/ ./jj1=2H W
.z/ D
(2)
R
b .d / is the Fourier where H 2 .0; 1/ and is a complex-valued function and W transform of a real Gaussian measure. We assume that is continuous and even
250
R. Marty and K. Sølna
and satisfies j ./j D Ojj!1 .jj1 /. Thus is a centered and stationary Gaussian process and its covariance takes the form Z r.z/ D EŒ.0/.z/ D
exp.i z/j ./j2 jj12H d :
(3)
R
An integration with respect to z gives Z
Z
r.z/ d z .H /Z 2H 1
(4)
0
Z
as Z!1 with
1
.H / D 2j .0/j2 0
sin./ d : 2H
We deduce from (4), depending onR the value of H , that the process R satisfies one 1 1 of the three following properties: 0 r.z/ d z D 1 if H 2 .1=2; 1/, 0 r.z/ d z 2 R1 .0; 1/ if H D 1=2 and 0 r.z/ d z D 0 if H 2 .0; 1=2/. In other words is long-range dependent when H > 1=2, antipersistent when H < 1=2 and satisfies a general mixing property when H D 1=2. We conclude this section by two examples of processes satisfying the assumptions cited above. These examples are defined in terms of the (non-standard) fractional Brownian motion defined for every z by 1 BH .z/ D .H /
Z R
e i z 1 b W .d / ijjH 1=2
where .H / is a normalizing constant defined by Z .H / D 2
R
je i 1j2 d : jj2H C1
The first example we present is the fractional white noise ıBH defined as 1 ıBH .z/ D BH .z C 1/ BH .z/ D .H /
Z exp.i z/ R
e i 1 b W .d /: ijjH 1=2
In this example ./ D
e i 1 : .H /i
The second example is the fractional Ornstein-Uhlenbeck (OU) process. Recall that the stationary fractional OU process XH can be written in terms of a the fractional Brownian motion BH as
Asymptotic Behavior of Oscillatory Fractional Processes
XH .z/ D BH .z/ e 1 D .H / so the corresponding function
Z R
z
Z
251
z 1
e y BH .y/ dy
b .d / exp.i z/ W 1 C i jjH 1=2
is ./ D
1 : .H /.1 C i/
2.2 Main Result Here we consider a general process as defined in the previous section and throughout the paper we assume that H 2 .0; 1/. We introduce the two processes !;" v!;" c and vs defined by v!;" c .z/
1 D "
Z
z
0
y "2
y cos ! dy "
(5)
y sin ! dy "
(6)
and v!;" s .z/ D
1 "
Z
z
0
y "2
with defined as in (2), ! D 6 0, while and are constants and " is a small parameter. We assume that 2 .0; 2 and is given by D 1 C .2 /.1=2 H /:
(7)
Notice that in the particular case D 2 we have D 1. We fix an n-dimensional vector of frequencies ˝ D .!1 ; :::; !n / such that !j 6D 0 for every j and !j 6D !k and !j 6D !k for every j 6D k. We define the 2ndimensional process V ˝;" by V ˝;" D vc!1 ;" ; vs!1 ;" ; vc!2 ;" ; vs!2 ;" ; ; vc!n ;" ; vs!n ;" :
(8)
We state the first main theorem of this paper. Theorem 1. The finite-dimensional distributions of the process V ˝;" converge to those of a 2n-dimensional Brownian motion B ˝ defined by B ˝ D Bc!1 ; Bs!1 ; Bc!2 ; Bs!2 ; ; Bc!n ; Bs!n
252
R. Marty and K. Sølna
whose coordinates are independent and satisfy !
!
EŒ.Bc j .1//2 D EŒ.Bs j .1//2 D C.!j /2
(9)
for every j and where the second moment coefficient is C.!/2 D
e.!/ Z 2
1
je i 1j2 d D e.!/ jj2
1
(10)
with e.!/ D j .0/ j2 j!j12H if 2 .0; 2/ and e.!/ D j .!/ j2 j!j12H if D 2. Notice that the cases 2 .0; 2/ (the slow-phase case) and D 2 (the fast-phase case) are dramatically different. The second moment depends on the whole function if D 2 while it does not if 2 .0; 2/. When D 2 the analogue in the wave case is that the wave resolves the full spectrum of the fluctuations. This theorem generalizes the diffusion-approximation theory set forth in [2]. Indeed, if we consider the case H D 1=2, we have Z
Z
A 0
0
C
Z
exp.i z/j ./j2 d d z
cos.!z/
1 D 2
Then, because
Z
A
r.z/ cos.!z/ d z D
R
Z
j .=A C !/j2 R
1 2
Z
e i 1 d i
j .=A !/j2 R
e i 1 d : i
is continuous and even we get Z
1
r.z/ cos.!z/ d z D j .!/j 0
2 R
sin./ d D j .!/j2 :
It then follows that the constant C.!/ (depending on the value of ) corresponds to the effective coefficients found in the diffusion-approximation presented in [2]. To conclude this section we make a remark about the super-fast case that corresponds to > 2. We can observe, following the lines of the proof of the theorem, that the process V ˝;" converges to 0R2n as already observed in the classical approximation-diffusion theory (for H D 1=2, see for instance [3]).
3 Proof of Theorem 1 In the first part of the proof we introduce some notation. Next we establish that, for !;" a fixed frequency !, the 2-dimensional process .v!;" c ; vs / defined by (5) and (6) converges in the sense of the finite dimensional distributions to the 2-dimensional
Asymptotic Behavior of Oscillatory Fractional Processes
253
Brownian motion .Bc! ; Bs! / defined by (9). In the third part we prove that for two different frequencies !1 and !2 the limits .Bc!1 ; Bs!1 / and .Bc!2 ; Bs!2 / are independent. In the proof we use technical lemmas whose derivations are postponed to the end of this section.
3.1 Preliminaries We introduce the complex-variable function g defined for every u 2 C by g.u/ D
exp.u/ 1 : u
By a direct integration we obtain Z
z
exp 0
and Z z
exp
0
y z ! ! g iz 2 cos ! dy D C g iz 2 C " 2 " " " "
y iz ! ! g iz 2 sin ! dy D g iz 2 C : " 2 " " " "
iy "2
iy "2
Then we can deduce that for every !, " and z we have Z v!;" c .z/
D R
b .d / gc!;" .z; /W
(11)
b .d / gs!;" .z; /W
(12)
and Z v!;" s .z/
D R
where gc!;" .z; /
./ z ! ! C g iz 2 C D H 1=2 g i z 2 2" jj " " " "
gs!;" .z; /
iz ! ! ./ g iz 2 C : D H 1=2 g i z 2 2" jj " " " "
and
It is convenient to introduce the integral
254
R. Marty and K. Sølna
I1" .!1 ; !2 ; z1 ; z2 /
z1 z2 D 2 4"
Z R
j ./j2 !1 !2 g iz1 2 g iz2 2 d ; jj2H 1 " " " "
and I2" .!1 ; !2 ; z1 ; z2 / D I1" .!1 ; !2 ; z1 ; z2 /; I3" .!1 ; !2 ; z1 ; z2 / D I1" .!1 ; !2 ; z1 ; z2 /; I4" .!1 ; !2 ; z1 ; z2 / D I1" .!1 ; !2 ; z1 ; z2 /:
3.2 Convergence at a Fixed Frequency !;" Now we can state the convergence of .v!;" c ; vs /. !;" Lemma 1. The finite dimensional distributions of .v!;" c ; vs / converge to those of ! ! .Bc ; Bs /.
Proof. We first consider v!;" c . For every z1 and z2 we have Z !;" EŒv!;" c .z1 /vc .z2 / D
D
R
gc!;" .z1 ; /gN c!;" .z2 ; / d
4 X
Ij" .!; !; z1 ; z2 /:
j D1
By the substitution !"2 C !"2 we can write I1" .!; !; z1 ; z2 /
z1 z2 D 4
ˇ2 Z ˇˇ 2 " C !"2 ˇ R
j" C !j2H 1
g .i z1 / g .i z2 / d
where we use D 1 C .2 /.1=2 H /. It now follows from Lemma 3 below that we have the limit lim
ˇ2 Z ˇˇ 2 " C!"2 ˇ
"!0 R
j" C !j2H 1
g .i z1 / g .i z2 / d D e.!/
Z R
g .i z1 / g .i z2 / d
where e.!/ D j .0/ j2 j!j12H if 2 .0; 2/ and e.!/ D j .!/ j2 j!j12H if D 2. We deal with the integral I4" .!; !; z1 ; z2 / in the same way so that we get lim .I1" .!; !; z1 ; z2 / C I4" .!; !; z1 ; z2 // D
"!0
z1 z2 e.!/ 2
Z R
g .i z1 / g .i z2 / d : (13)
Asymptotic Behavior of Oscillatory Fractional Processes
255
Now we deal with I2" .!; !; z1 ; z2 /. By the same substitution as for I1" .!; !; z1 ; z2 / we have Z j "2 C !"2 j2 z1 z2 " I2 .!; !; z1 ; z2 / D g .i z1 / g .i z2 . C 2!" // d : 4 R j" C !j2H 1 It is now enough to prove that the last integral converges to 0 and this is due to Lemma 4 below. We deal with the integral I3" .!; !; z1 ; z2 / in the same way so that we obtain again by Lemma 4 lim .I2" .!; !; z1 ; z2 / C I3" .!; !; z1 ; z2 // D 0
"!0
(14)
and thus, using (13), !;" lim EŒv!;" c .z1 /vc .z2 / D
"!0
z1 z2 e.!/ 2
Z R
g .i z1 / g .i z2 / d ;
which means that the finite dimensional distributions of v!;" c converge to those of the Brownian motion Bc! with the scaling (9). Using the same argument we also prove that the finite dimensional distribution of v!;" converge to those of the Brownian s motion Bs! with scaling (9). Now it remains to establish the independence between Bc! and Bs! . We can write for every z1 and z2 Z !;" EŒv!;" c .z1 /vs .z2 /
D R
Di
gc!;" .z1 ; /gN s!;" .z2 ; / d
4 X
.1/j Ij" .!; !; z1 ; z2 /:
j D1
From the asymptotic study of Ij" .!; !; z1 ; z2 / for j 2 f1; 2; 3; 4g just above, we get !;" lim EŒv!;" c .z1 /vs .z2 / D 0;
"!0
which proves that Bc! and Bs! are independent and this concludes the proof.
t u
3.3 Asymptotic Decorrelation for Different Frequencies Now we establish the asymptotic decorrelation between the vectors .vc!1 ;" ; vs!1 ;" / and .vc!2 ;" ; vs!2 ;" / when !1 6D !2 and !1 6D !2 . Lemma 2. Let z1 and z2 2 R and !1 and !2 such that !1 6D !2 and !1 6D !2 . We have
256
R. Marty and K. Sølna
lim E
"!0
vc!1 ;" .z1 /vc!2 ;" .z2 / vc!1 ;" .z1 /vs!2 ;" .z2 / vs!1 ;" .z1 /vc!2 ;" .z2 / vs!1 ;" .z1 /vs!2 ;" .z2 /
D
00 : 00
Proof. We consider !1 6D !2 and !1 6D !2 . To prove the lemma it is sufficient to establish the limit of EŒvc!1 ;" .z1 /vc!2 ;" .z2 / and EŒvc!1 ;" .z1 /vs!2 ;" .z2 /. The asymptotic study of EŒvs!1 ;" .z1 /vs!2 ;" .z2 / and EŒvs!1 ;" .z1 /vc!2 ;" .z2 / respectively will be similar. Let us start by EŒvc!1 ;" .z1 /vc!2 ;" .z2 /. We have EŒvc!1 ;" .z1 /vc!2 ;" .z2 / D
4 X
Ij" .!1 ; !2 ; z1 ; z2 /
j D1
As in the case of I2" .!; !; z1 ; z2 / in the proof of Lemma 1 we can write by the substitution !"2 C !1 "2 I1" .!1 ; !2 ; z1 ; z2 / Z j "2 C !1 "2 j2 z1 z2 D g .i z1 / g .i z2 . C .!1 !2 /" // d : 4 R j" C !1 j2H 1 Using Lemma 4 and the condition !1 6D !2 we get lim"!0 I1" .!1 ; !2 ; z1 ; z2 / D 0. Using the same approach and also either !1 6D !2 or !1 6D !2 we obtain lim"!0 Ij" .!1 ; !2 ; z1 ; z2 / D 0 for j D 2, 3 and 4 also, thus lim EŒvc!1 ;" .z1 /vs!2 ;" .z2 / D 0:
"!0
It remains to study EŒvc!1 ;" .z1 /vs!2 ;" .z2 /. We have Z EŒvc!1 ;" .z1 /vs!2 ;" .z2 / D Di
R
gc!1 ;" .z1 ; /gN s!2 ;" .z2 ; / d
4 X
.1/j Ij" .!1 ; !2 ; z1 ; z2 /:
j D1
Using the same calculation as just above for EŒvc!1 ;" .z1 /vc!2 ;" .z2 / we get lim EŒvc!1 ;" .z1 /vs!2 ;" .z2 / D 0;
"!0
which concludes the proof.
t u
Asymptotic Behavior of Oscillatory Fractional Processes
257
3.4 Technical Lemmas Lemma 3. We define e.!/ D j .0/ j2 j!j12H if 2 .0; 2/ and e.!/ D j .!/ j2 j!j12H if D 2. We have the following convergence lim
ˇ2 Z ˇˇ 2 " C!"2 ˇ
"!0 R
j" C !j2H 1
g .i z1 / g .i z2 / d D e.!/
Z R
g .i z1 / g .i z2 / d
for every z1 and z2 2 R and ! 6D 0. Proof. Without loss of generality and for simplicity of presentation we assume that ! > 0. We define for a and b in Œ1; 1 the "-dependent integral Z
b
S .a; b/ D "
j
2 " C !"2 j2 j" C !j2H 1
a
g .i z1 / g .i z2 / d
then, we need to study the convergence of S " .1; 1/. We fix > 0 sufficiently small and write S " .1; 1/ D S " .1; " / C S " . " ; " / C S " . " ; 1/: By means of the substitution !=" we get
Z
S .1; " / D " "
2 " . C !/ j2 i z1 " i z2 " 1 e 1 d e z1 z2 2 j C !j2H 1
j
1
so that jS " .1; " /j "
Z
1
k k21 z1 z2 2 j C !j2H 1
d !0
as "!0. We deal with S " . " ; 1/ in the same way and therefore it remains only to study S " . " ; " /. We have S " . " ; " / D
Z
1 1
" !;
./d
where " !;
./
D 1Œ " ; " ./
For !, and fixed we have
j
"2 C !"2 j2
j" C !j2H 1
g .i z1 / g .i z2 / :
258
R. Marty and K. Sølna " lim !;
./ D e.!/g .i z1 / g .i z2 / :
"!0
Moreover, observing that j" C !j12H maxf.! /12H ; .! C /12H g for every 2 Œ " ; " , we get ˇ ˇ ˇ ˇ " ˇ!; ./ˇ k k21 maxf.! /12H ; .! C /12H gjg .i z1 / g .i z2 / j: The function 7! g .i z1 / g .i z2 / is absolutely integrable, thus, by the bounded convergence theorem, Z
1
lim
"!0 1
" !;
./d
D e.!/
Z
1 1
g .i z1 / g .i z2 / d ; t u
which concludes the proof. Lemma 4. We have the following convergence Z lim
"!0 R
j
"2 C !"2 j2
j" C !j2H 1
g .i z1 / g .i z2 . C 2!" // d D 0
for every z1 and z2 2 R and ! 6D 0. Proof. Without loss of generality and for simplicity of presentation we assume that ! D 1. We have Z j "2 C "2 j2 (15) jg .i z1 / g .i z2 . C 2" //j d 2H 1 j" C 1j R Z jg .i z1 / g .i z2 . C 2" //j d k k21 j" C 1j2H 1 R Z j exp.i z/ 1jj exp.i z1 . C 2" // 1j D k k21 d ji zjj i z2 . C 2" /j j" C 1j2H 1 R 2 Z j exp.i z1 " / 1jj exp.i z2 . C 2/" / 1j k k1 D" d : jz1 z2 j R jjj C 2j j C 1j2H 1 We next define for a and b in Œ1; 1 the "-dependent integral Z
b
J " .a; b/ D " a
j exp.i z" / 1jj exp.i z. C 2/" / 1j jjj C 2j j C 1j2H 1
d :
Asymptotic Behavior of Oscillatory Fractional Processes
259
We fix a real number 0 < < < 1 sufficiently small and consider separately the integrals J " .1; 2 /, J " .2 ; 2 C /, J " .2 C ; /, J " . ; / and J " . ; 1/, which are defined such that Z
j exp.i z" / 1jj exp.i z. C 2/" / 1j
"
jjj C 2j j C 1j2H 1
R
d
(16)
D J " .1; 2 / C J " .2 ; 2 C / CJ " .2 C ; / C J " . ; / C J " . ; 1/: Using j exp.i u/ 1j 2 for every u 2 R, we get J " .1; 2 / C J " .2 C ; / C J " . ; 1/ (17) Z 2 Z
Z 1 d 4" C C : jjj C 2j j C 1j2H 1 1 2C
Using j exp.i u/ 1j min.juj; 2/ for every u 2 R, we get J " .2 ; 2 C / C J " . ; / Z 2C
Z
d d C 2jzj 2H 1 2H 1 2 jj j C 1j j C 2j j C 1j D O. / (independent of "/:
(18)
Then using (15), (16), (17) and (18) we get Z lim
"!0 R
j
"2 C "2 j2
j" C 1j2H 1
jg .i z/ g .i z . C 2" //j d D 0; t u
which concludes the proof.
4 Extension to More General Periodic Functions 4.1 Assumptions and Results In this section we generalize the above results and study the asymptotic behavior of the quantity defined for every z 0 by v"p .z/ D
1 "
Z
z
0
y y p dy "2 "
(19)
260
R. Marty and K. Sølna
where p is a zero-mean periodic function. More precisely we assume that there exists ! > 0 and two sequences fpc .k/; k 1g and fps .k/; k 1g such that for every z 0 p.z/ D
1 X
.pc .k/ cos.k!z/ C ps .k/ sin.k!z// :
kD1
We assume that the Fourier series of the right-hand side of the identity above is uniformly convergent and that the coefficients fpc .k/; k 1g and fps .k/; k 1g satisfy 1 X
k 1=2H .jpc .k/j C jps .k/j/ < 1:
(20)
kD1
Notice that Assumption 20 implies 1 X
k 12H pc .k/2 C ps .k/2 < 1:
(21)
kD1
Theorem 2. Under the assumptions above, as "!0 the finite dimensional distributions of v"p converge to those of the Brownian motion Bp satisfying
EŒBp .z1 /Bp .z2 / D
8 1 ˆ j .0/j2 X 12H ˆ ˆ minfz ; z g l pc .l/2 C ps .l/2 if < 2; ˆ 1 2 2H 1 < 2j!j l D1
1 ˆ j!j12H X j .l!/j2 ˆ ˆ pc .l/2 C ps .l/2 if D2: ; z g minfz ˆ 1 2 : 2H 1 2 l l D1
Next we consider another zero-mean periodic function pQ defined for every z by p.z/ Q D
1 X
.pQc .k/ cos.k !z/ Q C pQs .k/ sin.k !z// Q :
(22)
kD1
where the Fourier series of the right-hand side is uniformly convergent, !Q > 0 and the two sequences fpQc .k/; k 1g and fpQs .k/; k 1g satisfy 1 X
k 1=2H .jpQc .k/j C jpQs .k/j/ < 1:
kD1
From this function we define the process v"pQ for every z 0 by
(23)
Asymptotic Behavior of Oscillatory Fractional Processes
v"pQ .z/
1 D "
Z
z
0
y "2
pQ
261
y "
dy:
(24)
By applying Theorem 2 we get that the finite dimensional distributions of v"pQ converge to those of a Brownian motion. Now we address the question of the asymptotic correlation or decorrelation of the processes v"p and v"pQ . Theorem 3. Let z1 and z2 be two positive real numbers. Under the assumptions above, as "!0 the covariance EŒv"p .z1 /v"pQ .z2 / converges to C .z1 ; z2 / defined as 8 1 j .0/j2 X 12H ˆ ˆ minfz ˆ ; z g l .pc .nl/pQc .ml/Cps .nl/pQs .ml// 1 2 ˆ ˆ 2jn!j2H 1 ˆ ˆ l D 1 ˆ ˆ ˆ if !=! Q D n=m 2 Q with gcd.n; m/ D 1 and < 2; < 1 C .z1 ; z2 / D jn!j12H X j .nl!/j2 ˆ minfz ; z g .pc .nl/pQc .ml/Cps .nl/pQs .ml// ˆ 1 2 ˆ 2 l 2H 1 ˆ ˆ l0D1 ˆ ˆ ˆ ˆ if !=! Q D n=m 2 Q with gcd.n; m/ D 1 and D 2; ˆ : 0 if !=! Q … Q:
4.2 Proof of Theorems 2 and 3 First we establish a technical lemma regarding the quantities Ij" .!1 ; !2 ; z1 ; z2 / defined in Sect. 3. Lemma 5. For every z1 and z2 , there exists a constant CI .z1 ; z2 / > 0 such that for every j 2 f1; 2; 3; 4g, " 2 .0; 1/ and ! > 0 jIj" .!; !; z1 ; z2 /j .! 12H C ! 2H /CI .z1 ; z2 /: As a consequence, for every ! and z, there exists a constant CQ .!; z/ > 0 such that for every k and "
E vk!;" .z/2 C E vk!;" .z/2 CQ .!; z/k 12H c s Proof. Using the notation introduced in the proof of Lemma 3 and taking D !=2 we have (25) I1" .!; !; z1 ; z2 / ˇ ˇ Z ˇ 2 "2 C !"2 ˇ z1 z2 D g .i z1 / g .i z2 / d 4 R j" C !j2H 1 ! ! ! z1 z2 " ! D ; 1 : S 1; C S " ; C S " 4 2" 2" 2" 2"
262
R. Marty and K. Sølna
By making the substitution !=" and subsequently !! we obtain ˇ 2 Z !=2 ! ˇˇ d ˇ " k k1 ˇS 1; ˇ 4" 2" z1 z2 1 2 j C !j2H 1 Z k k21 1=2 d 4! 2H 2 z1 z2 1 j C 1j2H 1 We deal with S " .!=2" ; 1/ in a similar way. Regarding S " .!=2" ; !=2" / we have ˇ ! ! ˇ k k2 Z !=2" je i z1 1jje i z2 1j ˇ ˇ " 1 d : ˇS ; ˇ 2" 2" z1 z2 !=2" 2 j" C !j2H 1 Moreover, we have j" C !j12H maxf.!=2/12H ; .3!=2/12H g D ! 12H supf.1=2/12H ; .3=2/12H g for every 2 Œ!=2" ; !=2" , thus ˇ ! ! ˇ k k21 ˇ " ˇ ˇS ; ˇ ! 12H maxf.1=2/12H ; .3=2/12H g 2" 2" z1 z2 Z 1 i z1 je 1jje i z2 1j d : 2 1 Hence, the lemma is proved for I1" .!; !; z1 ; z2 /. We deal with I4" .!; !; z1 ; z2 / in the same way. Regarding I2" .!; !; z1 ; z2 / (and I3" .!; !; z1 ; z2 /) we observe that by Cauchy-Schwarz inequality we have jI2" .!; !; z1 ; z2 /j
q
I1" .!; !; z1 ; z1 /I4" .!; !; z2 ; z2 /:
Thus, using the bounds for I1" .!; !; z1 ; z2 / (and I4" .!; !; z1 ; z2 /) completes the proof of the first part of the lemma. To prove the second part, we let CQ .!; z/ D 8.! 12H C ! 2H /CI .z; z/: t u The following lemma deals with a uniform bound for the second moment of v"p .z/ for every z. Lemma 6. For every z and ! there exists a sequence fck .!; z/gk of positive numbers such that for every k and "
Asymptotic Behavior of Oscillatory Fractional Processes
263
r h r h i i k!;" 2 jpc .k/j E vc .z/ C jps .k/j E vk!;" .z/2 ck .!; z/ s and
1 X
ck .!; z/ < 1:
kD1
Proof. It is a direct consequence of (20) and Lemma 5.
t u
Proof. (Theorems 2 and 3) We have v"p .z/ D
1 "
Z
z
0
1 y X y y pc .k/ cos k! C ps .k/ sin k! dy: 2 " " " kD1
Because the convergence of the Fourier series is uniform we can change the order of the integral and the infinite sum to obtain v"p .z/
D
1 X
pc .k/vk!;" .z/ C ps .k/vk!;" .z/ : c s
(26)
kD1
Moreover, using Lemma 6 we get h i Q .z1 /vcj !;" .z2 /j jpc .k/pQc .j /jE jvk!;" c r h jpc .k/pQc .j /j
i h i j !;" Q E vk!;" .z1 /2 E vc .z2 /2 c
Q z2 / ck .!; z1 /cj .!;
(27)
where ck .!; z1 / and cj .!; Q z2 / are defined as in Lemma 6. Then we can write EŒv"p .z1 /v"pQ .z2 / D
1 X 1 X
h i Q pc .k/pQc .j /E vk!;" .z1 /vcj !;" .z2 / c
kD1 j D1
C
1 X 1 X
h i Q pc .k/pQs .j /E vk!;" .z1 /vsj !;" .z2 / c
kD1 j D1
C
1 X 1 X
h i j !;" Q ps .k/pQc .j /E vk!;" .z /v .z / 1 2 s c
kD1 j D1
C
1 X 1 X kD1 j D1
h i Q ps .k/pQs .j /E vk!;" .z1 /vjs !;" .z2 / : s
(28)
264
R. Marty and K. Sølna
We first deal with the convergence of the first term of the right hand side of the identity above. We know from Theorem 1 that for each .j; k/, h i C.j!/ j !;" Q .z /v .z / D lim E vk!;" inf.z1 ; z2 /Ifj!Dk !g 1 2 Q : c c "!0 2
(29)
Combining the bounded convergence theorem, (29), (27) and Lemma 6 then implies lim
"!0
1 1 X X
h i j !;" Q pc .k/pQc .j /E vk!;" .z /v .z / 1 2 c c
kD1 j D1 1
1
min.z1 ; z2 / X X D C.j!/2 Ifj!Dk !g Q pc .k/pQc .j /: 2 j D1 kD1
Proceeding in the same way for the other terms of the right hand side of (28) we get 1
lim EŒv"p .z1 /v"pQ .z2 / D
"!0
1
min.z1 ; z2 / X X C.j!/2 Ifj!Dk !g Q 2 j D1 kD1
.pc .k/pQc .j / C ps .k/pQs .j // :
(30)
If !=! Q 2 R Q then it is obvious that lim EŒv"p .z1 /v"pQ .z2 / D 0:
"!0
(31)
Let us now consider that !=! Q 2 Q. Then there exist two positive integers m and n such that gcd.m; n/ D 1 and !m Q D !n. If j! D k !Q then j D k n=m so that there exists an integer l such that k D ml and j D nl. We then conclude 1
lim EŒv"p .z1 /v"pQ .z2 / D
"!0
min.z1 ; z2 / X C.nl!/2 .pc .ml/pQc .nl/ C ps .ml/pQs .nl//: 2 lD1 (32) t u
5 Waves in Randomly Layered Media In this section we return to the discussion of applications to waves, the first motivation of our work. We consider wave propagation in randomly layered media. In the model that we consider the governing equations are the Euler equations giving conservation of moments and mass:
Asymptotic Behavior of Oscillatory Fractional Processes
265
@p " @u" .z; t/ C .z; t/ D 0 ; @t @z 1 @p " @u" .z; t/ C .z; t/ D 0 ; K " .z/ @t @z " .z/
where t is the time, z is the depth into the medium, p " is the pressure and u" the particle velocity. The medium parameters are the density " and the bulk-modulus K " (reciprocal of the compressibility). We assume that " is a constant identically equal to one in our non-dimensionalized units and 1=K " is modeled as 1 D " K .z/
(
z 1 C " 2 for z 2 Œ0; Z ; " 1 for z 2 R Œ0; Z ;
where > 0. We introduce the right- and left-going waves: A" D p " C u"
and
B " D u" p " ;
The boundary conditions are of the form A" .z D 0; t/ D f .t=" /
and
B " .z D Z; t/ D 0 ;
for a positive real number > 0 and a source function f . In order to deduce a description of the transmitted pulse, we open a window of size " in the neighborhood of the travel time of the homogenized medium and define the processes a" .z; s/ D A" .z; z C " s/
and
b " .z; s/ D B " .z; z C " s/ :
(33)
Observe that the background or homogenized medium in our scaling has a constant speed of sound equal to unity and that the medium is matched so that in the frame introduced in (33) the pulse shape is constant if 0 or if we consider the homogenized medium [2]. An important question is the study of the asymptotics of a" .Z; s/ when " goes to 0 [2]. In order to address this question we introduce next the Fourier transforms b a" and b b " of a" and b " respectively: Z b a" .z; !/ D that satisfy
e i !s a" .z; s/ds
and
b b " .z; !/ D
Z e i !s b " .z; s/ds ;
266
R. Marty and K. Sølna
z i! db a" " b ; a" e 2i !z=" b D 2 b dz 2" " z i! db b" " D 2 e 2i !z=" b a b b" ; dz 2" "
b.!/ ; b a " .0; !/ D f b b " .Z; !/ D 0:
Following [1, 2] we express the previous system of equations in term of the propagator P!" .z/ which can be written as P!" .z/
D
˛!" .z/ ˇ!" .z/ ˇ!" .z/ ˛!" .z/
! ;
and which satisfies z z 1 dP!" .z/ D H! ; 2 P!" .z/; dz " " "
! P!" .z D 0/ D
10 01
;
(34)
with i! 1 e 2i !z1 .z2 / 2i !z1 H! .z1 ; z2 / D e 1 2
! :
Defining next the transmission coefficient T!" and the reflection coefficient R!" by T!" .z/ D
1 ˇ!" .z/ " and R ; .z/ D ! ˛!" .z/ ˛!" .z/
(35)
we can then write 1 a .Z; s/ D 2
Z
"
b.!/ d! ; e i s! T!" .Z/f
(36)
b.!/ d! : e i s! R!" .Z/f
(37)
and 1 b .0; s/ D 2 "
Z
Hence we shall study the asymptotics of the propagator P!" in order to characterize a" as " goes to 0. What we wish to point out here is the dramatic influence of the range parameter H of . Let us recall two cases that have already been studied in previous works. For the moment we restrict ourself to the cases of slow phases 2 .0; 2/. • The case H D 1=2 or the mixing case. This corresponds to the framework studied in [1] and further developed in [2]. With 2 .0; 2/ we assume D 1 to get
Asymptotic Behavior of Oscillatory Fractional Processes
267
a non-trivial asymptotic behavior. We then have the convergence as " goes to 0 a" .Z; s/ ! e a.Z; s/ WD .G f /.s B/;
(38)
where G is a centered Gaussian density and B a Gaussian random variable. • The case H > 1=2 or the long-range case. Results on the propagation in a longrange medium have been derived in [9]. An important property of such a model is that there exists a constant cH > 0 such that EŒ.0/.z/ cH z2H 2 : In this case, we can prove that a.Z; s/ WD f .s B/ ; a" .Z; s/ ! e
(39)
where B a Gaussian random variable. In these two cases one of the most crucial step in the proof is the convergence of the propagator P!" . This consists of studying the asymptotics of the random differential equation (34). To do so, different tools are used depending on whether H D 1=2 or H > 1=2. If H D 1=2 then the diffusion-approximation and martingales techniques have been used [1,2]. If H > 1=2, because of the long-range property, the martingale approach cannot be used. In [9] the theory of rough paths [6, 7] that we describe below was used. A motivation for this work is to find a general framework dealing with the full range of H 2 .0; 1/ and in particular to address the case H 2 .0; 1=2/. Based on Theorem 1 we next present our conjecture regarding H 2 .0; 1=2/. We consider defined as throughout this paper, but to fix the ideas we may assume that is defined by .z/ D BH .z C 1/ BH .z/ where H 2 .0; 1=2/. Now we assume that D D 1 C .2 /.1 2H /=2 and < 1 so that 2 .0; 2/. Note that the case D 0 corresponds to the so called strong fluctuations regime, with medium fluctuations being of order one, see [2]. Recall that the propagator P!" satisfies (34) that we can write in the form i! " 2!;" F1 P!" ı d v0;" .z/ C F3 P!" ı d v2!;" .z/ ; (40) c .z/ C F2 P! ı d vc s 2 ! ! ! 1 0 0 1 0i where F1 D ; F2 D and F3 D and 0 1 1 0 i 0 dP!" .z/ D
v0;" c .z/ D
1 "
Z
z
0
y "2
dy D
" .12H /=2 "22H
Z
z
0
y "2
dy:
(41)
268
R. Marty and K. Sølna
Because of H < 1=2, (41) and the invariance principle that establishes that lim
"!0
Z
1 "22H
z
y
0
"2
Ddistribution BH ;
dy z
we can then observe the convergence distribution lim v0;" 0: c D
"!0
(42)
Let ˝ D .!1 ; ; !n / be a collection of frequencies. Because of (42) and Theorem 1, as " goes to 0, we may expect that the propagator vector P˝" D .P!"1 ; ; P!"n / converges to P˝ D .P!1 ; ; P!n / where the asymptotic propagator P! is solution of dP! .z/ D
iC.!/! F2 P! ı dBc! .z/ C F3 P! ı dBs! .z/ 2
where C.!/2 is defined by (10). Then we follow the same procedure as in [2] pp. 180–184 or as in [4] using complex martingales to get the limit distribution written as Z 1 ! 2 C.!/2 1 b.!/d!: exp i !s Z f e a.Z; s/ Ddistribution 2 1 4 Hence, we get the following conjecture. Conjecture 1. Under the above assumptions, as " goes to 0, fa" .Z; s/gs converges to the random process fe a.Z; s/gs that can be written as e a.Z; s/ D .GZ f / .s/ ;
(43)
where GZ is such that its Fourier transform is 2 cZ .!/ D exp j .0/j j!j32H Z : G 4 Note that this means that we in the antipersistent case have a modification in the pulse shape only, while we in the long range case have a correction in the travel time only and in the mixing case a modification in both. The technical challenge in the full proof would be the rigorous convergence of P˝" to P˝ as "!0. This could be based on the use of the theory of rough paths initiated by T. Lyons [7] and for which a good reference is [6]. To summarize, the main theorem establishes that the solution x of the differential system dx D F .x/d w driven by a continuous and multidimensional noise w D .w1 ; ::; wn / is a continuous function of the noise w for a topology based on the q-variations of iterated integrals of w for q large enough.
Asymptotic Behavior of Oscillatory Fractional Processes
269
This would imply that the convergence of P˝" could be reduced to the convergence of the noises of the form V ˝;" and their iterated integrals.
References 1. J.F. Clouet, J.P. Fouque, Spreading of a pulse travelling in random media. Ann. Appl. Probab. 4, 1083–1097 (1994) 2. J.P. Fouque, J. Garnier, G. Papanicolaou, K. Solna, Wave Propagation and Time Reversal in Randomly Layered Media (Springer, New York, 2007) 3. J. Garnier, A multi-scaled diffusion-approximation theorem. Applications to wave propagation in random media. ESAIM Probab. Statist. 1, 183–206 (1997) 4. J. Garnier, G. Papanicolaou, Analysis of pulse propagation through an one-dimensional random medium using complex martingales. Stoch. Dyn. 8, 127–138 (2008) 5. J. Garnier, K. Solna, Pulse propagation in random media with long-range correlation. SIAM Multiscale Model. Simul. 7, 1302–1324 (2009) 6. A. Lejay, An introduction to rough paths, Séminaire de Probabilités XXXVII, Lecture Notes in Mathematics (Springer, New York, 2003) 7. T. Lyons, Differential equations driven by rough signals. Rev. Mat. Iberoamer. 14(2), 215–310 (1998) 8. R. Marty, Asymptotic behavior of differential equations driven by periodic and random processes with slowly decaying correlations. ESAIM: Probab. Statist. 9, 165–184 (2005) 9. R. Marty, K. Solna, Acoustic waves in long range random media. SIAM J. Appl. Math. 69, 1065–1083 (2009) 10. R. Marty, K. Solna, A general framework for waves in random media with long-range correlations. Ann. Appl. Probab. 21(1), 115–139 (2011) 11. G. Samorodnitsky, M.S. Taqqu, Stable Non-Gaussian Random Processes (Chapman and Hall, New York, 1994)
Time Inversion Property for Rotation Invariant Self-similar Diffusion Processes Juha Vuolle-Apiala
Abstract We will show, using the skew product representation and the corresponding result for the radial process, that Shiga–Watanabe’s time inversion property of index ˛, ˛ positive, holds for all ˛-self-similar, rotation invariant diffusion processes on Rd , d 2, starting at 0. Keywords Time inversion • Self-similar • Bessel process • Diffusion • Rotation invariant • Skew product • Radial process AMS Classification: 60G18, 60J60, 60J25
1 Introduction The following property is well known for Brownian motion in Rd , d 1, starting at 0: .Xt / and .tX1=t / are equivalent diffusions under P0 : (1) Shiga and Watanabe [8] and Watanabe [13] showed that (1) is also true for all Bessel diffusions on non-negative real numbers. Bessel diffusions (including Brownian motion) form exactly the class of 1/2-selfsimilar diffusions on the non-negative real axis; see Lamperti [5]. For ˛-self-similar diffusions, ˛ strictly positive, (1) can immediately be generalized to (2): .Xt / and .t 2˛ X1=t / are equivalent diffusions under P0 :
(2)
It was in Vuolle-Apiala [11] showed that (2) is valid for all rotation invariant (RI) ˛-self-similar diffusions .Xt / on Rd , d 2; for which 0 is a polar set, that is, J. Vuolle-Apiala () Mathematical Department, Åbo Akademi University, FIN-20500 Åbo, Turku, Finland e-mail: [email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__13, © Springer-Verlag Berlin Heidelberg 2012
271
272
J. Vuolle-Apiala
after starting at 0 Xt will never return to it. The time inversion problem on R was studied in Graversen, Vuolle-Apiala [3] and on finite number of rays meeting at 0 in Vuolle-Apiala [10]. We call (2) the time inversion property of index ˛, (1) is a special case for ˛ D 1=2. In this note we will show that (2) is valid for all rotation invariant (RI) ˛-selfsimilar diffusions even if 0 is not a polar set. We have a proof which is valid, no matter if 0 is polar or non-polar. Our proof is based on: (1) The fact that (2) is true for the radial process. (2) The skew product representation of (Xt ), which is formulated at the end of this section (formula (6)). By ˛-self-similarity property, ˛ > 0, we mean that .Xt ; Px / has the same finite dimensional distributions as .a˛ Xat ; Pa x / for all x 2 Rd , for all positive a: ˛
(3)
By (RI) we mean .Xt ; Px / has the same finite dimensional distributions as .T 1 .Xt /; PT .x/ / for all rotations T 2 O.d /:
(4)
Obviously, (4) (and (3), of course) is fulfilled for Brownian motion on Rd , d 2. Remark 1. A related problem for (Xt ) under Px for arbitrary x has been studied in two papers: Gallardo and Yor [1] and Lawi [6]. In both articles the following property was called a time inversion property (of index 2˛ according to the definition by S. Lawi): (?) Assume (Xt ) under Px is a time-homogeneous Markov process, for all x 2 Rd . Then .t 2˛ X1=t / is a time-homogeneous Markov process under Px . For x D 0 this implies our definition (2). Lawi also obtained a converse property: (?) implies that (Xt ) is ˛-self-similar or h-transform of an ˛-self-similar Markov process (see Corollary 2.5 in Lawi [6]). In Vuolle-Apiala [11], where f0g was assumed to be a polar set, our main tool was the skew product representation for (Xt ; P0 ): .Xt / under P0 has the same finite dimensional distributions as ŒjXt j; R t jXh j1=˛ dh under P0 Q;
(5)
1
for some > 0, where (s ; Q/ is a stationary spherical Brownian motion, independent of the radial process (jXt j; P0 ), defined for 1 < s < C1 such that the law of (0 ) is the uniform spherical distribution m.d/. Equation (5) is a consequence of the result of Ito and McKean Jr. [4], p.275.
Time Inversion Property for Rotation Invariant Self-similar Diffusion Processes
273
The radial process (jXt j; Pjxj ), x 2 Rd , is an ˛-self-similar diffusion on [0, 1) and the time inversion property (2) for .jXt j; P0 / is therefore fulfilled. If f0g is a polar set then the skew product (5) at 0 completely characterizes .Xt ; P0 /. If 0 is a regular point for f0g, then the skew product representation (5) at 0 is not available. However, when Xt is started outside 0, then we have the following skew product representation: .Xt / under Px has the same finite dimensional distributions as ŒjXt j; R t jXh j-1/˛ dh under Pjxj Q ; t < T0 ; 0
(6)
for x ¤ 0 , for some > 0, where .s ; Q / is a spherical Brownian motion on S d 1 , starting at D arg.x/, s > 0, independent of .jXt j/ (see Graversen, Vuolle-Apiala [2] and Vuolle-Apiala [9]), T0 D infft > 0jXt D 0g. We apply (6) and the fact that the radial process fulfills (2) to construct a proof for (2) which is valid both in the case of f0g as a non-polar set and in the case of f0g as a polar set for .X /. Remark 2. All the ˛-self-similar (RI) diffusions on Rd , d 2, can be characterized as follows: As shown in Graversen, Vuolle-Apiala [2] and Vuolle-Apiala [9], every ˛-selfsimilar (RI) diffusion on Rd n f0g, d 2, killed at T0 if T0 is finite, can be written as a skew product (6). On the other hand, every skew product of type (6) gives an ˛-self-similar (RI) diffusion on Rd n f0g, the radial process can be chosen to be any ˛-self-similar diffusion on (0, 1) which is independent of the spherical process (see also Vuolle-Apiala [9], Remark on p. 965 and Lemma 2.1). As shown in VuolleApiala [9], Sect. 2, 0 can be added to the state space such that the extended process .Zt / is an ˛-self-similar (RI) diffusion on Rd which on Rd n f0g behaves like (6). f0g is polar for .Zt ; Px /, x 2 Rd , if and only if 0 is an entrance, non-exit boundary point of Œ0; 1/ for .jZt j/.
2 The Main Result Theorem 1. Let .Xt ; P0 / be an ˛-self-similar (RI) diffusion on Rd , d 2, ˛ > 0, starting at 0. Then the time inversion property (2) is valid. The following lemma plays a central role in the proof of Theorem 1: Lemma 1. Let .Xt ; Px /, x 2 Rd , be an ˛-self-similar (RI) diffusion on Rd , d 2, ˛ > 0. Then (i) the radial process .jXt j; Pjxj / is an ˛-self-similar diffusion on Œ0; 1/. Consequently, the process .jXt j; P0 / fulfills the time inversion property (2). (ii) For x ¤ 0, .Xt ; Px /, t < T0 , has the skew product representation (6).
274
J. Vuolle-Apiala
Proof. The time inversion property for the radial process in (i) was shown in Shiga and Watanabe [8] for ˛ D 12 , the generalization for ˛ > 0 is obvious. For the Proof of (ii) see Graversen and Vuolle-Apiala [2], Theorem 2.2, and Vuolle-Apiala [9], Lemma 2.1. t u Proof (Proof of Theorem 1). Assume for simplicity that ˛ D 1=2, the general case can be shown similarly. Since the two processes .Xt / and .tX1=t / have continuous paths on .0; 1), standard arguments of general theory of processes, based on the monotone classes theorem (See for example [7]), allow us to only check that these two processes have same finite dimensional distributions. We will prove that .Xt / and .tX1=t / have the same n-dimensional distributions for any positive integer n: Assume n D 2, the proof in the general case is analogous. Let Ii , i D 1; 2, be a Borel subset of .0; 1/ and Ji , i D 1; 2, a Borel subset of S d 1 and let Xt D Œrt ; t , rt and t are the radial and the spherical parts of Xt , respectively. Then for s < t we have P0 frs 2 I1 ; rt 2 I2 ; s 2 J1 ; t 2 J2 g D A C B; where A D P0 frs 2 I1 ; rt 2 I2 ; s 2 J1 ; t 2 J2 I s and t belong to different excursions of .ru / away from 0g D P0 frs 2 I1 ; rt 2 I2 ; s 2 J1 ; t 2 J2 I ru D 0 for some u 2 .s; t/g and B D P0 frs 2 I1 ; rt 2 I2 ; s 2 J1 ; t 2 J2 I s and t belong to the same excursion of .ru / away from 0g D P0 frs 2 I1 ; rs 2 I2 ; s 2 J1 ; t 2 J2 I ru ¤ 0 for all u 2 .s; t/g: Remark 3. (i) As shown by J. Lamperti [5], p.210, Lemma 2.5, either T0 < 1 a.s. or T0 D 1 a.s.. (Lamperti actually showed that T0 < 1 a.s. or T0 D 1 a.s. for .Xt ; Px /, x ¤ 0. However, if Xt starts at 0 then it immediately leaves 0 and enters Rd n f0g). (ii) If f0g is polar then obviously A D 0. Now A is equal to P0 frs 2 I1 ; rt 2 I2 ; s 2 J1 ; t 2 J2 jru D 0 for some u 2 .s; t/g P0 fru D 0 for some u 2 .s; t/g: Because of (RI) and independence of the excursions of .ru ; u /, away from 0 we have
Time Inversion Property for Rotation Invariant Self-similar Diffusion Processes
275
P0 frs 2 I1 ; rt 2 I2 ; s 2 J1 ; t 2 J2 jru D 0 for some u 2 .s; t/g D P0 frs 2 I1 ; s 2 J1 jru D 0 for some u 2 .s; t/g P0 frt 2 I2 ; t 2 J2 jru D 0 for some u 2 .s; t/g D m.J1 /P0 frs 2 I1 jru D 0 for some u 2 .s; t/g m.J2 /P0 frt 2 I2 jru D 0 for some u 2 .s; t/g D m.J1 /m.J2 /P0 frs 2 I1 ; rt 2 I2 jru D 0 for some u 2 .s; t/g
(7)
because of independence of the excursions of (ru /, m is the uniform spherical distribution on S d 1 . Thus A D m.J1 /m.J2 /P0 frs 2 I1 ; rt 2 I2 I ru D 0 for some u 2 .s; t/g: Similarly as in Vuolle-Apiala [10], p.257, using the fact that (1) is valid for the radial process .rt /, we can now conclude that A D m.J1 /m.J2 /P0 fsr1=s 2 I1 ; tr1=t 2 I2 I ur1=u D 0 for some u 2 .s; t/g D m.J1 /m.J2 /P0 fsr1=s 2 I1 ; t r1=t 2 I2 I ru D 0 for some u 2 .1=t; 1=s/g: Equation (7) implies, replacing s; t by 1=s and 1=t, (now 1=t < 1=s) and I1 , I2 by I1 =s, I2 =t, that P0 fr1=s 2 I1 =s; r1=t 2 I2 =t; 1=s 2 J1 ; 1=t 2 J2 jru D 0 for some u 2 .1=t; 1=s/g D m.J1 /m.J2 /P0 fsr1=s 2 I1 ; t r1=t 2 I2 jru D 0 for some u 2 .1=t; 1=s/g: which further implies A D P0 fr1=s 2 I1 =s; r1=t 2 I2 =t; 1=s 2 J1 ; 1=t 2 J2 jru D 0 for some u 2 .1=t; 1=s//g P0 fru D 0 for some u 2 .1=t; 1=s/g ˚ D P0 s r1=s 2 I1 ; t r1=t 2 I2 ; 1=s 2 J1 ; 1=t 2 J2 I 1=s and 1=t belong to different excursions of .ru / away from 0g :
Now B D P0 frs 2 I1 ; rt 2 I2 ; s 2 J1 ; t 2 J2 I ru ¤ 0 for all u 2 .s; t/g D E0 fP.rs ,s frt s 2 I2 ; t s 2 J2 ; t s < T0 gI rs 2 I1 ; s 2 J1 g
276
J. Vuolle-Apiala
because of the Markov property. Using the skew product representation (6) for ˛ D 12 we obtain B D E0 fP.rs ;s / frt s 2 I2 ; R t s r 2 dh 2 J2 gI rs 2 I1 ; s 2 J1 g: 0
h
This can be rewritten “ BD P.r;/ frt s 2 I2 ; R t s r 2 dh 2 J2 gP0 frs 2 dr; s 2 dg: 0
I1 xJ1
h
(RI) implies “ BD I1 x J1
P.r;/ frt s 2 I2 ; R t s r 2 dh 2 J2 gP0 frs 2 drgm.d/; 0
h
where m is the uniform spherical distribution on S d 1 . This is further equal to # Z "“ Z 1
t s
P.r;/ frt s 2 I2 ; v 2 J2 ; 0
0
I1 xJ1
rh2 dh 2 d vgP0 frs 2 drgm.d/ :
The independence of .ru / and .v / implies Z
“ P0 frs 2 drgm.d/
BD
1 0
I1 xJ1
Z Q fv 2 J2 gPr frt s 2 I2 ;
t s
0
rh2 dh 2 d vg :
which is equal to Z
Z J1
Z
1
Q fv 2 J2 g
m.d/
Z
I1
0
t s
P0 frs 2 drgPr frt s 2 I2 ; 0
rh2 dh 2 d vg :
Markov property of the radial process .ru / gives Z
Z BD J1
Z
1
Q fv 2 J2 gP0 frs 2 I1 ;
m.d/ 0
t
s
rh2 dh 2 d v; rt 2 I2 g :
The time-inversion property (1) is valid for .ru / and thus Z
Z BD
1
m.d/ 0
J1
"Z
Z D
m.d/ J1
0
1
Z Q fv 2 J2 gP0 fsr1=s 2 I1 ;
s
Q fv 2 J2 gP0 fr1=s 2 s 1 I1 ;
t
.hr1= h /2 dh 2 d v; t r1=t 2 I2 g ;
Z
1=s 1=t
# rh2 dh 2 d v; r1=t 2 t 1 I2 g :
Because .v / has a symmetric density with respect to m.d/ (Vuolle-Apiala, Graversen [12]), we have
Time Inversion Property for Rotation Invariant Self-similar Diffusion Processes
"Z
Z BD
m.d/ J2
0
1
Q fv 2 J1 gP fr1=s 2 s
0
1
Z I1 ;
1=s
1=t
277
# rh2 dh
2 d v; r1=t 2 t
1
I2 g :
Reversing the procedure above we get this equivalent to B D P0 fr1=t 2 t 1 I2 ; r1=s 2 s 1 I1 ; 1=t 2 J2 ; 1=s 2 J1 I ru ¤ 0 for all u 2 .1=t; 1=s/g;
which is further equivalent to ˚ B D P0 sr1=s 2 I1 ; t r1=t 2 I2 ; 1=s 2 J1 ; 1=t 2 J2 ; 1=s and 1=t belong to the same excursion of .ru / away from 0g :
t u
Acknowledgements The author wants to thank an anonymous referee for useful comments and suggestions which considerably improved this paper.
References 1. L. Gallardo, M. Yor, Some new examples of Markov processes which enjoy the time-inversion property. Probab. Theor. Relat. Field. 132(1), 150–162 (2005) 2. S.E. Graversen, J. Vuolle-Apiala, ˛-self-similar Markov processes. Probab. Theor. Relat. Field. 71(1), 149–158 (1986) 3. S.E. Graversen, J. Vuolle-Apiala, On Paul Levy’s arc sine law and Shiga-Watanabe’s time inversion result. Probab. Math. Statist. 20(1), 63–73 (2000) 4. K. Ito, H.P. McKean Jr., Diffusion Processes and Their Sample Paths (Springer, Berlin, 1974) 5. J.W. Lamperti, Semi-stable Markov processes I. Z. Wahrschein. verw. Gebiete 22, 205–225 (1972) 6. S. Lawi, Towards a characterization of Markov processes enjoying the time-inversion property. J. Theoret. Probab. 21(1), 144–168 (2008) 7. D. Revuz, M. Yor, Continuous martingales and Brownian motion (Springer, Berlin, 1991) 8. T. Shiga, S. Watanabe, Bessel diffusions as one parameter family of diffusion processes. Z. Wahrschein. verw. Gebiete 27, 37–46 (1973) 9. J. Vuolle-Apiala, On certain extensions of a rotation invariant Markov process. J. Theoret. Probab. 16(4), 957–969 (2003) 10. J. Vuolle-Apiala, Time inversion result for self-similar diffusions on finite number of rays, in Contributions to Management Science, Mathematics and Modelling, Acta Wasaensia 122, 253–260 (University of Vaasa, 2004) 11. J. Vuolle-Apiala, Shiga-Watanabe’s time inversion property for self-similar diffusion processes. Stat. Probab. Lett. 77(9), 920–924 (2007) 12. J. Vuolle-Apiala, S.E. Graversen, Duality theory for self-similar processes. Ann. Inst. Henri Poincaré 22, 323–332 (1986) 13. S. Watanabe, On time inversion of one-dimensional diffusion process. Z. Wahrschein. verw. Gebiete 31, 115–124 (1975)
On Peacocks: A General Introduction to Two Articles Antoine-Marie Bogso, Christophe Profeta, and Bernard Roynette
The aim of the following two articles, namely: 1. Some examples of peacocks in a Markovian set-up 2. Peacocks obtained by renormalisation, strong and very strong peacocks is to provide examples of integrable processes .Xt ; t 0/ which increase in the convex order, i.e., such that: i. For every t 0, EŒjXt j < 1; ii. For every convex function W R ! R and every 0 s < t, 1 < EŒ .Xs / EŒ .Xt /: Such processes are called peacocks (see [1]). In these two papers, the notions of conditional monotonicity and that of logconcave increments play an important role. In the first one, the emphasis is made on processes constructed from Markov processes (Sects. 2–4). We also establish a link between stochastic and convex orders (Sect. 5). The bulk of this article is taken up in Chaps. 1 and 8 of [1]. The second paper is devoted to processes obtained, from other processes, either by centering or normalisation (Sects. 1–3). The notions of strong and very strong peacock are developed in Sects. 4 and 5. This second article is a sort of continuation of the first paper. The reader interested in questions developed in these two articles may refer to [1] for more informations and results.
A.-M. Bogso () C. Profeta B. Roynette Institut Elie Cartan, Université Henri Poincaré, B.P. 239, 54506 Vandœuvre-lès-Nancy Cedex, France e-mail: [email protected]; [email protected]; [email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__14, © Springer-Verlag Berlin Heidelberg 2012
279
280
A.-M. Bogso et al.
Finally, we draw to the reader’s attention that each of these two papers may be read independently.
Reference 1. F. Hirsch, C. Profeta, B. Roynette, M. Yor, Peacocks and Associated Martingales, vol 3 (Bocconi-Springer, New York, 2011)
Some Examples of Peacocks in a Markovian Set-Up Antoine-Marie Bogso, Christophe Profeta, and Bernard Roynette
Abstract We give, in a Markovian set-up, some examples of processes which are increasing in the convex order (we call them peacocks). We then establish some relation between the stochastic and convex orders. Keywords Processes increasing in the convex order • Peacocks • Conditionally monotone processes • Stochastic order • Markov process
AMS Classification: 60J25, 32F17, 60G44, 60E15
1 Introduction 1.1 Definitions We start with a few definitions: (a) A real-valued process .Xt ; t 0/ is said to be increasing in the convex order if: 8t > 0; and, for every convex function
E ŒjXt j < 1
W R ! R:
t 2 RC 7! E Œ .Xt / 2 1; C1
is increasing.
(1)
A.-M. Bogso () C. Profeta B. Roynette Institut Elie Cartan, Université Henri Poincaré, B.P. 239, 54506 Vandœuvre-lès-Nancy Cedex, France e-mail: [email protected]; [email protected]; [email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__15, © Springer-Verlag Berlin Heidelberg 2012
281
282
A.-M. Bogso et al.
This notion plays an important role in many applied domains of probability; see, e.g., Shaked–Shanthikumar [20, 21]. We call such a process a peacock, an acronym derived from the French term: Processus Croissant pour l’Ordre Convexe. To prove (1), it suffices (see [10, Chap. 1]) to consider only the class: C WD f Note that if
is a convex function of C 2 class such that 2 C, then
0
00
has compact supportg:
is a bounded function.
(b) A real-valued process .Xt ; t 0/ is called a 1-martingale if there exists a martingale .Mt ; t 0/ (defined on a suitable filtered probability space) which has the same one-dimensional marginals as .Xt ; t 0/, that is to say, for each fixed t 0: (law) Xt D Mt : We say that such a martingale is associated to the process .Xt ; t 0/. From Jensen’s inequality, it is clear that a 1-martingale is a peacock. Conversely, a remarkable result due to Kellerer [13] states that any peacock is a 1-martingale. However, the proofs presented in Kellerer’s paper are not constructive, and in general, it is a difficult task to exhibit such a martingale. In this paper, we shall only tackle the question of exhibiting peacocks, and mainly focus on examples derived from diffusions.
1.2 Some Examples Let .Bs ; s 0/ be a standard Brownian motion. Carr et al. [3] proved that the process: Z Z 1 1 t s st ds D ds ; t 0 (2) exp Bs exp Bt s At WD t 0 2 2 0 is a peacock. Baker and Yor [2] then exhibited a martingale which is associated to this peacock, and is constructed from the Wiener sheet. This example was the starting point of many recent developments which we try to synthesize; consider, for every 0, a real-valued measurable process Z;˘ WD .Z;t ; t 0/ such that 8 2 RC ; 8t 2 RC ;
E e Z;t < 1;
and define, for any finite and positive measure on RC the process:
Some Examples of Peacocks in a Markovian Set-Up
Z ./ A
C1
WD 0
! e Z;t .dt/; 0 : E e Z;t
283
(3)
(Taking Z;t D Bt and .ds/ D 1Œ0;1 .ds/, we recover (2).) Now, this raises the following natural question: ./ Under which conditions is the process A ; 0 a peacock? ./
It is known that .A ; 0/ is a peacock in the following cases: • Z;t D tX with X a r.v., see [10]. • Z;t D Lt with .Lt ; t 0/ a Lévy process such that E e L1 < 1, (see [8]). • Z;t D G;t with, for every 0, .G;t ; t 0/ a Gaussian process such that the function ! E ŒG;t G;s is increasing for every s; t 0, (see [7]). In this paper, we shall exhibit several other families of peacocks. In Sect. 2, we introduce the notion of conditional monotonicity which will lead to a new large class of peacocks. In Sect. 3, we give many examples, among which the processes with independent log-concave increments and the “well-reversible” diffusions at fixed times. In Sect. 4, we present another condition, this time relying upon Laplace transforms, which implies the peacock property. Finally, in Sect. 5, we present a result which links the stochastic and convex orders, and makes it possible to recover some of the peacocks presented above.
2 A Class of Peacocks Under the Conditional Monotonicity Hypothesis In this section, we introduce and study the notion of conditional monotonicity, which already appear in [20, Chap. 4.B, pp. 114–126]. Definition 1 (Conditional monotonicity). A process .X ; 0/ is said to be conditionally monotone if, for every n 2 N , every i 2 f1; : : : ; ng, every 0 1 < < n and every bounded Borel function W Rn ! R which increases (resp. decreases) with respect to each of its arguments, we have: EŒ.X1 ; X2 ; : : : ; Xn /jXi D i .Xi /; where i W R ! R is a bounded increasing (resp. decreasing) function.
(4)
284
A.-M. Bogso et al.
Remark 1. 1. If there is an interval I of R such that, for every 0, X 2 I , we may assume in Definition 1 that is merely defined on I n , and i is defined on I . 2. Note that .X ; 0/ is conditionally monotone if and only if .X ; 0/ is conditionally monotone. 3. Let W R ! R be a strictly monotone and continuous function. It is not difficult to see that if the process .X ; 0/ is conditionally monotone, then so is ..X /; 0/. To prove that a process is conditionally monotone, we can restrict ourselves to bounded Borel functions increasing with respect to each of their arguments. Indeed, replacing by , the result then holds also for bounded Borel functions decreasing with respect to each of their arguments. Definition 2. We denote by En the set of bounded Borel functions W Rn ! R which are increasing with respect to each of their arguments. Theorem 1. Let .X ; 0/ a real-valued process which is right-continuous, conditionally monotone and which satisfies the following integrability conditions: For every compact K RC and every t 0: K;t WD sup exp.tX / D exp t sup X is integrable; 2K
(5)
2K
and kK;t WD inf E Œexp.tX / > 0:
(6)
2K
We set h .t/ D log EŒexp.tX /. Then, for every finite positive measure on RC : Z ./ At WD
1
e tX h .t / .d/ ; t 0
0
is a peacock. Proof (of Theorem 1). 1. By (5), for every 0 and every t 0, E Œexp.tX / < 1. This easily implies, thanks to the dominated convergence theorem, that h is continuous on RC , differentiable on 0; C1Œ, and h0 .t/e h .t / D E X e tX :
(7)
Since E e tX h .t / D 1, we obtain from (7): E .X h0 .t//e tX h .t / D 0:
(8)
Some Examples of Peacocks in a Markovian Set-Up
285
Moreover, we also deduce from (5) that, for every t 0, the function 0 7! h .t/ is right-continuous. 2. We first consider the case n X D ai ıi (9) i D1
where n 2 N , a1 0; : : : ; an 0, and 0 1 < : : : < n . Let 2 C. For t > 0, we have: # " n X @ h ./ i ./ 0 0 E At ai Xi hi .t/ exp .tXi hi .t// DE At @t i D1 Setting for i 2 f1; : : : ; ng, i D E
h
0
i ./ At Xi h0i .t/ exp .tXi hi .t//
we shall show that i 0 for every i 2 f1; : : : ; ng. Note that the function 0 .x1 ; : : : ; xn / 7!
0
@
n X
1 aj exp txj hj .t/ A
j D1
is bounded and increases with respect to each of its arguments, i.e., belongs to En . Hence, from the conditional monotonicity property of .X ; 0/: h h ii ./ i D E E 0 At .Xi h0i .t//e tXi hi .t / jXi D E .Xi h0i .t//e tXi hi .t / i .Xi / where i is a bounded increasing function. Besides, we have, .Xi h0i .t// i .Xi / i h0i .t/ 0: Therefore, i i h0i .t/ E .Xi h0i .t//e tXi hi .t / D 0 from (8): 3. We now assume that has compact support contained in a compact interval K. Since the function 7! exp .tX h .t// is right-continuous and bounded 1 from above by kK;t K;t which is finite a.s., there exists a sequence R R .n ; n 0/ of measures of the form (9), with supp .n / K, n .d/ D .d/ and for . / ./ every t 0, lim At n D At a.s. Moreover, from (5) and (6): n!C1
286
A.-M. Bogso et al.
. / jAt n j
K;t kK;t
Z .d/;
and from Point (2), for 0 s t: .n /
EŒ .As.n / / EŒ .At
/:
Therefore, since is sublinear, we can apply the dominated convergence theorem and pass to the limit when n ! C1 in this last inequality to obtain ./ that .At ; t 0/ is a peacock. .n / 4. In the general case, we set n .d/ D 1Œ0;n ./.d/ and observe is an Z that A x
increasing sequence of processes. Let be defined by .x/ D
.xz/
00
.z/d z.
0
An integration by parts yields, for 0 s t: i h . / . / E As n E At n i h i h . / . / E As.n / D E 0 .0/ At n As.n / C E At n i h . / E As.n / 0 D E At n i h i h . / . / from Point (3), and since E At n D E As n . Now, since is an increasing function on RC , the result follows from the monotone convergence theorem. t u Remark 2. Let W R ! R be a strictly monotone and continuous function, and denote a finite positive measure. From Remark 1, under the assumption that ..X /; 0/ still satisfies conditions (5) and (6), we obtain, denoting h; .t/ D log E Œexp .t.X //, that the process Z .;/ At WD
1
e t .X/h; .t / .d/ ; t 0
0
is a peacock. Note that only needs to be continuous and strictly monotone on an interval containing the image of X for every 0. Of course, Theorem 1 may have some practical interest only if we are able to exhibit enough examples of processes which enjoy the conditional monotonicity (4) property. Below, we shall see that there exists a large class of diffusions which enjoy this property. But to start with, let us first give a few examples which consist of processes with independent increments and Lévy processes in particular.
Some Examples of Peacocks in a Markovian Set-Up
287
3 Examples of Processes Satisfying the Conditional Monotonicity Property 3.1 Processes with Independent Increments Satisfying the Conditional Monotonicity Property We start by giving an assertion equivalent to (4) when dealing with processes with independent (not necessarily time-homogeneous) increments. Proposition 1. Let .X ; 0/ be a process with independent increments. Then, the conditional monotonicity hypothesis (4) is equivalent to the following: For every n 2 N , every 0 1 < < n and every function W Rn ! R in En , we have: E Œ .X1 ; : : : ; Xn / jXn D n .Xn / (10) where n is an increasing bounded function. Proof (of Proposition 1). The proof is straightforward. Indeed, let 2 En . For i 2 f1; : : : ; ng, the hypothesis of independent increments implies: E Œ .X1 ; : : : ; Xn / jXi D E ŒE Œ .X1 ; : : : ; Xn / jFi jXi D E E X1 ; : : : ; ; Xi ; Xi C1 Xi C Xi ; : : : ; Xn Xi C Xi jFi jXi DE e .X1 ; : : : ; Xi / jXi where e .x1 ; : : : ; xi / D E x1 ; : : : ; xi ; Xi C1 Xi C xi ; ; : : : ; Xn Xi C xi t u
belongs to Ei . 3.1.1 The Gamma Subordinator Is Conditionally Monotone The Gamma subordinator . ; 0/ is characterized by: E e t D
Z 1 x 1 tx e dx : D exp .1 e / .1 C t/ x 0
In particular, is a gamma random variable with parameter . From (10), we wish to show that for every n 2 N , every 0 1 < < n and every function W Rn ! R in En : EŒ. 1 ; : : : ; n /j n D n . n /;
(11)
288
A.-M. Bogso et al.
where n is an increasing function. The explicit knowledge of the law of and the fact that . ; 0/ has time-homogeneous independent increments imply the well-known result that, given f n D xg, the vector . 1 ; 2 1 ; : : : ; n n1 / follows the Dirichlet law with parameters .1 ; 2 1 ; : : : ; n n1 / on Œ0; x. In other words, the density fnx of . 1 ; 2 ; : : : ; n1 / conditionally on f n D xg equals: fnx .x1 ; : : : ; xn1 / D
C x n 1
x11 1 .x2 x1 /2 1 1 : : :
.xn1 xn2 /n1 n2 1 .x xn1 /n n1 1 1Sn;x ; where C WD C.1 ; : : : ; n / is a positive constant and Sn;x D f.x1 ; : : : ; xn1 / 2 Rn1 W 0 x1 xn1 xg: Hence, EŒ. 1 ; : : : ; n /j n D x Z .x1 ; : : : ; xn1 ; x/fnx .x1 ; : : : ; xn1 /dx1 : : : dxn1 D Sn;x
Z
DC Sn;1
.xy1 ; : : : ; xyn1 ; x/y11 1 .y2 y1 /2 1 1 : : :
.yn1 yn2 /n1 n2 1 .1 yn1 /n n1 1 dy1 : : : dyn1 after the change of variables: xi D xyi ; i D 1; : : : ; n1. It is then clear that since increases with respect to each of its arguments, this last expression is an increasing function with respect to x. Corollary 1. Let . ; 0/ be the gamma subordinator. Then, for every finite positive measure on RC , and for every p > 0, the process: Z .;p/ At WD
1
e
t . /p h;p .t /
.d/ ; t 0
(12)
0
is a peacock. Here, the function h;p is defined as: h;p .t/ D log E Œexp .t. /p / : Proof (of Corollary 1). p By Remark 2 with .x/ D x p for x 0, the process .X WD ; 0/ is conditionally monotone. Since it is a negative process, (5) is obviously satisfied. Moreover, since . ; 0/ is an increasing process, (6) is easily verified. Finally, Theorem 1 holds. t u
Some Examples of Peacocks in a Markovian Set-Up
289
Remark 3. Actually, for p D 1, Corollary 1 holds more generally with a signed measure, see [8].
3.1.2 The Simple Random Walk Is Conditionally Monotone Let ."i ; i 2 N / be a sequence of independent and identically distributed r. v.’s such that, for every i 2 N : P."i D 1/ D p;
P."i D 1/ D q
with p; q > 0 and p C q D 1:
Let .Sn ; n 2 N/ be the random walk defined by: S0 D 0 and Sn D
n X
"i ;
for every n 2 N
i D1
We shall prove that .Sn ; n 2 N/ is conditionally monotone; i.e: for every r 2 2; C1, every 0 n1 < n2 < < nr < C1 and every function W Rr1 ! R in Er1 , k 2 Inr 7! EŒ.Sn1 ; Sn2 ; : : : ; Snr1 /jSnr D k is an increasing function
(13)
where Ix x; x denotes the set of all the values the r.v. Sx can take. It is not difficult to see that (13) holds if and only if: for every N 2 2; C1 and every function W RN 1 ! R in EN 1 : k 2 IN 7! EŒ.S1 ; : : : ; SN 1 /jSN D k is an increasing function on IN : (14) We shall distinguish two cases: 1. If N and k are even, we set N D 2n .n 2 1; C1/ and k D 2x .x 2 n; n/. 2x , the set For every n 2 1; C1 and every x 2 n; n, let us denote by J2n of polygonal lines ! WD .!i ; i 2 0; 2n/ such that !0 D 0, !pC1 D !p ˙ 1, 2x .p 2 0; 2n 1/ and !2n D 2x. Observe that any ! 2 J2n has n C x positive slopes and n x negative ones. This implies that: nCx 2x jJ2n j D C2n ;
where j j denotes cardinality. It is well known that, conditionally on fS2n D 2xg, 2x the law of the random vector .S0 ; S; : : : ; S2n / is the uniform law on J2n . Let n 2 1; C1 and x 2 n; n be fixed and consider, for every i 2 1; n C x C 1 the map: 2xC2 2x ˘i W J2n ! J2n 2xC2 defined by: for every ! 2 J2n , ˘i .!/ has the same negative slopes and the same positive slopes as ! except the i th positive slope which is replaced by a negative one.
290
A.-M. Bogso et al.
2xC2 For every ! 2 J2n and every function W R2n ! R in E2n ,
.!/ .˘i .!//: Summing this relation, we obtain: X
.n C x C 1/
X
nCxC1 X
2xC2 !2J2n
i D1
.!/
2xC2 !2J2n
D
X
nCxC1 X
2x !2J2n
i D1
D .n x/
.˘i .!//
j˘i1 .!/j.!/
X
.!/:
2x !2J2n
Thus, we have proved the following: Lemma 1. For every n 2 N and every W R2n ! R in E2n , X
1 2xC2 jJ2n j
.!/
2xC2 !2J2n
X 1 .!/; 2x jJ2n j 2x
(15)
!2J2n
which means that .Sn ; n 2 N/ is conditionally monotone. 2. It is not difficult to establish a similar result when k and N are odd. CorollaryP2. For every odd and positive integer p, and for every positive finite measure an ın on N: n2N C1 X
! an e
t .Sn /p hn;p .t /
;t 0
is a peacock.
nD0
Here, the function hn;p is defined by: hn;p .t/ D log E Œexp .t.Sn /p /.
3.1.3 The Processes with Independent Log-Concave Increments Are Conditionally Monotone We first introduce the notions of PF2 and log-concave random variables (see [4]). Definition 3 (R-valued PF2 r.v.’s). An R-valued random variable X is said to be PF2 if: 1. X admits a probability density f ,
Some Examples of Peacocks in a Markovian Set-Up
291
2. For every x1 x2 , y2 y1 , det
f .x1 y1 / f .x1 y2 / f .x2 y1 / f .x2 y2 /
0:
Definition 4 (Z-valued PF2 r.v.’s). A Z-valued random variable X is said to be PF2 if, setting f .x/ D P.X D x/ .x 2 Z/, one has: for every x1 x2 , y2 y1 ,
f .x1 y1 / f .x1 y2 / det f .x2 y1 / f .x2 y2 /
0:
Definition 5 (R-valued log-concave r.v.’s). An R-valued random variable X is said to be log-concave if: 1. X admits a probability density f , 2. The set Sf WD ff > 0g is convex (i.e., is an interval) and log f is concave on Sf (i.e., the second derivative of log f (in the distribution sense) is a negative measure). Definition 6 (Z-valued log-concave r.v.’s). A Z-valued random variable X is said to be log-concave if, with f .x/ D P.X D x/, .x 2 Z/, the set Sf WD ff > 0g is an interval of Z, and for every n; n1; nC12Sf , f 2 .n/ f .n 1/f .n C 1/I in other words, the discrete second derivative of log f is negative on Sf . We then deduce the equivalence: Theorem 2 (see [1] or [4]). An R-valued (or Z-valued) random variable is PF2 if and only if it is log-concave. Example 1. Many common density functions on R (or Z) are PF2 . Indeed, the normal density, the uniform density, the exponential density, the negative binomial density, the Poisson density and the geometric density are PF2 . We refer to [1] for more examples. Note that: 1. A gamma random variable of parameter a .with density fa .x/ D 1 x a1 e x 1Œ0;C1Œ .x/, a > 0) is not PF2 if a < 1, .a/ 2. A Bernoulli random variable X such that P.X D 1/ D p D 1 P.X D 1/ is not PF2 . The following result is due to Efron [5] (see also [22]). Theorem 3. Let n 2 1; C1, X1 ; XP 2 ; : : : ; Xn be independent R-valued .or Zvalued/ PF2 random variables, Sn D niD1 Xi , and W Rn ! R belonging to En . Then,
292
A.-M. Bogso et al.
EŒ.X1 ; X2 ; : : : ; Xn /jSn D x
is increasing in x:
Thanks to Theorem 3, we obtain the following result: Theorem 4. Let .Z ; 2 RC or 2 N/ be a R-valued .or Z-valued) process satisfying (5) and (6), with independent .not necessarily time-homogeneous/ PF2 increments. Then, .Z ; 0/ is conditionally monotone, and for every positive measure on RC (or N) with finite total mass, Z
C1
e
t Z h .t /
.d/; t 0
is a peacock,
0
where the function h is defined by: h .t/ D log E e t Z . Proof (of Theorem 4). It suffices to show that .Z ; 0/ satisfies (10). Let n 2 1; C1 and W Rn ! R belonging to En . For every 0 1 < 2 < < n and k 2 R (or Z), EŒ.Z1 ; Z2 ; : : : ; Zn /jZn D k h i DE b .Z1 ; Z2 Z1 ; : : : ; Zn Zn1 /jZn D k ; where the function b is given by: b .x1 ; x2 ; : : : ; xn / D .x1 ; x2 C x1 ; : : : ; x1 C x2 C C xn /: It is obvious that b belongs to En . Thus, applying Theorem 3 with: X1 D Z1 and Xi C1 D Zi C1 Zi i D 1; : : : n 1, one obtains the desired result. t u Remark 4. 1. Theorem 4 does not apply neither in the case of the Gamma subordinator, nor in the case of the random walk whose increments are Bernoulli with values in f1; 1g. Nevertheless, its conclusion remains true in these cases, see Sects. 3.1.1 and 3.1.2. 2. We deduce from Corollary 4 that the Poisson process and the random walk with geometric increments are conditionally monotone. We shall give below a direct proof, i.e., without using Theorem 3. 3.1.4 The Poisson Process Is Conditionally Monotone Let .N ; 0/ be a Poisson process with parameter 1 and let .Tn ; n 1/ be its successive jumps times. Then N D #fi 1 W Ti g:
Some Examples of Peacocks in a Markovian Set-Up
293
In order to prove that .N ; 0/ is conditionally monotone, we shall show that for every 0 1 < < n and every function W Rn ! R in En , we have: EŒ.N1 ; : : : ; Nn /jNn D n .Nn /;
(16)
where n W R ! R increases. But, conditionally on fNn D kg, the random vector .T1 ; : : : ; Tk / is distributed as .U1 ; : : : ; Uk /, U1 ; : : : ; Uk being the increasing rearrangement of k independent random variables, uniformly distributed on Œ0; n . We go from k to k C 1 by adding one more point. Thus, with obvious notation, it .kC1/ .k/ N . Then, the conditional monotonicity is clear that: for all 2 Œ0; n , N property follows immediately. Corollary 3. Let .N ; 0/ be a Poisson process and let be a finite positive measure on RC . Then, for every p > 0, the process: Z .;p/ At WD
1
e t .N /
p h ;p .t /
.d/; t 0
(17)
0
is a peacock with: h;p .t/ D log E Œexp .t.N /p / :
3.1.5 The Random Walk with Geometric Increments Is Conditionally Monotone Let ."i ; i 2 1; C1/ be a sequence of independent geometric variables with the same parameter p; i.e., such that: P."i D k/ D p k .1 p/
.k 0; 0 < p < 1/:
We consider the random walk .Sn ; n 2 N/ defined by: S0 D 0 and Sn D
n X
"i ; for every n 2 N :
i D1
For n 2 N , Sn is distributed as a negative binomial random variable with parameters n and p; more precisely: k P.Sn D k/ D CnCk1 p n .1 p/k ; for every k 2 N:
As in Sect. 3.1.2, we only need to prove that: for every N 2 N and every function W RN ! R in EN : k 7! EŒ.S1 ; : : : ; SN /jSN C1 D k
is an increasing function on N:
(18)
294
A.-M. Bogso et al.
Let JkN denote the set: JkN WD f.x1 ; : : : ; xN / 2 NN W 0 x1 xN kg:
(19)
For every k 0 and N 1, it is well known that jJkN j D CNk Ck . Now, we have: E Œ .S1 ; : : : ; SN / jSN C1 D k X
D
.l1 ; : : : ; lN /
P.S1 D l1 ; : : : ; SN D lN ; SN C1 D k/ P.SN C1 D k/
.l1 ; : : : ; lN /
P.S1 D l1 ; S2 S1 D l2 l1 ; : : : ; SN C1 SN D k lN / P.SN C1 D k/
.l1 ; : : : ; lN /
P.S1 D l1 /P.S2 S1 D l2 l1 / : : : P.SN C1 SN D k lN / P.SN C1 D k/
.l1 ;:::;lN /2JkN
X
D
.l1 ;:::;lN /2JkN
X
D
.l1 ;:::;lN /2JkN
X
D
.l1 ; : : : ; lN /
p.1 p/l1 p.1 p/l2 l1 : : : p.1 p/klN CNk Ck p N C1 .1 p/k
.l1 ;:::;lN /2JkN
D
D
1
X
CNk Ck .l1 ;:::;lN /2JkN 1
X
jJkN j .l1 ;:::;lN /2JkN
.l1 ; : : : ; lN /
.l1 ; : : : ; lN / :
Therefore, the law of the random vector .S1 ; : : : ; SN / conditionally on fSN C1 D kg is the uniform law on the set JkN . Hence, we will obtain (18) if we prove that: for every k 2 N, every N 2 N and every function W RN ! RC in EN : X 1 1 X .x/ kC1 .x/: k jJN j jJN j k kC1 x2JN
(20)
x2JN
Let us notice that: J0N D f.0; : : : ; 0/g; „ ƒ‚ …
for every N 2 1; C1
N times
and Jk1 D f.0/; .1/; : : : ; .k/g;
for every k 2 0; C1:
For k 2 0; C1 and N 2 1; C1, we define: WD JkC1 n JkN D f.x1 ; : : : ; xN / 2 JkC1 W xN D k C 1g: kC1 N N N
(21)
Some Examples of Peacocks in a Markovian Set-Up
295
and set 0N D ;. By Pascal’s formula, kC1 k kC1 kC1 jkC1 N j D CkC1CN CkCN D CN Ck D jJN 1 j;
.with N 2 2; C1/:
kC1 On one hand, we consider, for N 2 2; C1, the map W JkC1 defined N 1 ! N by: Œ.x1 ; : : : ; xN 1 / D .x1 ; : : : ; xN 1 ; k C 1/: (22)
The map is bijective, and for every non empty pair of subsets G and H of JkC1 N 1 , there is the equivalence: 8 8 f W RN 1 ! R 2 EN 1 ; ˆ ˆ <
8 8 W RN ! R 2 EN ; ˆ ˆ < ” P P 1 1 1 P 1 P ˆ ˆ ˆ ˆ .z/ .z/ f .x/ f .x/ : : j .G/j z2 .G/ j .H /j z2 .H / jGj x2G jH j x2H On the other hand, for N 2 2; C1, let ƒ W kN ! kC1 be the injection given N by: ƒŒ.x1 ; : : : ; xN 1 ; k/ D .x1 ; : : : ; xN 1 ; k C 1/: (23) For every z 2 kN and function W RN ! R in EN , .z/ .ƒ.z//: Therefore, for every non empty subset K of kN , X 1 X 1 .z/ .u/: jKj z2K jƒ.K/j
(24)
u2ƒ.K/
since jKj D j .K/j. Furthermore, one notices that: 1 Œƒ.kN / D JkN 1
kC1 1 .kC1 N / D JN 1
and
where 1 denotes the inverse map of . Hence, the following is easily obtained: Lemma 2. Let k 2 1; C1 and N 2 2; C1. Assume that for every function f W RN 1 ! R in EN 1 W 1 jJkN 1 j
X x2JkN 1
f .x/
1 jJkC1 N 1 j
Then, for every function W RN ! R in EN ,
X x2JkC1 N 1
f .x/:
(25)
296
A.-M. Bogso et al.
X
1 jkN j
.y/
y2kN
X
1 jkC1 N j
.y/:
(26)
y2kC1 N
Now, we are able to prove (20) by induction on N 2 1; C1 and k 2 0; C1. Proposition 2. Let k 2 0; C1, N 2 1; C1 and let W RN ! R be any function in EN . Then, X 1 X 1 .z/ .z/I jJkN j k jJkC1 N j kC1 z2JN
(27)
z2JN
in other words, .Sn ; n 2 N/ is conditionally monotone. Proof (of Proposition 2). 1. It is obvious that (27) holds for .k; N / 2 0; C1f1g, and for .k; N / 2 f0g 1; C1. 2. Let .k; N / 2 1; C12; C1. We assume that: 8 .l; m/ 2 D WD 0; k 1 1; C1 [ fkg 1; N 1 (see Fig. 1) and any function f W Rm ! R in Em : X 1 1 X f .x/ lC1 f .x/: l jJm j jJm j l lC1 x2Jm
(28)
x2Jm
By taking .l; m/ D .k; N 1/ in (28), Lemma 2 yields: 1
X
jkN j y2kN
.y/
1
X
jkC1 N j y2kC1 N
.y/:
(29)
On the other hand, from the definition of kC1 N , (27) is equivalent to: X 1 X 1 .y/ .y/: jJkN j jkC1 N j k kC1 y2JN
(30)
y2N
Using (28) with .l; m/ D .k 1; N /, we have: 1 X 1 X .y/ .y/: jJkN j jkN j k k y2JN
(31)
y2N
The comparison of (29) with (31) yields (30) which is equivalent to (27) (Fig. 1). t u
Some Examples of Peacocks in a Markovian Set-Up Fig. 1 D WD 0; k 1 1; C1[ fkg 1; N 1
297
m ( k; N )
N
D
N −1
1
k −1
1
0
P
Corollary 4. For every positive finite measure
k
l
an ın on N and every p > 0 W
n2N C1 X
! an e
t .Sn /p hn;p .t /
;t 0
is a peacock,
nD0
where the function hn;p is defined by: hn;p .t/ D log E Œexp .t.Sn /p /. Remark 5. The result in this example has to be compared with that of Sect. 3.1.1: we replace the gamma r.v’s by geometric ones.
3.2 Diffusions Which Are “Well-Reversible” at Fixed Times Are Conditionally Monotone Let us now present an important class of conditionally monotone processes: that of the “well-reversible” diffusions at a fixed time. 3.2.1 The Diffusion .X ; 0I Px ; x 2 R/ Let W RC R ! R and b W RC R ! R be two Borel measurable functions and let .Bu ; u 0/ be a standard Brownian motion starting from 0. We consider the SDE: Z Z
X D x C 0
.s; Xs / dBs C
b.s; Xs / ds; 0:
(32)
0
We assume that: .x/
A1. For every x 2 R, this SDE admits a unique pathwise solution .X ; 0/, .x/ and furthermore the mapping x 7! .X ; 0/ may be chosen measurable.
298
A.-M. Bogso et al. .x/
As a consequence of (A1), from Yamada-Watanabe’s theorem, .X ; 0/ is a strong solution of (32), and it enjoys the strong Markov property; finally the .x/ transition kernel P .x; dy/ D P.X 2 dy/ is measurable. .y/ We now remark that, for x y, the process .X ; 0/ is stochastically greater .x/ than .X ; 0/ in the following sense: for every a 2 R and 0, .y/ .x/ P X a P X a : .x/
(33)
.y/
Indeed, assuming that both .X ; 0/ and .X ; 0/ are defined on the same probability space, and setting .x/
.y/
T D inff 0I X D X g .x/
.y/
.D C1 if f 0I X D X g D ;/; it is clear that, on fT D C1g, .y/
.x/
X X
.since y x/
while on fT < C1g, we have: .y/
.x/
X > X and
.y/
.x/
X D X
for every 2 Œ0; T Œ for every 2 ŒT; C1Œ
since, as a consequence of our hypothesis (A1), (32) admits a unique strong Markovian solution. On the other hand, (33) is equivalent to: for every bounded and increasing (resp. decreasing) function, and for every 0: Z x ! Ex Œ.X / D
R
P .x; dy/.y/ is increasing (resp. decreasing):
(34)
Lemma 3. Let ..X /0 ; .F /0 ; .Px /x2R / be a Markov process in R which satisfies (33). Then, for every n 1, every 0 < 1 < < n , every i 2 f1; : : : ; ng, every function W Rn ! R in En , and for every x 0, i .X1 ; : : : ; Xi /; Ex Œ.X1 ; : : : ; Xn /jFi D e
(35)
where e i W Ri ! R belongs to Ei . In particular, x ! Ex Œ.X1 ; : : : ; Xn / is increasing:
(36)
Some Examples of Peacocks in a Markovian Set-Up
299
Proof (of Lemma 3). If i D n, (35) is obvious. If i D n 1, then (35) is satisfied since: Ex Œ.X1 ; : : : ; Xn1 ; Xn /jFn1 Z D .X1 ; : : : ; Xn1 ; y/Pn n1 .Xn1 ; dy/ R
and then, for i D n1, (35) follows immediately from (34). Thus, Lemma 3 follows by iteration of this argument. t u Observe that as a consequence of Lemma 3, the conditional monotonicity property (4) for these diffusions is equivalent to (10).
3.2.2 Time-Reversal at a Fixed Time Let x 2 R fixed. We assume that: A2. For every > 0, .; / is a differentiable function and X admits a C 1;2 density function p on 0; C1ŒR. By setting a.; y/ WD 2 .; y/
for every 0 and y 2 R;
we define successively, for any fixed 0 > 0 and for y 2 R: 8 a 0 .; y/ D a.0 ; y/; ˆ ˆ ˆ ˆ <
.0 0 /
@ 1 0 ˆ a.0 ; y/ p.0 ; y/ ; b .; y/ D b.0 ; y/ C ˆ ˆ p.0 ; y/ @y ˆ : .0 < 0 / (37) and the differential operator L0 , .0 < 0 /: L0 f .x/ D
1 0 0 a .; y/f 00 .y/ C b .; y/f 0 .y/ 2
for f 2 Cb2 :
Under some suitable conditions on a and b, Haussmann and Pardoux [6] (see also Meyer [17]) proved that: 0
A3. The process .X ; 0 < 0 / obtained by time-reversing .X ; 0 < 0 / at time 0 : 0 .X ; 0 < 0 / WD .X0 ; 0 < 0 / is a diffusion and there exists a Brownian motion .B u ; 0 u 0 /, 0 independent of X0 , such that .X ; 0 < 0 / solves the SDE:
300
A.-M. Bogso et al.
8 ˆ ˆ < d Y
D 0 .; Y /d B C b .; Y /d
ˆ ˆ :Y
D X0
0
.0 < 0 / (38)
0
.with .; y/ D .0 ; y//: 0
0
Note that the coefficients b and 0 depend on x. A4. We assume furthermore that the SDE (38) admits a unique strong solution on Œ0; 0 Œ; thus, this strong solution is strongly Markovian. Note that, a priori, the solution of (38) is only defined on Œ0; 0 Œ, but it can be 0 extended on Œ0; 0 by setting X 0 D x. 3.2.3 Our Hypotheses and the Main Result Our goal here is not to give optimal hypotheses under which the assertions (A1)– (A4) are satisfied. We refer the reader to [6] or [18] for more details. Instead, we shall present two hypotheses (H1) and (H2), either of them implying the preceding assertions: H1. We assume that: i. The functions .; y/ 7! .; y/ and .; y/ 7! b.; y/ are of C 1;2 class on 0; C1ŒR, locally Lipschitz continuous in y uniformly in , and the solution of (32) does not explode on Œ0; 0 , ii. There exists ˛ > 0 such that: a.; y/ 2 .; y/ ˛ and
for every y 2 R and 0 0 ;
@2 a 2 L1 .0; 0 RC /: @y 2
H2. We assume that: i. The functions and b are of C 1;2 class, locally Lipschitz continuous in y uniformly in , and the solution of (32) does not explode on Œ0; 0 , ii. The functions a and b are of C 1 class on 0; C1ŒR in .; y/ and the differential operator @ LD C L @ is hypoelliptic (see Ikeda–Watanabe [11, p. 411] for the definition and properties of hypoelliptic operators), where .L ; 0/ is the generator of the diffusion (32): L D
d 1 d a.; / 2 C b.; / : 2 dy dy
(39)
Some Examples of Peacocks in a Markovian Set-Up
301
Then, under either (H1) or (H2), the assertions .Ai /i D1:::4 of both Sects. 3.2.1 0 and 3.2.2 are satisfied, see [6]. In particular, .X ; 0 < 0 / is a strong solution of (38), see Meyer [17]. Let us now give the main result of this section. Theorem 5. Under either (H1) or (H2), and for every x2R, the process .X ; >0/ is conditionally monotone under Px . Proof (of Theorem 5). Let n 2 N and let W Rn ! R in En . For every 0 < 1 < < n and every i 2 f1; : : : ; ng: Ex Œ.X1 ; : : : ; Xn /jXi D z D Ex ŒEx Œ.X1 ; : : : ; Xn /jFi jXi D z D Ex e i .X1 ; : : : ; Xi /jXi D z .by Lemma 3, where e i W Ri ! R belongs to Ei i h i i i D Ex e .by time-reversal at i / i .X i 1 ; : : : ; X 0 /jX 0 D z h i i i i .X i 1 ; : : : ; X i i 1 ; z/ D Ez e i
and, by applying (36) to the reversed process .X ; 0 < i /, this last expression is a bounded function which increases with respect to z. t u Remark 6. Observe that we were careful to exclude the point 1 D 0 in Theorem 5, 0 since “well-reversible” diffusions can be only reversed a priori on 0; 0 : .X ; 0 < 0 / WD .X0 ; 0 < 0 /. Corollary 5. Let .X ; > 0/ the unique strong solution of (32), taking values in RC , where b and satisfy either (H1) or (H2). Then, for every finite positive measure on 0; C1Œ and for every p > 0, the process: Z .;p/ At WD
1
e t .X /
p h ;p .t /
.d/; t 0
(40)
0
is a peacock, with: h;p .t/ D log Ex Œexp .t.X /p / : Proof (of Corollary 5). Let " > 0 and define ."/ to be the restriction of to the interval Œ"; C1Œ: ."/ WD jŒ";C1Œ . As .X ; "/ is a continuous positive process, conditions (5) and (6) are satisfied, and we may apply Theorems 5 and 1 to obtain that, for every 2 C and every 0 s t: h i h i ."/ .."/ / E .As. / / E .At / :
302
A.-M. Bogso et al.
Then, proceeding as in Point (4) of the proof of Theorem 1, the result follows by letting " tend to 0. u t
3.2.4 A Few Examples of Diffusions Which are “Well-Reversible” at Fixed Times Example 2 (Brownian motion with drift ). We take 1; b.s; y/ D and X D x C B C . Then, p.t; x; y/ D p
.y .x C t//2 exp ; 2t 2 t 1
0
and .X ; 0 < 0 / is the solution of: Y D
0 X0
Z
C B C 0 0
x Yu du 0 u .x/
with .B ; 0 < 0 / independent from X 0 D X0 . See Jeulin–Yor [12] for similar computations. Example 3 (Bessel processes of dimension ı 2). ı1 We take 1 and b.s; y/ D , with ı D 2. C 1/, ı 2. Then, 2y i. For x > 0: 2 1 y C1 xy x C y2 p.t; x; y/ D I ; exp t x 2t t where I denote the modified Bessel function of index (see Lebedev [14, p.110] 0 for the definition of I ), and .X ; 0 < 0 / is the solution of: Y D
0 X0
Z
C B C 0
x I0 1 Yu C 2Yu 0 u 0 u I
xYu 0 u
ii. For x D 0: p.t; 0; y/ D
2 1 y 2C1 y ; exp C1 2 t . C 1/ 2t
0
and .X ; 0 < 0 / is the solution of: Y D
0 X0
Z
C B C 0
2 C 1 Yu du 2Yu 0 u
d u:
Some Examples of Peacocks in a Markovian Set-Up
303
This examples has a strong likelihood with Bessel processes with drift, see Watanabe [24]. Example 4 (Squared Bessel processes of dimension ı > 0). p We take .s; y/ D 2 y and b ı. Then: i. For x > 0: p xy 1 y =2 xCy p.t; x; y/ D I ; exp 2t x 2t t 0
and .X ; 0 < 0 / is the solution of: 0
Y D X 0 C 2
Z
Z p Yu dBu C 2 2
0
0
p p Yu xYu I0 xYu d u: u 0 u I 0 u
ii. For x D 0: p.t; 0; y/ D
1 2t
ı=2
y 1 ı y 2 1 exp ; .ı=2/ 2t
0
and .X ; 0 < 0 / is the solution of: 0
Y D X 0 C 2
Z
Z p Yu dBu C ı
0
0
2Yu d u: u
Note that we could also have obtained this example by squaring the results on Bessel processes. Remark 7. All the above examples have a strong link with initial enlargement of a filtration (by the terminal value). We refer the reader to Mansuy–Yor [16] for further examples.
4 Another Class of Markovian Peacocks We shall introduce another set of hypotheses on the Markov process .X ; 0/ which ensures that: Z 1 ./ tX h .t / At WD e .d/; t 0 0
is a peacock. Definition 7. A right-continuous Markov process .X ; 0I Px ; x 2 RC /, with values in RC , is said to satisfy condition (L) if both (i) and (ii) below are satisfied:
304
A.-M. Bogso et al.
i. This process increases in the stochastic order with respect to the starting point x; in other words, for every a 0 and 0, and for every 0 x y: Py .X a/ Px .X a/:
(41)
ii. The Laplace transform Ex Œe tX is of the form: Ex Œe tX D C1 .t; / exp.x C2 .t; //;
(42)
where C1 and C2 are two positive functions such that: • For every t > 0 and 0, @ C2 .t; / > 0: @t
(43)
• For every t 0 and every compact K, there exist two constants kK .t/ > 0 and e k K .t/ < C1 such that: kK .t/ inf C1 .t; /I 2K
sup C2 .t; / e k K .t/:
(44)
2K
Taking x D 0 in (42), we see that C1 .˘; / is completely monotone (and hence infinitely differentiable) on 0; C1Œ and continuous at 0. Consequently, C2 .˘; / is also infinitely differentiable on 0; C1Œ and continuous at 0. Moreover, we have for t > 0 and 0: @ @ Ex X e tX D C1 .t; / C xC1 .t; / C2 .t; / exp .xC2 .t; // @t @t and we introduce: 8 ˆ <˛.t; /
@ C1 .t; / 0 @t ˆ :ˇ.t; / WD C1 .t; / @ C2 .t; / > 0 @t WD
(45)
We can now state the main result of this section. Theorem 6. Let .X ; 0I Px ; x 2 RC / be a Markov process which satisfies condition (L). Then, for every x 0 and every finite positive measure on RC , Z ./ At WD
1
e
tX h .t /
.d/; t 0
0
is a peacock under Px . Here, the function h is defined by: h .t/ D log Ex e tX :
Some Examples of Peacocks in a Markovian Set-Up
305
Before proving Theorem 6, let us give two examples of processes .X ; 0I Px ; x 2 RC / which satisfy condition .L/. Example 5. Let .X ; 0I Qx ; x 2 RC / be the square of a ı-dimensional Bessel process (denoted BESQı , ı 0, see [19, Chap. XI or [23]]). This process satisfies condition .L/ since: • It is stochastically increasing with respect to x; indeed, it solves a SDE which enjoys both existence and uniqueness properties, hence the strong Markov property (see Sect. 3.2.1). • For every t > 0, we have: Qx e tX D
1 ı
.1 C 2t/ 2
exp
tx ; 1 C 2t
which yields Point (ii) of Definition 7. ./
In particular, for .Xt ; t 0/ a squared Bessel process of dimension 0, .At ; t 0/ is a peacock. This case was outside the scope of Example 4. Example 6 (A generalization of the preceding example for ı D 0). Let .X ; 0I Px ; x 2 RC / be a continuous state branching process (denoted CSBP) (see [15]).We denote by P .x; dy/ the law of X under Px , (with x ¤ 0), and by the convolution product. Then .P / satisfies: P .x; ˘/ P .x 0 ; ˘/ D P .x C x 0 ; ˘/
for every 0; x 0 and x 0 0
which easily implies (41) (see [15, pp. 21–23]). On the other hand, one has: Ex e tX D exp.x C.t; //;
(46)
where the function C W RC RC ! RC satisfies: • For every 0, the function C.˘; / is continuous on RC and differentiable on 0; C1Œ, and @C .t; / > 0 for every t > 0; @t • For every t 0 and every compact K, there exists a constant kK .t/ < 1 such that: sup C.t; / kK .t/: (47) 2K
Thus, .X ; 0/ satisfies (42). Corollary 6. Let .X ; 0I Px ; x 2 RC / be either a BESQı or a CSBP. Then, for any finite positive measure on RC , and for every x 0: Z ./ At WD
1 0
e tX h .t / .d/; t 0
306
A.-M. Bogso et al.
is a peacock under Px with: h .t/ D log Ex e tX : Remark 8. This example generalizes the previous one in the following sense. Let .Yt ; t 0/ be a Lévy process of characteristic exponent ./ D c˛ , .c > 0; ˛ 2 1; 2/: E e Yt D exp .ct˛ / : We denote by .Ht ; t 0/ the height process associated to .Yt ; t 0/. This process admits a family of local times .Lat .H /; t 0; a 0/ and, denoting by r .H / WD inffs 0I L0s .H / > rg its right-continuous inverse, it is known (see [15]) that the process .Lar .H / ; a 0/ is a stable CSBP of index ˛. Then, observe that for ˛ D 2 and c D 12 , .Yt WD Bt ; t 0/ is a standard Brownian motion started from 0, (law)
.Ht ; t 0/ D .jBt j; t 0/ has the same law as a reflected Brownian motion, and that, from the Ray-Knight theorem, .Lar .H / ; a 0/ is a squared Bessel process of dimension 0 started from r. We refer the interested reader to [10, Chap. 4] for a description of other peacocks constructed from CSBP, and their associated martingales. Proof (of Theorem 6). Let .X ; 0/ be a process which enjoys condition .L/. 1. .X ; 0/ being a negative process, condition (5) clearly holds. Moreover, by (44), (6) also holds. Thus, following the proof of Theorem 1, it suffices to show ./ that .At ; t 0/ is a peacock when is a finite linear combination of Dirac measures with positive coefficients. 2. For t 0, a1 0; : : : ; an 0 and 0 1 < < n , we set: At WD
n X
ai e tXi hi .t / :
i D1
Let
2 C. One has: @ Ex Œ .At / D Ex @t
" 0
.At /
n X i D1
# ai e
tXi hi .t /
.h0i .t/
C Xi /
and, we shall prove as in the proof of Theorem 1 that, for every i 2 f1; : : : ; ng, i 0, with:
.At /e tXi hi .t / .h0i .t/ C Xi / D Ex 0 .At /ei .Xi / ;
i D Ex
0
Some Examples of Peacocks in a Markovian Set-Up
307
and where we have set ei .z/ WD e t zhi .t / .h0i .t/ C z/: We note, since E e tXi hi .t / D 1, that: Ex Œei .Xi / D 0:
(48)
Since the function 0 .x1 ; : : : ; xn / !
0
@
n X
1 aj e txj hj .t / A
j D0
is bounded and decreases with respect to each of its arguments, it suffices to show that: for every bounded Borel function W Rn ! RC which decreases with respect to each of its arguments, and for every i 2 f1; : : : ; ng, Ex Œ.X1 ; : : : ; Xn /ei .Xi / 0:
(49)
3. We now show (49). (a) We may suppose i D n. Indeed, thanks to (41) and to Lemma 3, we have, for i < n: Ex Œ.X1 ; : : : ; Xn /ei .Xi / D Ex ŒEx Œ.X1 ; : : : ; Xn /jFi ei .Xi / D Ex Œe i .X1 ; : : : ; Xi /ei .Xi /; where e i W Ri ! R is a bounded Borel function which decreases with respect to each of its arguments. (b) On the other hand, one has: Ex Œe i .X1 ; : : : ; Xi /ei .Xi / D Ex Œe i .X1 ; : : : ; Xi /e tXi hi .t / .h0i .t/ C Xi / i .X1 ; : : : ; Xi 1 ; h0i .t//ei .Xi / Ex Œe i .X1 ; : : : ; h0i .t//.h0i .t/ C Xi /) (since e i .X1 ; : : : ; Xi /.h0i .t/ C Xi / e h i e D Ex e i .X1 ; : : : ; Xi 1 /ei .Xi / ; e where e i W Ri 1 ! R is a bounded Borel function which decreases with respect to each of its arguments, and is defined by: e e i .z1 ; : : : ; zi 1 ; h0i .t//: i .z1 ; : : : ; zi 1 / D e
(50)
308
A.-M. Bogso et al.
(c) We now end the proof of Theorem 6 by showing the following lemma. Lemma 4. For every i 2 f1; : : : ; ng and j 2 f0; 1; : : : ; i 1g, let W Rj ! R be a bounded Borel function which decreases with respect to each of its arguments. Then, Ex Œ.X1 ; : : : ; Xj /ei .Xi / 0: (51) In particular, Ex Œ.X1 ; : : : ; Xi 1 /ei .Xi / 0:
(52)
Proof (of Lemma 4). We prove this lemma by induction on j . • For j D 0, is constant and one has: Ex Œ ei .Xi / D Ex Œei .Xi / D 0 .from (48)/ • On the other hand, if one assumes that (51) holds for 0 j < i 1, then Ex Œ.X1 ; : : : ; Xj ; Xj C1 /ei .Xi / D Ex Œ.X1 ; : : : ; Xj ; Xj C1 /Pi j C1 ei .Xj C1 / .by the Markov property/ X
C .t;
/h .t /
D Ex Œ.X1 ; : : : ; Xj ; Xj C1 /e j C1 2 i j C1 i : ˛.t; i j C1 / C Xj C1 ˇ .t; i j C1 / (53) where, from (42) and (45), ˇ > 0; and ˛ depends on a and h0i
˛.t; i j C1 / Pi j C1 ei .Xj C1 / Ex X1 ; : : : ; Xj ; ˇ .t; i j C1 / .X1 ; : : : ; Xj /ei .Xi / 0 .by the induction hypothesis/; D Ex e where e W Rj ! R is defined by: ˛.t; i j C1 / e : .z1 ; : : : ; zj / D z1 ; : : : ; zj ; ˇ .t; i j C1 / t u
5 Stochastic and Convex Orders The purpose of this section is different from that of the previous sections. Here, we do not look a priori for peacocks, but rather study a link between the stochastic and convex orders. As a byproduct, this will provide us with some new peacocks.
Some Examples of Peacocks in a Markovian Set-Up
309
Definition 8. Let and be two probability measures on RC . We shall say that is stochastically greater than , and we write: (st)
if for every t 0, F .t/ WD .Œ0; t/ F .t/ WD .Œ0; t/: 1 In [10], the authors prove that if .Mt ; t 0/ is a martingale in Hloc (thus, it is a peacock), and ˛ W RC ! RC is a continuous and strictly increasing function such that ˛.0/ D 0, then the process
1 ˛.t/
Z
t
Mu d˛.u/ ; u 0
0
is a peacock. In other words, for every 0 s t: Z
C1
Mu 0
Z C1 (c) 1 1 1Œ0;t .u/ d˛.u/ 1Œ0;s .u/ d˛.u/: Mu ˛.t/ ˛.s/ 0
Now, it is clear that:
(st) 1 1 1Œ0;t .u/ d˛.u/ 1Œ0;s .u/ d˛.u/; ˛.t/ ˛.s/
and this leads to the following question: which processes .Xt ; 0/ satisfy, for every (st)
couple of probabilities .; / such that , the property: Z
C1
A./ WD
(c)
Z
C1
Xu .d u/ A./ WD 0
Xu .d u/ ‹
(54)
0
Note that such a process .Xt ; t 0/ must be a peacock. Indeed, taking for 0s t, (c)
D ıt and D ıs , we deduce from (54) that Xt Xs , i.e., .Xt ; t 0/ is a peacock. Here is a partial answer to this question: Theorem 7. Let .Xt ; t 0/ be an integrable right-continuous process satisfying both following conditions: i. For every bounded and increasing function W R ! RC and every 0 s t, EŒ.Xt /jFs is an increasing function of Xs . ii. For every n 2 N , every 0 t1 < < tnC1 and every W Rn ! R in En , we have: E .Xt1 ; : : : ; Xtn / .XtnC1 Xtn / 0:
310
A.-M. Bogso et al. (st)
Let and two probability measures on RC such that . Moreover, we assume that either: and have compact supports, and for every compact K RC , sup jXt j is t 2K
integrable, or: sup jXt j is integrable. t 0
Then:
Z
C1
A./ WD
Z
(c)
C1
Xu .d u/ A./ WD 0
Xu .d u/: 0
Remark 9. (a) Observe that condition (ii) implies that the process .Xt ; t 0/ is a peacock. Indeed, if is a convex function of C 1 class, then, for 0 s t: EŒ .Xt / EŒ .Xs / EŒ
0
.Xs /.Xt Xs / 0:
In particular, EŒXt does not depend on t. (b) Note also that condition (i) implies that .Xt ; t 0/ is Markovian. Before proving Theorem 7, we shall give some examples of processes which satisfy both conditions (i) and (ii). Example 7. Let X be a r.v. such that for every t 0, EŒe tX < 1. We define .tX D exp.tX hX .t//; t 0/ where hX .t/ D log EŒe tX . Then, .tX ; t 0/ satisfies the conditions of Theorem 7. Indeed, condition (i) is obvious, and condition (ii) follows from: h i E .tX1 ; : : : ; tXn /.tXnC1 tXn / i h .e t1 ˇn hX .t1 / ; : : : ; e tn ˇn hX .tn / /E tXnC1 tXn D 0; hX .tnC1 / hX .tn / . In particular, we recover that, if ˛ W RC ! RC tnC1 tn is a continuous and strictly increasing function such that ˛.0/ D 0, then the process where ˇn D
1 ˛.t/
Z
t
e uX hX .u/ d˛.u/ ; t 0
0
is a peacock. Example 8 (Martingales). Clearly, martingales satisfy condition (ii). Here are some examples of martingales satisfying also condition (i):
Some Examples of Peacocks in a Markovian Set-Up
311
(a) Let .Xt ; t 0/ be an integrable process with independent and centered increments. Then EŒ.Xt /jFs D EŒ.Xs C Xt Xs /jFs D EŒ.x C Z/; (law)
where Xs D x and Z D Xt Xs , is an increasing function of x. (b) Let .Lt ; t 0/ be an integrable right-continuous process with independent increments, and such that, for every ; t 0, EŒe Lt < 1. Then, the process
Xt WD e Lt hLt ./ ; t 0
where hLt ./ D log E e Lt
is a martingale which, as in item (a), satisfies condition (i). (c) Let .Xt ; t 0/ be a diffusion process which satisfies an equation of type .x/
Xt
Z
t
DxC
.Xs.x/ / dBs :
0
Then condition (i) follows from the stochastic comparison theorem (see Point (A1)). Example 9 (“Well-reversible” diffusions). Let .Zt ; t 0/ be a “well-reversible” diffusion satisfying (32) and such that b is an increasing function. Then .Xt WD Zt EŒZt ; t > 0/ satisfies both conditions (i) and (ii). Indeed, condition (i) is clearly satisfied from Lemma 3. As for condition (ii), setting h.t/ D EŒZt , we have, with 0 < t1 < : : : < tnC1 , by time reversal at tnC1 : E .Xt1 ; : : : ; Xtn / .XtnC1 Xtn / .t / h .t / i .tnC1 / .tnC1 / nC1 nC1 X tnC1 DE X tnC1 t1 ; : : : ; X tnC1 tn .X 0 tn / i .t / h h .t / i .tnC1 / .tnC1 / nC1 nC1 X tnC1 tn / DE E X tnC1 t1 ; : : : ; X tnC1 tn jF tnC1 tn .X 0 h .t / .t / i .tnC1 / nC1 nC1 X tnC1 X / .X DE e 0 tnC1 tn tn where e is an increasing function, DE e .Xtn / .XtnC1 Xtn / : Now
from (32): Z E e .Ztn h.tn // Z D
tnC1 tn
tnC1 tn
Z .s; Zs /dBs C
tnC1
b.s; Zs /ds h.tnC1 / C h.tn /
tn
E e .Ztn h.tn // .b.s; Zs / h0 .s// ds
312
A.-M. Bogso et al.
Z
tnC1
D
tn
h i E e .Ztn h.tn // .e b.s; Ztn / h0 .s// ds
e where x 7! h i b.s; x/ WD EŒb.s; Zs /jZtn D x is an increasing function such that E e b 1 its right-continuous inverse, b.s; Zt / D EŒb.s; Zs / D h0 .s/. Denoting by e s
n
we finally obtain: E e .Xtn / .XtnC1 Xtn / Z tnC1 h i 0 0 e e b.s; Z b 1 .h .s// h.t / E / h .s/ ds D 0: e n t n s tn
Example 10. Let .Bt ; t 0/ be a Brownian motion started from 0 and ' be a strictly increasing odd function of C 2 class such that ' 0 is convex. It is known, see [10, Chap. 1, Sect. 5] that the process .'.Bt /; t 0/ is a peacock. As a consequence 1 1Œ0;t .u/d˛.u/ and of Example 9 and of Theorem 7 applied with .d u/ D ˛.t/ 1 1Œ0;s .u/d˛.u/, where ˛ W RC ! RC is a continuous and strictly .d u/ D ˛.s/ increasing function such that ˛.0/ D 0, we deduce that the process
1 ˛.t/
Z
t
'.Bu /d˛.u/ ; t 0
0
is also a peacock. Indeed, from Itô’s formula, .'.Bu /; u 0/ satisfies (32) with b D 12 ' 00 ı ' 1 increasing. Proof (of Theorem 7). 1. Since Z
Z
1
A./ WD 0
Z
1
Xs d.s/ D
1
Xs dF .s/ D 0
0
XF1 .u/ d u;
it suffices, by approximation of dF with a linear combination of Dirac measures (as in the proof of Theorem 1), to show that for every n 2 N , for every a1 ; a2 ; : : : ; an and for every t1 s1 ; : : : ; tn sn , n X i D1
2. Let
(c)
ai Xti
n X i D1
W R ! R in C. By convexity, we have:
ai Xsi :
(55)
Some Examples of Peacocks in a Markovian Set-Up n X
!
n X
D
ai Xti
i D1
ai Xsi C
i D1 n X
313
n X
! ai .Xti Xsi /
i D1
!
C
ai Xsi
n X
0
i D1
! ai Xsi
i D1
n X
aj .Xtj Xsj /:
j D1
Then, taking the expectation leads to: " E
n X
!# ai Xti
" E
i D1
!#
n X
ai Xsi
i D1
2 C E4
0
n X i D1
We set .x1 ; : : : ; xn / WD
0
n X
! ai Xsi
n X
3 aj .Xtj Xsj /5 : (56)
j D1
! ai xi . Thus, 2 En . Let j be fixed and
i D1
assume that:
0 s1 < : : : < sj < : : : < sj Cr < tj < sj CrC1 < : : : < sn : We write: .Xs1 ; : : : ; Xsn /.Xtj Xsj / D.Xs1 ; : : : ; Xsn /.Xtj Xsj Cr C Xsj Cr : : : C Xsj C1 Xsj / D.Xs1 ; : : : ; Xsn /.Xtj Xsj Cr / C
r1 X
.Xs1 ; : : : ; Xsn /.Xsj CkC1 Xsj Ck /
kD0
and we study the expectation of each term separately. From condition (i), we obtain by iteration: E .Xs1 ; : : : ; Xsn /.Xtj Xsj Cr / DE e .Xs1 ; : : : ; Xsj Cr ; Xtj /.Xtj Xsj Cr / E e .Xs1 ; : : : ; Xsj Cr ; Xsj Cr /.Xtj Xsj Cr / 0 from condition (ii). The other terms can be dealt with in the same way. t u
314
A.-M. Bogso et al.
Remark 10. R t 1. Note that, in general, the process 1t 0 Xu d u; t 0 may be a peacock even if .Xt ; t 0/ is not a peacock. For example, this is the case for the process .Xt D e t G; t 0/ where Gaussian r.v. Similarly, .Xu ; u 0/ RG is a centered 1 t may be a peacock while t 0 Xu d u; t 0 is not; for example, take the process .Xt D .1Œ0;1 .t/ 11;C1Œ .t//G; t 0/ where G is a centered Gaussian r.v. 2. Theorem 7 answers partially a question raised in [9], namely, for which martingales does (54) hold? Remark 11. In this paper, our aim has been to give several examples of peacocks. On the other hand, we did not exhibit associated martingales (see Point (b)) of the introduction). We refer the interested reader to [10] where numerous martingales associated to given peacocks are presented. However, for most of the peacocks presented in this paper, we do not know how to exhibit an associated martingale. Acknowledgements We are grateful to F. Hirsch and M. Yor for numerous fruitful discussions during the preparation of this work.
References 1. An, M.Y., Log-concave probability distributions: Theory and statistical testing. Duke University Dept of Economics, Working Paper No. 95-03. SSRN, pp. i–29 (1997) 2. D. Baker, M. Yor, A Brownian sheet martingale with the same marginals as the arithmetic average of geometric Brownian motion. Elect. J. Prob. 14(52), 1532–1540 (2009) 3. P. Carr, C.-O. Ewald, Y. Xiao, On the qualitative effect of volatility and duration on prices of Asian options. Finance Res. Lett. 5(3), 162–171 (2008) 4. H. Daduna, R. Szekli, A queueing theoretical proof of increasing property of Pólya frequency functions. Stat. Probab. Lett. 26(3), 233–242 (1996) 5. B. Efron, Increasing properties of Pólya frequency functions. Ann. Math. Stat. 36, 272–279 (1965) 6. U.G. Haussmann, É. Pardoux, Time reversal of diffusions. Ann. Probab. 14(4), 1188–1205 (1986) 7. F. Hirsch, B. Roynette, M. Yor, From an Itô type calculus for Gaussian processes to integrals 992 of log-normal processes increasing in the convex order. J. Math. Soc. Japan, 63(3), 887–891 (2011) 8. F. Hirsch, B. Roynette, M. Yor. Unifying constructions of martingales associated with processes increasing in the convex order, via Lï¿ 12 vy and Sato sheets. Expo. Math. 28(4), 299–324 (2010) 9. F. Hirsch, B. Roynette, M. Yor, Applying Itô’s motto: “Look at the infinite dimensional picture” by constructing sheets to obtain processes increasing in the convex order. Period. Math. Hungar. 61(1), 195–211 (2010) 10. F. Hirsch, C. Profeta, B. Roynette, M. Yor. Peacocks and associated martingales, with explicit constructions. Bocconi & Springer Series, 3. Springer, Milan; Bocconi University Press, Milan, 2011. xxxii+384 pp 11. N. Ikeda, S. Watanabe, in Stochastic Differential Equations and Diffusion Processes. NorthHolland Mathematical Library, vol. 24, 2nd edn. (North-Holland, Amsterdam, 1989) 12. T. Jeulin, M. Yor, in Inégalité de Hardy, semimartingales, et faux-amis. Séminaire de Probabilités, XIII (University of Strasbourg, Strasbourg, 1977/78). Lecture Notes in Math., vol. 721, pp. 332–359 (Springer, Berlin, 1979)
Some Examples of Peacocks in a Markovian Set-Up
315
13. H.G. Kellerer, Markov-Komposition und eine Anwendung auf Martingale. Math. Ann. 198, 99–122 (1972) 14. N.N. Lebedev, Special Functions and Their Applications. Revised English edition (trans. and ed. by R.A. Silverman) (Prentice-Hall, Englewood Cliffs, 1965) 15. J.-F. Le Gall, in Spatial Branching Processes, Random Snakes and Partial Differential Equations. Lectures in Mathematics ETH Zürich (Birkhäuser, Basel, 1999) 16. R. Mansuy, M. Yor, in Random Times and Enlargements of Filtrations in a Brownian Setting. Lecture Notes in Mathematics, vol. 1873 (Springer, Berlin, 2006) 17. P.-A. Meyer, in Sur une transformation du mouvement brownien dûe à Jeulin et Yor. Séminaire de Probabilités, XXVIII. Lecture Notes in Mathematics, vol. 1583, pp. 98–101 (Springer, Berlin, 1994) 18. A. Millet, D. Nualart, M. Sanz, Integration by parts and time reversal for diffusion processes. Ann. Probab. 17(1), 208–238 (1989) 19. D. Revuz, M. Yor, in Continuous Martingales and Brownian Motion. Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences], vol. 293, 3rd edn. (Springer, Berlin, 1999) 20. M. Shaked, J.G. Shanthikumar, in Stochastic Orders and Their Applications. Probability and Mathematical Statistics (Academic, Boston, 1994) 21. M. Shaked, J.G. Shanthikumar, in Stochastic Orders. Springer Series in Statistics (Springer, New York, 2007) 22. J.G. Shanthikumar, On stochastic comparison of random vectors. J. Appl. Probab. 24(1), 123–136 (1987) 23. T. Shiga, S. Watanabe, Bessel diffusions as a one-parameter family of diffusion processes. Z. Wahrscheinlichkeitstheorie und Verw. Gebiete 27, 37–46 (1973) 24. S. Watanabe, On time inversion of one-dimensional diffusion processes. Z. Wahrscheinlichkeitstheorie und Verw. Gebiete 31, 115–124 (1974/75)
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks Antoine-Marie Bogso, Christophe Profeta, and Bernard Roynette
Abstract This paper contains two parts: Part I. Let .Vt ; t 0/ be an integrable right-continuous process such that E ŒjVt j < 1, for every t 0. Let us consider the three types of processes: 0/, 1. .C t WD Vt E ŒVt ; t Vt ; t 0 , with E ŒVt > 0 for every t 0, 2. Nt WD E ŒVt Vt ; t 0 , where EŒVt D 0 for every t 0 and, ˛ W RC ! RC is a 3. Qt WD ˛.t/ Borel function which is strictly positive. We shall give some classes of processes .Vt ; t 0/ such that C , N or Q are peacocks, i.e.: whose one-dimensional marginals are increasing in the convex order. Part II. We introduce the notions of strong and very strong peacocks which lead to the study of new classes of processes. Keywords Peacocks • Conditionally monotone processes • Strong peacocks • Very strong peacocks
AMS Classification: 60J25, 32F17, 60G44, 60E15
A.-M. Bogso () C. Profeta B. Roynette Institut Elie Cartan, Université Henri Poincaré, B.P. 239, 54506 Vandœuvre-lès-Nancy Cedex, France e-mail: [email protected]; [email protected]; [email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__16, © Springer-Verlag Berlin Heidelberg 2012
317
318
A.-M. Bogso et al.
1 Introduction This article deals with processes which increase in the convex order. The investigation of this family of processes has gained renewed interest since the work of Carr, Ewald and Xiao. Indeed, they showed that in the Black-Scholes model, the price of an arithmetic average Asian call option increaseswith s ; s 0/ Z maturity, i.e.: if .B 1 t Bs s 2 ds; t 0 e increases is a Brownian motion issued from 0, the process t 0 in the convex order. Since, many classes of processes which increase in the convex order have been described and studied (see e.g., [11]). The aim of this paper is to complete the known results by exhibiting new families of processes which increase in the convex order. Let us start with some elementary definitions and results. Definition 1. Let U and V be two real-valued r.v.’s. U is said to be dominated by V for the convex order if, for every convex function W R ! R such that EŒj .U /j < 1 and EŒj .V /j < 1, one has: EŒ .U / EŒ .V /:
(1)
We denote this order relation by: (c)
U V:
(2)
We refer to Shaked and Shanthikumar ([18] and [19]) for background on stochastics orders. Definition 2. We denote by C the class of convex C 2 -functions W R ! R such that 00 has a compact support, and by CC the class of convex functions 2 C such that is positive and increasing. We note that if
2 C:
0
• j j is a bounded function, • There exist k1 and k2 0 such that: j .x/j k1 C k2 jxj:
(3)
The next result is proved in [11]. Lemma 1. Let U and V be two integrable real-valued r.v.’s. Then, the following properties are equivalent: (c)
1. U V 2. For every 2 C: EŒ .U / EŒ .V / 3. EŒU D EŒV and for every 2 CC : EŒ .U / EŒ .V /. Definition 3. 1. A process .Zt ; t 0/ is said to be integrable if, for every t 0, EŒjZt j < 1.
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
319
2. A process .Zt ; t 0/ is said to be increasing (resp. decreasing) in the convex (c)
(c)
order if, for every s t, Zs Zt (resp. Zt Zs ). 3. An integrable process which is increasing (resp decreasing) in the convex order will be called a peacock (resp. a peadock). Note: Most of this paper is devoted to the study of peacocks. However, some statements also feature peadock. To draw the reader’s attention, we shall always underline the d. If .Zt ; t 0/ is a peacock, then it follows from Definitions 1 and 3, applied with .x/ D x and .x/ D x, that EŒZt does not depend on t. In the sequel, when two processes .Ut ; t 0/ and .Vt ; t 0/ have the same 1-dimensional marginals, we shall write (1.d)
U t D Vt
(4)
and say that .Ut ; t 0/ and .Vt ; t 0/ are associated. From Jensen’s inequality, every martingale .Mt ; t 0/ is a peacock; conversely, a result due to Kellerer [13] states that, for any peacock .Zt ; t 0/, there exists (at least) one martingale .Mt ; t 0/ such that: (1.d)
Zt D Mt :
(5)
We say that such a martingale is associated with Z. Note that: (a) In general, there exist several different martingales associated with a given peacock. (b) Sometimes, one may be fortunate enough to find directly a martingale associated to the process Z, thus proving that Z is a peacock, see e.g., Example 2. Many examples of peacocks with a description of associated martingales are given in [11]. One may also refer to [2, 9, 10] where the notions of Brownian and Lévy Sheet play an essential role in constructing associated martingales to certain peacocks. On the contrary, we note that for most of the peacocks given in this article, the question of finding an associated martingale remains open. A table summing-up our main results is found at the end of the paper, see Table 1.
2 Part I: Peacocks Obtained by Normalisation, Centering and Quotient 2.1 Preliminaries 2.1.1 The Aim of This Part Let .Vt ; t 0/ be an integrable right-continuous process and, let us consider the three families of processes:
320
A.-M. Bogso et al.
1. .C 0/, t WD Vt E ŒVt ; t Vt ; t 0 , with 0 < E ŒVt < C1 for every t 0, 2. Nt WD E ŒVt Vt ; t 0 , where EŒVt D 0 for every t 0 and ˛ W RC ! RC is a 3. Qt WD ˛.t/ Borel function which is strictly positive. We adopt the notation C for centering, N for normalisation and Q for quotient. We note that, for every t 0, E ŒCt D E ŒQt D 0 and E ŒNt D 1. Since E ŒCt , E ŒNt and E ŒQt do not depend on t, it may be natural to ask under which conditions on .Vt ; t 0/ the processes C , N and Q are peacocks. Let us first recall the following elementary lemma (see [11]). Lemma 2. Let U be a real-valued integrable random variable. Then, the following properties are equivalent: 1. For every real c, E 1fU cg U 0, 2. For every bounded and increasing function h W R ! RC : EŒh.U /U 0; 3. EŒU 0.
2.1.2 Some Examples We now deal with some examples. Z t s Example 1. i. If Vt D e Bs 2 ds, then Carr et al. [5] showed that: 0
Z 1 t Bs s 2 e ds; t 0 is a peacock, Nt WD t 0 when .Bs ; s 0/ is a Brownian motion issued from 0. We note that EŒVt D t for every t 0. An associated martingale is provided in [2]. This example, called the “guiding example” in [11] has been at the origin of many generalizations, as will be clear Z t from now on. Z t ii. If Vt D Ms d˛.s/ (resp. Vt D .Ms M0 /d˛.s/), where .Ms ; s 0/ is 0
0
1 and ˛ W RC ! RC , a continuous and increasing function a martingale in Hloc such that ˛.0/ D 0, then it is shown in [11], Chap. 1, that:
Nt WD
1 ˛.t/
Z
t
Ms d˛.s/; t 0 0
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
(resp.
321
Z t Ct WD .Ms M0 / d˛.s/; t 0 / 0
is a peacock. Note that, for every t 0, EŒVt D ˛.t/EŒN0 (resp. EŒVt D 0). In that generality, we do not know how to associate a martingale to these peacocks. We shall generalize this result in Theorem 11 thanks to the notion of very strong peacock. Example 2. i. If Vt D tX , where X is an integrable centered real-valued r.v., then .Ct WD tX; t 0/ is a peacock (see [11], Chap. 1). ii. If Vt D e tX , where X is a real-valued r.v. such that, for every t 0, EŒe tX < 1, then, it is proved in [11] (see also Example 4) that: Nt WD
e tX ;t 0 EŒe tX
is a peacock.
In particular, if .Gu ; u 0/ is a measurable centered Gaussian process with covariance function K.s; t/ WD EŒGs Gt , and if is a positive Radon measure on RC , then: Z
0 B ./ BNt WD @
t
1
Gu .d u/ exp C Z0 t ; t 0C A is a peacock E exp Gu .d u/ 0
as soon as Z tZ
t
t 7! .t/ WD
K.u; v/.d u/.d v/ is an increasing function. 0
Indeed,
0
Z
t
(law)
Gu .d u/ D
p .t/G;
0
with G a reduced r.v. and, if .B t ; t 0/ is a Brownian motion issued normal .t / from 0, then Mt WD exp B.t / 2 ; t 0 is a martingale associated to ./ Nt ; t 0 . Example 3. Let .Xt ; t 0/ be an integrable centered process and ˛; ˇ W RC ! RC Xt .˛/ ; t 0 is be two strictly positive Borel functions. We suppose that Qt WD ˛.t/ a peacock.
322
A.-M. Bogso et al.
(a) If t 7!
˛.t/ is increasing, then: ˇ.t/ Xt .ˇ/ Qt WD ; t 0 is a peacock. ˇ.t/
Xt .ˇ/ ; t 0 is a peadock and if, for every t 0, Xt is not Qt WD ˇ.t/ identically 0, then ˛.t/ is decreasing. t 7! ˇ.t/
(b) If
Proof (of Example 3). (a) For every
2 C and every 0 s < t, we have:
Xt ˛.t/ Xs ˛.t/ Xt DE E E ˇ.t/ ˛.t/ ˇ.t/ ˛.s/ ˇ.t/ Xt .˛/ (since Qt WD ; t 0 is a peacock) ˛.t/ Xs ˛.s/ E ˛.s/ ˇ.s/
(from point (i)) of Example 2 since Xs is centered and t7! DE
Xs ˇ.s/
˛.t/ is increasing). ˇ.t/
(b) Let us suppose that there exist 0 s < t such that: ˛.t/ ˛.s/ < ; ˇ.s/ ˇ.t/ then,
E
and we may choose E
Xs ˛.s/ ˛.s/ ˇ.s/
Xs ˇ.s/
DE
Xs ˛.s/ ˛.s/ ˇ.s/
such that:
<E
Xs ˛.t/ ˛.s/ ˇ.t/
(from point (i)) of Example 2 since Xs is centered and not identically 0)
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
323
Hence, E
Xs ˇ.s/
<E E
Xs ˛.t/ ˛.s/ ˇ.t/ Xt ˛.t/ ˛.t/ ˇ.t/
(since Q.˛/ is a peacock).
This contradicts the fact that Q.ˇ/ is a peadock.
t u
Example 4. Let .Xt ; t 0/ be a measurable RC -valued process such that 0 < EŒXt < 1: (a) If, for every 0 s t, x 7!
Xt ;t 0 EŒXt
(b) If, for every 0 s t, x 7!
1 EŒXt jXs D x is an increasing function, then: x is a peacock.
1 EŒXs jXt D x is an increasing function, then: x
Xt ;t 0 EŒXt
is a peadock.
Proof (of Example 4). (a) For every
2 CC and every 0 s < t, we have:
Xs Xt E E EŒXt EŒXs Xt Xs Xs 0 E (since is convex) EŒXs EŒXt EŒXs Xs EŒXs EŒXt jXs Xs 0 DE 1 0 EŒXs EŒXs EŒXt Xs
(since x 7!
1 EŒXt jXs D x is increasing and using a slight x
extension of Lemma 2)
324
A.-M. Bogso et al.
(b) For every
2 CC and every 0 s < t, one has:
Xs Xt E EŒXt EŒXs Xt Xs Xt 0 E (since is convex) EŒXt EŒXt EŒXs Xt EŒXt EŒXs jXt Xt 0 DE 1 0 EŒXt EŒXt EŒXs Xt
E
(since x 7!
1 EŒXs jXt D x is increasing and using a slight x
extension of Lemma 2) t u In particular, i. Let W RC R ! RC such that: • For every t 0, x 7! .t; x/ is increasing on R, .t; x/ is increasing. • For every s < t, x 7! .s; x/ Then, if X is a r.v. such that EŒ.t; X / < 1 for every t 0, Nt WD
.t; X / ;t 0 EŒ.t; X /
is a peacockI
ii. Let f W RC ! RC be an increasing C 1 -function such that x 7! decreasing. If .t ; t 0/ denotes the Gamma subordinator, then Nt WD
f .t / ;t 0 EŒf .t /
xf 0 .x/ is f .x/
is a peadock.
This assertion follows from point b) above and from a well-known property of theGamma subordinator: for every t 0, the (Dirichlet) process s ; 0 s t is independent of the r.v. t . t Example 5. Let .Xt ; t 0/ be a càdlàg RC -valued process such that " E
# sup jXu j < 1 0ut
and let be a positive Radon measure on RC . Suppose that, for every 0 s t, 1 's;t W x 7! EŒXs jXt D x is an increasing function. Then: x
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
Z
0
1
t
Xu .du/
B BNt WD @
Z0
t
E
325
Xu .du/
C ; t 0C A is a peadock.
(6)
0
Proof (of (6)). By approximation, we may assume that is absolutely continuous with respect to the Lebesgue measure and admits a continuous density f : .dt/ D f .t/dt: Z
t
Then, with h.t/ WD E
Xu .d u/ and
(7)
2 C:
0
d E dt
"
Rt 0
Xu f .u/d u h.t/
!#
! # Z Xu f .u/d u Xt f .t/ h0 .t/ t 2 Xu f .u/d u DE h.t/ h.t/ h .t/ 0 Z h0 .t/ t 1 0 Xt f .t/ Xt f .t/ E Xu f .u/d u h.t/ h0 .t/ h.t/ 0 "
Rt
0
since
0
0
is increasing and, if
h0 .t/ Xt f .t/ 2 h.t/ h .t/ then Xt f .t/ h0 .t/
Z
t 0
Rt 0
h0 .t/ Xt f .t/ 2 Xu f .u/d u .resp. h.t/ h .t/ Xu f .u/d u h.t/
Xt f .t/ .resp. 0 h .t/
Rt 0
Z
t
Xu f .u/d u/; 0
Xu f .u/d u /: h.t/
Thus, "
!# Xu f .u/d u h.t/ Z h0 .t/ t 1 0 Xt f .t/ E Xt f .t/ Xu f .u/d u h.t/ h0 .t/ h.t/ 0 Z 1 h0 .t/ t 0 Xt f .t/ E Xt f .t/ D EŒXu jXt f .u/d u h.t/ h0 .t/ h.t/ 0
d E dt
Rt 0
326
A.-M. Bogso et al.
1 E D h.t/
0
Z h0 .t/ t Xt f .t/ Xt f .t/ 'u;t .Xt /f .u/d u h0 .t/ h.t/ 0
(where 'u;t is an increasing function) 0
t u
(by Lemma 2).
In particular, if f W RC ! RC is an increasing C 1 -function such that x 7! is increasing and if .t ; t 0/ stands for the Gamma subordinator, then: Z
0
1
t
f .s / ds
B BNt WD @
Z0
t
E
xf 0 .x/ f .x/
f .s / ds
C ; t 0C A is a peadock.
0
One may compare this result with the second point of Remark 7. Example 6. i. Let L WD .Lt ; t 0/ be an integrable Lévy process and be a positive Radon measure on RC . Then: Z t 1 ./ (a) If L is centered, then Qt WD Lu .d u/; t 0 is a .0; t/ 0 peacock Z t 1 0 L .d u/ u C B ./ 0 ; t 0C (b) B A is a peadock. @Nt WD Z t u .d u/ 0
Proof (of Example 6). The assertion (a) is deduced from point (ii) of Example 1 since .Lt ; t 0/ is a centered martingale. To prove (b), we shall make some computations which are close to those in the proof of Example 5 (although here, L does not take values in RC ). We may suppose, without loss of generality, that L is centered and, as in the proof Zof Example 5, that .d u/ D f .u/d u, where f t
is continuous. Let us set h.t/ WD
uf .u/d u. Then, for every
2 C:
0
d E dt
"
Rt 0
Lu f .u/d u h.t/
!#
! # Z Lu f .u/d u Lt f .t/ tf .t/ t 2 Lu f .u/d u DE h.t/ h.t/ h .t/ 0 Z t 1 tf .t/ Lt 0 Lt E Lu f .u/d u h.t/ t t h.t/ 0 "
Rt
0
0
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
tf .t/ E D h.t/
0
Lt t
1 Lt t h.t/
Z
t
E 0
Lu jFtC
327
f .u/d u
(where FtC D .Lu ; u t/).
Lt Observing that ; t 0 is an inverse martingale with respect to the filtration t ˇ
C Lt Ls ˇˇ C F , see [12]), we D Ft ; t 0 (i.e., for every 0 < s t, E t ˇ s t obtain: "
!# Lu f .u/d u h.t/ Z t Lt 1 tf .t/ 0 Lt 1 E uf .u/d u D 0: h.t/ t t h.t/ 0
d E dt
Rt 0
t u
ii. Therefore, the following question arises naturally: for which values of ˛ Z 1 t Lu d u; t 0 is either a peacock (it is true for ˛ 1) or a peadock t˛ 0 (it is true for ˛ 2) or neither of them? (a) If .LZs ; s 0/ is a Brownian motion (issued from 0), then, by scaling, 1 t 3 Lu d u; t 0 is a peacock (resp. a peadock) if and only if ˛ t˛ 0 2 3 (resp. ˛ ). 2 (b) Let .Ls ; s 0/ be a Lévy which is stable of index (1 < 2) Zprocess 1 t and symmetric. Then ˛ Lu d u; t 0 is a peacock (resp. a peadock) t 0 1 1 (resp. ˛ 1 C ). Indeed, by scaling, if and only if ˛ 1 C Z t
(1.d)
1
Lu d u D t 1C S , where the r.v. S is symmetric and stable of index
0
(see point (i) of Example 2). (c) Let .Ls ; s 0/ be a square integrable and centered Lévy process. We have " E
1 t˛
Z
2 #
t
Lu d u 0
D
EŒL21 32˛ t : 3
328
A.-M. Bogso et al.
Hence:
Z t 3 1 , ˛ Lu d u; t 0 is not a peadock, 2 t Z0 1 t 3 Lu d u; t 0 is not a peacock. • If ˛ > , ˛ 2 t 0
• If ˛ <
In some specific situations, we may obtain simultaneously a peacock with one of its associated martingale. The results in the following example are close to those obtained in [11]. Therefore, we state them without proof and refer the reader to [11, Chap. 2]. Example 7. Let .Lt ; t 0/ be a Lévy process such that: Z t E exp Ls ds < 1; for every t 0: 0
Then: Z t 1 Ls ds exp B C BNt WD Z0 t ; t 0C @ A is a peacock E exp Ls ds
1.
0
0
and
Z
1 s dL exp s C B BMt WD ; t 0C Z0 t A @ E exp s dLs 0
t
0
is a martingale associated to .Nt ; t 0/. 2. Z t 0 1 1 L ds exp s B C t 0 BN e t WD ; t 0C Z t @ A is a peacock 1 E exp Ls ds t 0 and
0
Z
B Bf BM t WD @
1
exp 0 Z E exp
.L/ Wu;t 1
du
.L/
Wu;t d u
1
C C ; t 0C A
0 .L/ e t ; t 0/, where .Wu;t 0/-martingale associated to .N ; u 0; t 0/ is the Lévy sheet associated to .Lt ; t 0/ and .L/ is a .Gt ; t
.L/
Gt
.L/ D Wu;s ; u 0; 0 s t (see [8]).
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
329
2.1.3 Relation Between the Peacock Properties of C and N The peacock properties of C and N are linked as is shown in the following: Theorem 1. Suppose that .Vt ; t 0/ satisfies 0 < EŒVt < C1 for every t 0 and t 7! EŒVt is monotone. 1. If t 7! EŒVt increases, we have the implication: .Nt ; t 0/ is a peacock ) .Ct ; t 0/ is a peacock. 2. If t 7! EŒVt decreases, we have the reverse implication: .Ct ; t 0/ is a peacock ) .Nt ; t 0/ is a peacock. We shall give two proofs of Theorem 1. In the first proof, we use Kellerer’s theorem which is not necessary in the second one.
First Proof of Theorem 1 1. We first assume that t 7! EŒVt is an increasing function and that .Nt WD Vt ; t 0 is a peacock. Then, from Kellerer’s theorem, there exists a EŒVt martingale .Mt ; t 0/ such that: Vt (1.d) D Mt ; EŒVt
(1.d)
or, equivalently, Vt D Mt EŒVt :
We note that EŒMt D 1 for every t 0. For every 2 CC and every 0 < s t, one has: EŒ .Ct / EŒ .Cs / D EŒ .Vt EŒVt / EŒ .Vs EŒVs / D EŒ ..Mt 1/EŒVt / EŒ ..Ms 1/EŒVs / E 0 ..Ms 1/EŒVs / ..Mt 1/EŒVt .Ms 1/EŒVs / (by convexity) D E 0 ..Ms 1/EŒVs / ..Ms 1/EŒVt .Ms 1/EŒVs / (taking the conditional expectation) D E 0 ..Ms 1/EŒVs / .Ms 1/ .EŒVt EŒVs / „ ƒ‚ … 0
0
.0/.EŒVt EŒVs /EŒMs 1 D 0 (since
0
is increasing).
330
A.-M. Bogso et al.
2. We now assume that t 7! EŒVt is a decreasing function and that .Ct WD Vt EŒVt ; t 0/ is a peacock. From Kellerer’s theorem, there exists a martingale .Mt ; t 0/ such that: (1.d)
Vt EŒVt D Mt
(1.d)
or, equivalently, Vt D Mt C EŒVt :
We note that EŒMt D 0 for every t 0. Let Vt EŒVt
2 CC and 0 < s t:
Vs E EŒ .Nt / EŒ .Ns / D E EŒVs Mt Ms DE C1 E C1 EŒVt EŒVs Mt Ms Ms E 0 (by convexity) C1 EŒVs EŒVt EŒVs 1 Ms 1 DE 0 C 1 Ms EŒVs EŒVt EŒVs „ ƒ‚ …
0
(taking the conditional expectation) 1 1 EŒMs D 0 0 .1/ EŒVt EŒVs (since
0
is increasing).
t u
Second Proof of Theorem 1 For every t 0, we set ˛.t/ D EŒVt . Then, for every convex function 2 C, we have: 1 Ct .Vt ˛.t// C 1 D E e ; (8) EŒ .Nt / D E ˛.t/ ˛.t/ where e.x/ WD
.x C 1/.
1. To prove the first point, we assume without loss of generality that e.0/ D e0 .0/ D 0
(9)
Let us assume that .Nt ; t 0/ is a peacock. Then, for every 0 < s t, (8) implies that:
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
331
Cs Ct Cs e e e D E Œ .Nt / E Œ .Ns / D E E E ˛.t/ ˛.s/ ˛.t/ (10) 1 1 and, from (9), e increases on Œ0; C1Œ and decreases on ˛.t/ ˛.s/ 1; 0. Hence, since
Cs Ct e e E ; E ˛.t/ ˛.t/ x stands for any convex function. ˛.t/ 2. The second point follows from (8). Indeed, if .Ct ; t 0/ is a peacock, then for every convex function e 2 C such that e.0/ D e0 .0/ D 0 and every 0 < s t, we have: Ct E Œ .Nt / D E e ˛.t/ Cs e E (since .Ct ; t 0/ is a peacock) ˛.t/ 1 1 Cs (since , Cs E e ˛.s/ ˛.t/ ˛.s/ where x 7 !e
is centered and e.0/ D e0 .0/ D 0/ D E Œ .Ns / :
t u
Illustration of Theorem 1 The next example shows that N may be a peacock while C is not. Example 8. LetX be a real-valued random variable and let be the law of X . tX < 1 for every t 0 and supp D R. Let ˛.t/ WD Suppose tX that E jX je E e . Then: e tX 1. Nt WD ; t 0 is a peacock. ˛.t/
2. Ct WD e tX ˛.t/; t 0 is a peacock if and only if ˛ is increasing. We note that, if EŒX > 0, then, from Lemma 2, ˛ 0 .t/ D EŒXe tX 0:
(11)
332
A.-M. Bogso et al.
Proof (of Example 8). The first point is a particular case of Example 4. To prove the second point, we note that @ 0 D EŒCt D E Xe tX ˛ 0 .t/ ; for every t 0 (12) @t and, for every convex function 2 CC : @ EŒ .Ct / D EŒ @t
0
.e tX ˛.t//.Xe tX ˛ 0 .t//:
(13)
i. Let us suppose, on one hand, that ˛ is increasing. The function ft W x 7! xe tx ˛ 0 .t/ has exactly one zero a 0 and ft .x/ > 0; for every x > a: see Fig. 1: Indeed, the derivative function ft 0 of ft is strictly positive on Œ0; 1Œ; hence, ft is a continuous and strictly increasing function, i.e., a bijection map from Œ0; 1Œ to Œ˛ 0 .t/; 1Œ, 0 2 Œ˛ 0 .t/; 1Œ since ˛ 0 .t/ 0 for every t 0 and, ft 1 .0/ D a; moreover, ft .x/ < 0 for every x < 0; therefore, distinguishing the cases X a and X a, we have: @ EŒ .Ct / @t
0
ta e ˛.t/ E Xe tX ˛ 0 .t/ D 0:
Then, Ct WD e tX ˛.t/; t 0 is a peacock if ˛ increases. ii. On the other hand, if ˛ is not increasing, then there exists t0 > 0 such that ˛ 0 .t0 /<0. The function ft0 W x 7! xe t0 x ˛ 0 .t0 / has exactly two zeros a1 < a2 < 0 and 8 < strictly positive if x < a1 ; ft0 .x/ is strictly negative if a1 < x < a2 ; see Fig. 2 : strictly positive if x > a2 :
1
− t
a 0 −
Fig. 1 Graph of ft when ˛ is strictly increasing and t > 0
− ® (t) 1
− ft ( − t )
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks 1
Fig. 2 Graph of ft0
−t
0
a1
333
− − α (t0) a2 0 −
ft0 (−
1 t0
)
Denoting by the law of X , we then observe that: Z
a1 1
and, (12) implies:
Z
ft0 .x/.dx/ > 0
(14)
ft0 .x/.dx/ D 0:
(15)
1 1
From (14) and (15), we deduce that:
E 1fX >a1 g Xe t0 X ˛ 0 .t0 / D
Z
1
ft0 .x/.dx/ < 0 (since supp D R/: a1
t u e tx ˛.t/ D 1Œa1 ;1Œ .x/ in (13). tX e ; t 0 is a peacock, then, by Let us note that, if ˛ is increasing and ˛.t/
Theorem 1, Ct WD e tX ˛.t/; t 0 is a peacock. This provides another proof of point (i) of the preceding proof. Then, the result follows by taking
0
3 Peacocks Obtained from Conditionally Monotone Processes 3.1 Definition of Conditionally Monotone Processes and Examples Let us first introduce the notion of conditional monotonicity, which appears in [17, Chap. 4.B, pp. 114–126] and is studied in [3]. Definition 4 (Conditional monotonicity). A process .X ; 0/ is said to be conditionally monotone if, for every n 2 N , every i 2 f1; : : : ; ng, every 0 < 1 < < n and every bounded Borel function W Rn ! R which increases (resp. decreases) with respect to each of its arguments, we have: EŒ.X1 ; X2 ; : : : ; Xn /jXi D i .Xi /;
(16)
334
A.-M. Bogso et al.
where i W R ! R is a bounded increasing (resp. decreasing) function. To prove that a process is conditionally monotone, we may restrict ourselves to bounded Borel functions which are increasing with respect to each of their arguments. Indeed, replacing by , the result then also holds for bounded Borel functions which are decreasing with respect to each of their arguments. Definition 5. We denote by En the set of bounded Borel functions W Rn ! R which are increasing with respect to each of their arguments. Remark 1. 1. Note that .X ; 0/ is conditionally monotone if and only if .X ; 0/ is conditionally monotone. 2. Let W R ! R be a strictly monotone and continuous function. It is not difficult to see that if the process .X ; 0/ is conditionally monotone, then so is . .X /; 0/. In [3], the authors exhibited a number of examples of processes enjoying the conditional monotonicity (16) property. Among them are: i. The processes with independent and log-concave increments, ii. The Gamma subordinator, iii. The well-reversible diffusions at a fixed time, such as, for example: • Brownian motion with drift , • Bessel processes of dimension ı 2, • Squared Bessel processes of dimension ı > 0. We refer the reader to [3] and ([11], Chap. 1, Sect. 4) for more details. The next lemma follows immediately from Definition 4. Lemma 3. Let .Xt ; t 0/ be a real-valued right-continuous process which is conditionally monotone and, let q W RC R ! R be a continuous function such that, for every s 0, qs W x 7! q.s; x/ is increasing. Then, for every positive function 2 E1 , every positive Radon measure on RC and every t > 0: Z t ˇ ˇ E q.s; Xs / .ds/ ˇˇ Xt D t .Xt /;
(17)
0
where t is an increasing function.
3.2 Peacocks Obtained by Centering Under a Conditional Monotonicity Hypothesis Theorem 2. Let .Xt ; t 0/ be a real-valued right-continuous process which is conditionally monotone. Let q W RC R ! RC be a positive and continuous function such that, for every s 0, qs W x 7! q.s; x/ is increasing and
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
335
EŒq.s; Xs / > 0. Let W RC ! RC a positive, increasing and convex C 1 -function satisfying: Z t 8 t 0; E q.s; Xs /ds <1 (18) 0
"
and 8 a > 0; Then:
E
sup q.t; Xt /
0
0
Z
#
t
< 1:
q.s; Xs /ds
(19)
0
Z t Ct WD q.s; Xs /ds h.t/; t 0 is a peacock, 0
Z t where h.t/ WD E q.s; Xs /ds . 0
Proof (of Theorem 2). For every convex function 2 CC , we have: Z t d 0 0 0 EŒ .Ct / D E .Ct / q.t; Xt / q.s; Xs /ds h .t/ dt 0 Z t h0 .t/ D E 0 .Ct /q.t; Xt / 0 q.s; Xs /ds EŒq.t; Xt / 0 q.t; Xt / 1 C E 0 .Ct /h0 .t/ EŒq.t; Xt / WD K1 .t/ C K2 .t/: Let us prove that K1 .t/ 0. We note that, for every t 0: Z t 0 E q.t; Xt / q.s; Xs /ds 0
since
h0 .t/ EŒq.t; Xt /
D0
(20)
Z t E q.t; Xt / 0 q.s; Xs /ds D h0 .t/: 0
Then, since 0 is increasing, one has: Z t 0 K1 .t/ D E .Ct /q.t; Xt / 0 q.s; Xs /ds 0
h .t/ h.t/ EŒq.t; Xt / Z t 0 E q.t; Xt / q.s; Xs /ds 0 01 ı
h0 .t/ EŒq.t; Xt /
0
0
h0 .t/ EŒq.t; Xt /
D 0:
336
A.-M. Bogso et al.
Let us prove that K2 .t/ 0. We have:
q.t; Xt / 1 .Ct / K2 .t/ D h .t/E EŒq.t; Xt / q.t; Xt / 0 0 D h .t/E EŒ .Ct /jXt 1 : EŒq.t; Xt /
0
0
But, by Lemma 3, EŒ
0
.Ct /jXt D E
0
Z t ˇ ˇ q.s; Xs /ds h.t/ ˇˇ Xt D 't .Xt /; 0
where 't is an increasing function. Hence, from Lemma 2, q.t; Xt / 1 0 K2 .t/ D h0 .t/E 't .Xt / EŒq.t; Xt / since qt W x 7! q.t; x/ is increasing for every t 0.
t u
Example 9. Suppose .Xt ; t 0/ and q W RRC ! RC be chosen as in Theorem 2. If we take successively .x/ D x and .x/ D e x , we obtain: Z
t 0
Z
t
q.s; Xs /ds E
q.s; Xs /ds ; t 0 is a peacock
(21)
0
and Z t Z t q.s; Xs /ds E exp q.s; Xs /ds ; t 0 is a peacock. (22) exp 0
0
3.3 Peacocks Obtained by Normalisation from a Particular Class of Conditionally Monotone Processes We now consider the class of processes with independent and log-concave increments. (Note that these processes are conditionally monotone (see [3])). Let us recall some definitions and properties, see [1] or [6]. Definition 6 (R-valued log-concave r.v.’s). An R-valued random variable X is said to be log-concave if: 1. X admits a probability density g, 2. The function log g is concave; i.e., the second derivative of log g (in the distribution sense) is a negative measure.
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
337
Definition 7 (Z-valued log-concave r.v.’s). A Z-valued random variable X is said to be log-concave if, with g.n/ D P.X D n/ .n 2 Z/, one has: for every n 2 Z, g 2 .n/ g.n 1/g.n C 1/I in other words, the discrete second derivative of log g is negative. Example 10. Many common density functions on R (or Z) are log-concave. Indeed, the normal density, the uniform density, the exponential density, the Poisson density and the geometric density are log-concave. The following properties of log-concave random variables are well-known (see [16] or [7]). Lemma 4. An R-valued .resp. Z-valued/ random variable X is log-concave if and only if its probability density g satisfies: 1. The support of g is an .finite or infinite/ interval I R .resp. I Z/, 2. For every x2 x1 , y2 y1 , det
g.x1 y1 / g.x1 y2 / g.x2 y1 / g.x2 y2 /
0:
(23)
Lemma 5. ([16]) i. Every log-concave density is bounded. ii. If g and h are two log-concave densities, then their convolution g h given by: Z
1
g h.x/ D
g.y/h.x y/dy 1
is also log-concave, i.e: the sum of two independent log-concave random variables is log-concave. The main result of this section is the following: Theorem 3. Let .Xt ; t 0/ be a right-continuous R-valued process with independent and log-concave increments issued from 0, and ˛ W RC ! RC a right-continuous and increasing function satisfying ˛.0/ D 0. Let q W RC R ! R be a continuous function such that, for every t 0: i. The variable
!
t WD exp ˛.t/ sup q.s; Xs /
is integrable
(24)
0st
and,
t WD E exp ˛.t/ inf q.s; Xs / > 0; 0st
(25)
338
A.-M. Bogso et al.
ii. The function x 7! q.t; x/ is increasing (resp. decreasing). Then,
Z t 1 q.s; X / d˛.s/ exp s B C BNt WD Z0 t ; t 0C @ A is a peacock. E exp q.s; Xs / d˛.s/ 0
0
In particular, if .Bt ; t 0/ denotes a Brownian motion starting from 0, then Z t 0 1 q.s; Bs / d˛.s/ exp B .B/ C BNt WD Z0 t ; t 0C @ A is a peacock. E exp q.s; Bs / d˛.s/ 0
To prove Theorem 3, we need the following lemma. Lemma 6. Let .Xt ; t 0/ be a R-valued process with independent and logconcave increments and .fk W R ! RC ; k 2 N / a family of strictly positive Borel functions such that, for every p 2 N and every 0 t1 < < tp : # " p Y fk .Xtk / < 1: E kD1
Then, for every n 2, every 0 1 < 2 < < n and every bounded Borel function W Rn1 ! RC which increases (resp. decreases) with respect to each of its arguments: ˇ # " n1 ˇ Y ˇ fk .Xk /ˇ Xn D z E .X1 ; : : : ; Xn1 / ˇ kD1 ˇ K.n; z/ D " n1 # ˇ Y ˇ E fk .Xk /ˇ Xn D z ˇ kD1
is an increasing (resp. decreasing) function of z. Proof (of Lemma 6). i. We give the proof of Lemma 6 only in the continuous case (the proof in the discrete one is similar). ii. By truncation and regularisation, it suffices to prove Lemma 6 when the functions fk .k 2 N / are bounded and when the support of all increments’ densities is R. iii. In this proof we deal only, without loss of generality, with bounded Borel functions which increase with respect to each of its arguments. We now prove this lemma by induction on n 2.
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
339
• For n D 2: we denote by g1 (resp. e g 2 ) the density function of X1 (resp. X2 X1 ). For every bounded and increasing Borel function W R ! RC : K.2; z/ D
E Œ.X1 /f1 .X1 /jX2 D z E Œf1 .X1 /jX2 D z Z
D
1
1
Z
.u/f1 .u/g1 .u/e g 2 .z u/ d u :
1
1
f1 .u/g1 .u/e g 2 .z u/ d u
Since is bounded and increasing, then, by approximation, we can restrict ourselves to the case: l X D ci 1Œxi ;1Œ i D1
where l 2 N and, for every i 2 1; l, ci is a positive constant and xi a real number. Hence, the function z 7! K.2; z/ increases if and only if, for every x 2 R, Z
1
f1 .u/g1 .u/e g 2 .z u/ d u z 7! Z x1 1
is increasing. f1 .u/g1 .u/e g 2 .z u/ d u
This is also equivalent to say that: for every x 2 R, Z
1
f1 .u/g1 .u/e g 2 .z u/ d u L.x; z/ D Z xx 1
f1 .u/g1 .u/e g 2 .z u/ d u
is an increasing function of z. Since e g 2 is log-concave, then, for every x 2 R, z 2 R and > 0, we have: e g 2 .z u/ e g 2 .z C u/ ; for every u x e g 2 .z C x/ e g 2 .z x/ and
e g 2 .z C u/ e g 2 .z u/ ; for every u x: e g 2 .z C x/ e g 2 .z x/
Therefore, for every x 2 R, z 2 R and > 0, one has:
340
A.-M. Bogso et al.
Z1 f1 .u/g1 .u/e g 2 .z C u/ d u x L.x; z C / D Z x 1 Z 1
f1 .u/g1 .u/e g 2 .z C u/ d u
e g 2 .z C u/ du e g 2 .z C x/ D Z xx e g 2 .z C u/ du f1 .u/g1 .u/ e g 2 .z C x/ 1 Z 1 e g 2 .z u/ du f1 .u/g1 .u/ e g 2 .z x/ Z xx e g 2 .z u/ du f1 .u/g1 .u/ e g 2 .z x/ 1 f1 .u/g1 .u/
D L.x; z/I which means that L.x; z/ increases with z. Then, z 7! K.2; z/ increases for every bounded and increasing Borel function W R ! RC . • For n 3: we assume that, for every bounded Borel function ' W Rn2 ! RC which increases with respect to each of its arguments, z 7! K.n 1; z/ increases and, we denote by gn1 (resp. e g n ) the density function of Xn1 (resp. Xn Xn1 ). Since the variables Xn Xn1 and Xn1 are independent, then, for every bounded Borel function W Rn1 ! R which increases with respect to each of its arguments, we have: ˇ # ˇ ˇ E .X1 ; : : : ; Xn1 / fk .Xk /ˇ Xn1 C Xn Xn1 D z ˇ kD1 K.n; z/ D ˇ " n1 # ˇ Y ˇ E fk .Xk /ˇ Xn1 C Xn Xn1 D z ˇ "
n1 Y
kD1
Z
1 1
D Z
1
1
" E .X1 ; : : : ; Xn2 ; z y/
n2 Y kD1
E
" n2 Y kD1
ˇ # ˇ ˇ fk .Xk /ˇ Xn1 D z y ˇ
fn1 .zy/gn1 .zy/e gn .y/ dy : ˇ # ˇ ˇ fk .Xk /ˇ Xn1 D z y fn1 .z y/gn1 .z y/e gn .y/ dy ˇ
After the change of variable: x D z y, we obtain:
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
Z
1
K.n; z/ D
"
1
Z
E .X1 ; : : : ; Xn2 ; x/
n2 Y kD1
1
1
E
" n2 Y kD1
For x 2 R, we define:
341
ˇ # ˇ ˇ fk .Xk /ˇ Xn1 D x ˇ
g n .z x/ dx fn1 .x/gn1 .x/e ˇ : # ˇ ˇ fk .Xk /ˇ Xn1 D x fn1 .x/gn1 .x/e g n .z x/ dx ˇ
ˇ # ˇ ˇ E .X1 ; : : : ; Xn2 ; x/ fk .Xk /ˇ Xn1 D x ˇ kD1 m.x/ D ˇ n2 ˇ Q E fk .Xk /ˇˇ Xn1 D x "
n2 Y
kD1
and f .x/ D E
" n2 Y kD1
ˇ # ˇ ˇ fk .Xk /ˇ Xn1 D x : ˇ
Hence, (a) Since W Rn1 ! RC is bounded and increasing with respect to each of its arguments, the induction hypothesis implies that m W R ! RC is a bounded and increasing Borel function, (b) We have: Z
1 1
Z
K.n; z/ D
m.x/f .x/fn1 .x/gn1 .x/e g n .z x/ dx :
1
1
f .x/fn1 .x/gn1 .x/e g n .z x/ dx
Using the log-concavity of gn and the case n D 2 computed above, we have: for every y 2 R, Z
1
f .x/fn1 .x/gn1 .x/e g n .z x/ dx z 7! Z
y y 1
f .x/fn1 .x/gn1 .x/e g n .z x/ dx
is an increasing function of z. Then, the function z 7! K.n; z/ increases for every bounded Borel function W Rn1 ! RC which increases with respect to each of its arguments. t u
342
A.-M. Bogso et al.
Proof (of Theorem 3). We prove this theorem only in the case where x 7! q.; x/ is increasing. Let T > 0 be fixed. 1. We first consider the case 1Œ0;T d˛ D
r X
ai ıi
i D1
where r 2 2; 1, a1 0; a2 0; : : : ; ar 0, 2 < < r T . Let us prove that: Nn WD exp
n X
Pr
!
i D1 ai
D ˛.T / and 0 1 <
!
ai q.i ; Xi / h.n/ ; n 2 1; r
is a peacock,
i D1
where " h.n/ WD log E exp
n X
!# ai q.i ; Xi /
; for every n 2 1; r:
i D1
We have: EŒNn Nn1 D 0; for every n 2 2; r with
Nn Nn1 D Nn1 e an q.n ;Xn /h.n/Ch.n1/ 1 ;
and, for every convex function
2 C,
EŒ .Nn / EŒ .Nn1 / E
0
.Nn1 /Nn1 e an q.n ;Xn /h.n/Ch.n1/ 1
D E EŒNn1 jXn K.n; Xn / e an q.n ;Xn /h.n/Ch.n1/ 1 ; where K.n; z/ D
EŒ
0
.Nn1 /Nn1 jXn D z : EŒNn1 jXn D z
The positive and bounded C 0 -function W Rn1 ! RC given by: !# " n1 X 0 .x1 ; : : : ; xn1 / D ai q.i ; xi / h.n 1/ exp i D1
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
343
increases with respect to each of its arguments. For i 2 N , let us define: fi .x/ D e ai q.i ;x/ ; for every x 2 RI then, for every n 2 2; r, we have: Nn1 D e h.n1/
n1 Y
fk .Xk /:
kD1
Hence,
ˇ # ˇ ˇ E .X1 ; : : : ; Xn1 / fk .Xk /ˇ Xn D z ˇ kD1 K.n; z/ D ˇ : " n1 # ˇ Y ˇ E fk .Xk /ˇ Xn D z ˇ "
n1 Y
kD1
Moreover, for every n 2 1; r, " E
n Y
#
"
fk .Xk / D E exp
kD1
n X
!# ai q.i ; Xi /
kD1
" E exp
sup q.; X / 0T
" E exp
sup q.; X / 0T
"
n X
!# ai
kD1 r X kD1
! ai
# _1
!
#
D E exp ˛.T / sup q.; X / _ 1 0T
D EŒ T _ 1 < 1: Therefore, thanks to Lemma 6, K.n; z/ is an increasing function of z. For 0, we denote by q1 , the right-continuous inverse of x 7! q.; x/. Let us also consider the variable:
V .n; Xn / WD K.n; Xn /EŒNn1 jXn e an q.n ;Xn /h.n/Ch.n1/ 1 : Then: i. If Xn then
q1 n
h.n/ h.n 1/ ; an
344
A.-M. Bogso et al.
e an q.n ;Xn /h.n/Ch.n1/ 1 0 and 1 h.n/ h.n 1/ V .n; Xn / K n; qn an
a q. ;X /h.n/Ch.n1/ 1 ; EŒNn1 jXn e n n n ii. If Xn
q1 n
h.n/ h.n 1/ ; an
then e an q.n ;Xn /h.n/Ch.n1/ 1 0 and h.n/ h.n 1/ V .n; Xn / K n; q1 n an
a q. ;X /h.n/Ch.n1/ 1 : EŒNn1 jXn e n n n Thus, EŒ .Nn / EŒ .Nn1 / EV .n; Xn /
1 h.n/ h.n 1/ E EŒNn1 jXn e an q.n ;Xn /h.n/Ch.n1/ 1 K n; qn an
h.n/ h.n 1/ D K n; q1 E Nn1 e an q.n ;Xn /h.n/Ch.n1/ 1 n an 1 h.n/ h.n 1/ D K n; qn E ŒNn Nn1 D 0: an Hence, for every r 2 2; 1: Nn WD exp
n X
!
!
ai q.i ; Xi / h.n/ ; n 2 1; r
i D1
2. We now set D 1Œ0;T d˛ and, for every 0 t T , Z ./
Nt
D
t
q.u; Xu /.d u/ exp Z0 t : E exp q.u; Xu /.d u/ 0
is a peacock.
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
345
Since the function 2 Œ0; T 7! q.; X / is right-continuous and bounded from above by sup jq.; X /j which is finite a.s., there exists a sequence .n ; n 0/ 0T R R of measures of type considered in (1), with supp n Œ0; T , n .d u/ D .d u/ and, for every 0 t T , Z
t
lim exp
n!1
Z t q.u; Xu /n .d u/ D exp q.u; Xu /.d u/ a.s.
0
(26)
0
Moreover, for every 0 t T and every n 0, Z n0
t
q.u; Xu /n .d u/
sup exp 0
Z exp
sup q.; X /
n .d u/
0T
0
Z D exp
!
T
n .d u/ _ 1
sup q.; X / 0
0T
Z D exp
!
t
T
sup q.; X / 0T
! .d u/ _ 1 D T _ 1
0
which is integrable from (24). Thus, the dominated convergence theorem yields
Z
lim E exp
n!1
t
q.u; Xu /n .d u/
Z
D E exp
0
t
q.u; Xu /.d u/
:
(27)
0
Using (26) and (27), we obtain: .n /
lim Nt
n!1
./
D Nt
a.s., for every 0 t T:
(28)
Now, from (1), . / Nt n ; 0 t T is a peacock for every n 0; i.e., for every 0 s < t T and for every E Then, since
(29)
2 C:
h i . / .Ns.n / / E .Nt n / :
ˇ ˇ ˇ . / ˇ T _ 1 sup sup ˇNt n ˇ ; T ^ 1 0t T n0
(30)
(31)
346
A.-M. Bogso et al.
which is integrable from (24) and (25), it remains to apply the dominated conver./ gence theorem in (30) to obtain that .Nt ; 0 t T / is a peacock for every T > 0. t u
4 Peacocks Obtained from a Diffusion by Centering and Normalisation Let us consider two Borel functions W RC R ! R and b W RC R ! R such that s .x/ WD .s; x/ and bs .x/ WD b.s; x/ are Lipschitz continuous with respect to x, locally uniformly with respect to s and .Xt ; t 0/ a process with values in an interval I R and which solves the SDE: Z t Z t .s; Ys /dBs C b.s; Ys /ds (32) Yt D x0 C 0
0
where x0 2 I and .Bs ; s 0/ denotes a standard Brownian motion started at 0. For s 0, let Ls denotes the second-order differential operator: Ls WD
1 2 @2 @ .s; x/ 2 C b.s; x/ : 2 @x @x
(33)
The following results concern peacocks of C and N types.
4.1 Peacocks Obtained by Normalisation Theorem 4. Let .Xt ; t 0/ be a solution of (32) taking values in I . Let W I !RC be an increasing C 2 -function such that: 1. For every s 0: vs W x 2 I 7!
Ls .x/ is an increasing function .x/
2. The process Z t Mt WD .Xt / .x0 / Ls .Xs /ds; t 0 0
is a martingale.
(34)
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
Nt WD
Then:
.Xt / ;t 0 EŒ .Xt /
347
is a peacock.
(35)
Proof (of Theorem 4). For every t 0, let h.t/ D EŒ .Xt /. We note that h is strictly positive. Let and 0 s < t. By ItOo’s formula
.Xt / h.t/
D
.Xs / h.s/
Z
t
0
C Z
s t
0
s
C
1 2
Z
t s
.Xu / h.u/
dMu Lu .Xu /d u C h.u/ h.u/
2C
.Xu / h0 .u/ .Xu/ du h.u/ h2 .u/ 1 00 .Xu / d hM iu: 2 h.u/ h .u/
Hence, it suffices to see that, for every s u < t: K.u/ WD E We note that:
since u 7!
0
.Xu / h.u/
Lu .Xu / h0 .u/ .Xu / h.u/ h2 .u/
0:
(36)
Lu .Xu / h0 .u/ .Xu / E D0 h.u/ h2 .u/
(37)
1 E Œ .Xu / is constant and h.u/ d E Œ .Xu / D EŒLu .Xu /: du
(38)
Hence, for every s u < t, since, by hypothesis (34), x 7! vu .x/ is increasing, we have: .Xu / .Xu / h0 .u/ vu .Xu / K.u/ D E 0 h.u/ h.u/ h.u/ 0 1 0 h .u/ v1 u 0 B C h.u/ C E .Xu / vu .Xu / h .u/ 0B D 0: @ A h.u/ h.u/ h.u/ t u
348
A.-M. Bogso et al.
4.2 Peacocks Obtained by Centering Theorem 5. Let .Xt ; t 0/ be a solution of (32) taking values in I . Let W I ! RC be an increasing C 2 -function such that: 1. For every s 0, x 7! Ls .x/ is increasing. 2. The process Z t Mt WD .Xt / .x0 / Ls .Xs /ds; t 0 0
is a martingale. Then: .Ct WD .Xt / EŒ .Xt /; t 0/ is a peacock. Proof (of Theorem 5). Let 2 C, h.t/ D EŒ .Xt / and 0 s < t. From ItOo’s formula, we have: . .Xt / h.t// . .Xs / h.s// Z t 0 D . .Xu / h.u//ŒdMu C Lu .Xu /d u h0 .u/d u s
1 C 2
Z
t
00
. .Xu / h.u//d hM iu:
s
Hence, it is sufficient to show that, for every s u < t:
. .Xu / h.u//.Lu .Xu / h0 .u// 0:
(39)
EŒLu .Xu / h0 .u/ D 0; for every s u < t
(40)
E
0
But, (39) follows from:
and E
. .Xu / h.u//.Lu .Xu / h0 .u//
0 .Lu /1 .h0 .u// h.u/ E Lu .Xu / h0 .u/ D 0: 0
t u Remark 2. In Theorem 4, if we suppose furthermore that Ls .x/ 0 for every s 0 and x 2 I , then .Ct WD .Xt / EŒ .Xt /; t 0/ is a peacock.
(41)
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
349
Indeed, for every t 0: Z
t
.Xt / D .x0 / C Mt C
Lu .Xu /d u: 0
Thus, h W t 7! EŒ .Xt / is increasing and the result follows from both Theorems 1 and 4.
4.3 Peacocks Obtained from an Additive Functional by Normalisation Let As be the space-time differential operator given by: As WD
1 @ @2 @ C 2 .s; x/ 2 C b.s; x/ : @s 2 @x @x
(42)
We shall prove the following result: Theorem 6. Let .Xt ; t 0/ be a conditionally monotone process with values in I and which solves (32) and, let q W RC I ! RC be a strictly positive C 2 -function such that, 1. For every s 0, EŒq.s; Xs / > 0, qs W x 2 I 7! q.s; x/ is increasing and fs W x 2 I 7!
As q.s; x/ is an increasing function q.s; x/
(43)
2. The process Z t Zt WD q.t; Xt / q.0; x0 / As q.s; Xs /ds; t 0 0
is a martingale. Then, for every positive Radon measure on RC : Z
0 B BNt WD @
1
t
q.s; Xs /.ds/ Z0
t
E
q.s; Xs /.ds/
C ; t 0C A is a peacock.
0
One may find in [11], Chap. 1, many examples of SDEs solutions which are conditionally monotone. This fact is related to the “well-reversible” property of these diffusions.
350
A.-M. Bogso et al.
Proof (of Theorem 6). We set:
u WD
1 ; for every u 0: EŒq.u; Xu /
For every u 0, ItOo’s formula yields: Z u Z
v dZv C
u q.u; Xu / D 1 C 0
u 0
(44)
0
v q.v; Xv / C v Av q.v; Xv / d v
WD Mu C Hu Z Mu WD 1 C
where
u
v dZv ; u 0
is a martingale
(45)
0
and Z Hu WD
u 0
0
v q.v; Xv / C v Av q.v; Xv / d v; u 0
is a centered process (46)
since EŒ u q.u; Xu / D 1 for every u 0 and d EŒq.u; Xu / D EŒAu q.u; Xu /: du
(47)
Hence, by setting Z
t
1 .d u/ D E
u
h.t/ WD 0
Z
t
q.u; Xu /.d u/ ;
(48)
0
one has: Nt D
1 h.t/
Z
t
q.u; Xu /.d u/ D 0
1 h.t/
1 D h.t/
Z
t
u q.u; Xu / Z
0 t
1 .d u/
u
.Mu C Hu /dh.u/:
0
Thus, integrating by parts, we obtain: dNt D Z
with .h/
Mt
t
D
dh.t/ .h/ .h/ M C H t t h2 .t/ Z .h/
h.u/dMu and Ht 0
D
t
h.s/dHs : 0
(49)
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
351
2 CC and every 0 s < t, we have:
Then, for every
Z
t
0
EŒ .Nt / EŒ .Ns / D E Z DE Z
t
D s
.Nu /dNu
s
dh.u/
.Nu / Mu.h/ C Hu.h/ 2 h .u/ s Z u
.h/ dh.u/ 00 .h/ 2 dh.v/ E .Nv / Mv C Hv h2 .u/ h2 .v/ 0 Z u Z t
dh.u/ 0 .h/ .h/ C E : .Nv / dMv C dHv 2 s h .u/ 0 t
0
Hence, it remains to see that, for every u > 0: Z
u
0
E 0
.Nv /dHv.h/
Z
u
0
DE 0
.Nv /h.v/ v Av q.v; Xv / C v0 q.v; Xv / d v 0
(50)
But, for every 0 v u,
.Nv /h.v/ v Av q.v; Xv / C v0 q.v; Xv /
v0 0 (with fv defined by (43)) E .N /q.v; X / f .X / C D h.v/ v v v v v
v
0 D h.v/ v E EŒ 0 .Nv /jXv q.v; Xv / fv .Xv / C v
v
K.v/ WD E
0
But, by Lemma 3, EŒ
0
.Nv /jXv D E
0
1 h.v/
Z 0
v
ˇ ˇ q.u; Xu /.d u/ ˇˇ Xv D 'v .Xv /;
(51)
where 'v is an increasing function and,
v0 D 0: E q.v; Xv / fv .Xv / C
v
(52)
Then,
0
0 K.v/ h.v/ v 'v fv1 v E q.v; Xv / fv .Xv / C v D 0:
v
v
t u
352
A.-M. Bogso et al.
4.4 Peacocks Obtained by Normalisation from Markov Processes with Independent and Log-Concave Increments Let us give two extensions of Theorem 3. The first one concern random walks and the second result deals with continuous Markov processes. Theorem 7. Let .Xn ; n 2 N/ be a random walk on R with independent and logconcave increments, issued from 0 and let P be its transition kernel. Let W R ! RC be an increasing (resp. decreasing) Borel function such that: Mn WD .Xn /
n1 Y kD0
.Xk / ;n 2 N P .Xk /
! is a martingale.
Let q W R ! R be an increasing (resp. decreasing) Borel function such that x 7! q.x/ C log
P .x/ is increasing (resp. decreasing), .x/
(53)
and for every n 1 E
"
.Xn / C P .Xn /e
q.Xn /
n1 X
exp
!# < 1:
q.Xk /
(54)
kD0
Then, Nn WD .Xn / exp
n1 X
! q.Xk / h.n/ ; n 2 N
! is a peacock,
kD0
n1 P where h.n/ WD log E .Xn / exp q.Xk / . kD0
Proof (of Theorem 7). 1. We only consider the case where the functions , q and x 7! q.x/ C log are increasing. We set A0 D 0, An WD exp
n1 X kD0
! n1 Y P .Xk / ; for every n 1; q.Xk / h.n/ .Xk / kD0
and Fn WD .Xp ; p n/; for every n 0:
P .x/ .x/
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
353
Then, for every n 1, An is Fn1 -measurable and Nn D Mn An . It follows from (54) that, for every n 1, the variables Nn WD Mn An and Mn1 An are integrable. Moreover, since .Nn ; n 0/ has a constant mean and .Mn ; n 0/ is a .Fn /-martingale, we obtain, for n 1: 0 D EŒNn Nn1 D EŒMn An Mn1 An1 D EŒAn .Mn Mn1 / C EŒMn1 .An An1 / D EŒMn1 .An An1 /: But,
An 1 An1
Mn1 .An An1 / D Nn1
P .Xn1 / q.Xn1 /h.n/Ch.n1/ e D Nn1 1 .Xn1 / q.Xn1 / 1 ; D Nn1 ee where, from (53), the function x 7! e q .x/ WD q.x/ C log is increasing. Hence:
2. For every
P .x/ h.n/ C h.n 1/ .x/
h i q.Xn1 / E Nn1 ee 1 D 0
2 C, one has:
EŒ .Nn / EŒ .Nn1 / E 0 .Nn1 /.Nn Nn1 / (by convexity) D E 0 .Nn1 /An .Mn Mn1 / C E 0 .Nn1 /Mn1 .An An1 / (since An is Fn1 -integrable and .Mn ; n 0/ is a martingale) D E 0 .Nn1 /Mn1 .An An1 / i h q.Xn1 / 1 D E 0 .Nn1 /Nn1 ee EŒ 0 .Nn1 /Nn1 jXn1 q.Xn1 / e EŒNn1 jXn1 e 1 : DE EŒNn1 jXn1 Now, for every n 1 and x 2 R, we define: K.n; x/ WD and
EŒ
0
.Nn1 /Nn1 jXn1 D x EŒNn1 jXn1 D x
(55)
354
A.-M. Bogso et al.
q.x/ V .n; x/ WD EŒNn1 jXn1 D x ee 1 : We note that:
8 n 1; EŒV .n; Xn1 / D 0;
(from (55))
and from Lemma 6, the function x 7! K.n; x/ is increasing. Let .e q /1 denotes the right-continuous inverse of e q . Then, distinguishing the cases Xn1 .e q /1 .0/ and Xn1 .e q /1 .0/, we obtain
q /1 .0/ V .n; Xn1 /; 8 n 1; K.n; Xn1 /V .n; Xn1 / K n; .e and finally, EŒ .Nn / EŒ .Nn1 / EŒK.n; Xn1 /V .n; Xn1 /
K n; .e q /1 .0/ EŒV .n; Xn1 / D 0:
t u
We now deal with an extension of Theorem 3 for continuous Markov processes. Theorem 8. Let .Xt ; t 0/ be a continuous (non necessary homogeneous) Markov process with independent and log-concave increments, issued from 0 and let A be the infinitesimal generator of the space-time process associated with X . Let W RC R ! RC be a continuous function such that x 7! .t; x/ is increasing (resp. decreasing) for every t 0 and Z t A .u; Xu / d u ; t 0 Mt WD .t; Xt / exp 0 is a continuous local martingale. C R ! R be a continuous function such that x 7! q.t; x/ and x 7! Let q W R A .t; x/ are increasing (resp. decreasing) for every t 0. We assume qC furthermore that, for every t 0: " ! Z # t
E
sup . C jqj C jA j/.s; Xs / exp 0st
jqj.s; Xs / ds
"
and
(56)
!#
E .t; Xt / exp t sup q.s; Xs /
< 1:
(57)
0st
Then,
<1
0
Z
t
Nt WD .t; Xt / exp
q.s; Xs / ds h.t/ ; t 0
0
Z
t
is a peacock, where h.t/ WD log E .t; Xt / exp
q.s; Xs / ds 0
.
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
355
Proof (of Theorem 8). 1. We start with Lemma 7. Let .Xt ; t 0/ be a right-continuous process with independent and logconcave increments issued from 0. Let q W RC R ! R and W RC R ! RC be two continuous functions such that, for every t 0, x 7! q.t; x/ and x 7! .t; x/ are increasing, and " !# E .t; Xt / exp t sup q.s; Xs /
< 1:
(58)
0st
Z
If we set
t
8 t 0; Nt WD .t; Xt / exp
q.s; Xs / ds ;
0
then, for every t 0 and every increasing and bounded continuous function W R ! R, EŒ.Nt /Nt jXt D x x 7! K.t; x/ WD is increasing. EŒNt jXt D x The proof of this result is quite similar to that of Lemma 6. 2. To prove Theorem 8, it suffices to consider only the case where the functions A .t; x/ are increasing for every x 7! .t; x/, x 7! q.t; x/ and x 7! q C t 0. The condition (56) ensures that the function h is differentiable and
where
8 t 0; h0 .t/ D E Œe q .t; Xt /Nt ;
(59)
A q .t; x/ WD q C 8 .t; x/ 2 RC R; e .t; x/:
(60)
Indeed, if for every t 0, we set: Z t A .u; Xu / d u qC Ct WD exp 0 Z
and
t
Lt WD .t; Xt / exp
q.u; Xu / d u ;
0
then Lt D Mt Ct and, for every t 0, ItOo’s formula yields: Z t Z t Cu dMu C Mu dCu Lt L0 D 0
0
Z t A f .u; Xu /Mu Cu d u qC D Mt C 0 Z t Df Mt C e q .u; Xu /Lu d u; 0
356
A.-M. Bogso et al.
where Z t f M t WD Cu dMu ; t 0 is a continuous local martingale. 0
Moreover, for every 0 s t: ˇ ˇ ˇf M s ˇ jL0 j C jLs j C
Z
s
q .u; Xu /j Lu d u je 0
jL0 j C .1Cs/
!
Z
jqj.u; Xu / d u
0us
0
! jL0 j C .1Ct/
s
sup . C jqj C jA j/.u; Xu / exp Z
sup . C jqj C jA j/.u; Xu / exp 0ut
t
jqj.u; Xu/ d u 0
#
" which is integrable from (56). Thus, E
Ms sup f
< 1 and then .f M t ; t 0/
0st
belongs to the class (DL) (see [14], Chap. IV, Definition 1.6 and Proposition 1.7). ft ; t 0 is a true martingale. Hence: Therefore, M Z
t
8 t 0; EŒLt EŒL0 D
E Œe q .u; Xu /Lu d u 0
and d EŒLt E Œe q .t; Xt /Lt d dt h .t/ D log EŒLt D D D E Œe q .t; Xt /Nt dt EŒLt EŒLt 0
which is equivalent to:
8 t 0; E e q .t; Xt / h0 .t/ Nt D 0 (since EŒNt D 1/: Likewise, for every t 0, we set Z t A .u; Xu / d u h.t/ : qC Dt WD exp 0 Then, and, for every 0 s < t,
8 t 0; Nt D Mt Dt
(61)
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
Z
Z
t
Nt Ns D Mt Dt Ms Ds D
357
t
Du dMu C
Mu dDu
s
s
Z t A .u; Xu / h0 .u/ Mu Du d u qC s Z t ft M fs C e q .u; Xu / h0 .u/ Nu d u; DM
Df Mt f Ms C
s
where x 7! e q .u; x/ is increasing for every u 0. 3. Now, let 2 C. Then, for every 0 s < t, Z
t
EŒ .Nt / EŒ .Ns / E Z DE
t
0
0
.Nu / dNu
s
.Nu / d f Mu C E
0
q .u; Xu / h0 .u/ d u .Nu /Nu e
s
Z
t
D
E
0
q .u; Xu / h0 .u/ d u: .Nu /Nu e
s
Moreover, for every u 0, we have:
q .u; Xu / h0 .u/ .Nu /Nu e
EŒ 0 .Nu /Nu jXu EŒNu jXu e q .u; Xu / h0 .u/ : DE EŒNu jXu
E
0
For every u 0 and x 2 R, let us define: K.u; x/ WD and
EŒ
0
.Nu /Nu jXu D x EŒNu jXu D x
V .u; x/ WD EŒNu jXu D x e q .u; x/ h0 .t/ :
We note that: 8 u 0; EŒV .u; Xu / D 0: On the other hand, it follows from (57) and Lemma 7 that the function x 7! K.u; x/ is increasing. For u 0, let .e q u /1 denotes the right-continuous inverse of the function x 7! e q .u; x/. Thus, distinguishing the cases Xu .e q u /1 .0/ and Xu .e q u /1 .0/, we obtain
8 u 0; K.u; Xu /V .u; Xu / K u; .e q u /1 .0/ V .u; Xu /; and finally, for every 0 s < t:
358
A.-M. Bogso et al.
Z
t
EŒ .Nt / EŒ .Ns /
EŒK.u; Xu /V .u; Xu / d u Z
s t
K u; .e q u /1 .0/ EŒV .u; Xu / d u D 0:
t u
s
Remark 3. Suppose that .Xt ; t 0/ is a Brownian motion issued from 0. Let qN W R ! RC satisfies Z
1
.1 C jxj/q.x/ N dx < 1 and lim inf jxj2˛ q.x/ N > 0; for some ˛ < 1; x!1
0
and let be the unique solution of the Sturm-Liouville equation: 8 N < 00 .x/ D .x/q.x/ r 2 : lim 0 .x/ D ; x!1
lim .x/ D 0:
x!1
Then, Z 1 t q.X N u / d u ; t 0 is a local martingale, Nt WD .Xt / exp 2 0 and it is shown in [15] that .Nt ; t 0/ is a true martingale. We are here in the situation of Theorem 8 with qC
1 1 00 A D qN C D 0: 2 2
In other words, in this specific case, .Mt ; t 0/ is better than a peacock: it is a martingale.
5 Part II: Strong and Very Strong Peacocks 5.1 Strong Peacocks 5.1.1 Definition and Examples Definition 8. An integrable real-valued process .Xt ; t 0/ is said to be a strong peacock (resp. a strong peadock) if, for every 0 s < t and every increasing and bounded Borel function W R ! R: EŒ.Xt Xs /.Xs / 0
(62)
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
359
(resp. EŒ.Xt Xs /.Xt / 0:/ Remark 4. 1. The definition of a peacock involves only its 1-dimensional marginals. On the other hand, the definition of a strong peacock involves its 2-dimensional marginals. 2. If .Xt ; t 0/ is a strong peacock, then EŒXt does not depend on t (it suffices to apply (62) with D 1 and D 1). Every strong peacock is a peacock; indeed, if 2 CC , then: EŒ .Xt / EŒ .Xs / EŒ
0
.Xs /.Xt Xs / 0:
3. If .Xt ; t 0/ is a strong peacock such that E Xt2 < 1 for every t 0, then: EŒXs .Xt Xs / 0; for every 0 s < t:
(63)
4. If X and Y are two processes which have the same 1-dimensional marginals, it may be possible that X is a strong peacock while Y is not. For example, let us Bt 1 consider .Xt WD t 4 B1 ; t 0/ and Yt WD 1 ; t 0 , where .Bt ; t 0/ is a t4 Brownian motion started at 0. By Lemma 2, .Xt ; t 0/ is a strong peacock while .Yt ; t 0/ is not. Indeed, for every a 2 R and 0 < s t: 2 6 8 E6 41< B
s
:
1
s4
9 =
> a;
3 # " Bt 1 Bs 7 1
< 0: 1 7 1 E 1 1 Bs 1 1 5D Bs > as 4 t4 s4 t4 s4
More generally, for every nonnull martingale .Mt ; t 0/ and every increasing Mt ; t 0 is not a strong peacock. Borel function ˛ W RC ! RC , ˛.t/ 5. Theorem 1 remains true if one replaces peacock by strong peacock. Example 11. Some examples of strong peacocks: • Martingales: Indeed, if .Mt ; t 0/ is a martingale with respect to some filtration .Ft ; t 0/, then, for every bounded Borel function W R ! R: EŒ.Ms /.Mt Ms / D EŒ.Ms /.EŒMt jFs Ms / D 0: 1 and ˛ W RC ! RC is a strictly • If .Mu ; u 0/ is a martingale belonging to Hloc increasing Borel function such that ˛.0/ D 0, then
(see [11, Chap. 1])
1 ˛.t/
Z
t
Mu d˛.u/ 0
is a strong peacock
360
A.-M. Bogso et al.
• The process .tX; t 0/ where X is a centered and integrable r.v. tX e ; t 0 (see [11, Chap. 1]). • The process EŒe tX In the case of Gaussian processes, we obtain a characterization of strong peacocks using the covariance function. Indeed, one has: Proposition 1. A centered Gaussian peacock .Xt ; t 0/ is strong if and only if, for every 0 < s t: EŒXs Xt EŒXs2 : (64) We note that a centered gaussian process .Xt ; t 0/ is a peacock if and only if (65) t 7! E Xt2 is increasing and of course, (64) implies (65); indeed, for every 0 < s t: 1 1 E Xs2 E ŒXs Xt E Xs2 2 E Xt2 2 ; (from Schwartz’s Inequality) which implies (65). Proof (of Proposition 1). Let .Xt ; t 0/ be a centered Gaussian strong peacock. 1. By taking .x/ D x in (62), we have: EŒXs .Xt Xs / 0; for every 0 < s t; i.e., K.s; t/ K.s; s/; for every 0 < s t: 2. Conversely, if (64) holds, then, for every 0 < s t and every increasing Borel function W R ! R: EŒ.Xs /.Xt Xs / D EŒ.Xs /.EŒXt jXs Xs / K.s; t/ 1 EŒ.Xs /Xs 0 (from Lemma 2). D K.s; s/
t u
Example 12. We give two examples: • An Ornstein-Uhlenbeck process with parameter c 2 R: Z Xt D Bt C c
t
Xu d u; 0
where .Bt ; t 0/ is a Brownian motion started at 0, is a peacock for every c and a strong peacock if and only if c 0. Indeed, for every t 0,
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
Z Xt D e
t
ct
361
e cs dBs
0
and, for every 0 < s t, since EŒXs Xt D we have:
Thus,
sinh.cs/ ct e ; c
sinh.cs/ ct EŒXs Xt E Xs2 D Œe e cs : c EŒXs Xt E Xs2 0 if and only if c 0:
• A fractional Brownian motion .Xt ; t 0/ with index H 2 Œ0; 1 is a peacock for 1 every H and a strong peacock if and only if H . This follows from the fact 2 that, 1 K.s; t/ K.s; s/ D .t 2H s 2H .t s/2H / 2 1 is positive for every 0 < s t if and only if H , where K denotes the 2 covariance function of .Xt ; t 0/.
5.1.2 Upper and Lower Orthant Orders Let X D .X1 ; X2 ; : : : ; Xp / and Y D .Y1 ; Y2 ; : : : ; Yp / be two Rp -valued random vectors. The following definitions are taken from Shaked and Shanthikumar [17], p. 140. Definition 9. (Upper orthant order). X is said to be smaller than Y in the upper orthant order (notation: X Y ) if one u.o
of the two following equivalent conditions is satisfied: 1. For every p-tuple 1 ; 2 ; : : : ; p of reals: P.X1 > 1 ; X2 > 2 ; : : : ; Xp > p / P.Y1 > 1 ; Y2 > 2 ; : : : ; Yp > p / (66) 2. For every p-tuple l1 ; l2 ; : : : ; lp of nonnegative increasing functions: " E
p Y i D1
# li .Xi / E
"
p Y i D1
# li .Yi /
(67)
362
A.-M. Bogso et al.
Definition 10. (Upper orthant order for processes). A process .Xt ; t 0/ is smaller than a process .Yt ; t 0/ for the upper orthant order (notation: .Xt ; t 0/ .Yt ; t 0/) if, for every integer p and every 0 u.o
t1 < t2 < < tp :
.Xt1 ; Xt2 ; : : : ; Xtp / .Yt1 ; Yt2 ; : : : ; Ytp /:
(68)
u.o
If X and Y are two càdlàg processes, (68) is equivalent to: for every h W R ! R càdlàg: P.for every t 0; Xt h.t// P.for every t 0; Yt h.t//:
(69)
Definition 11. (Lower orthant order). X is said to be smaller than Y in the lower orthant order (notation: X Y ) if one l.o
of the two following equivalent conditions is satisfied: 1. For every p-tuple 1 ; 2 ; : : : ; p of reals: P.X1 1 ; X2 2 ; : : : ; Xp p / P.Y1 1 ; Y2 2 ; : : : ; Yp p / (70) 2. For every p-tuple l1 ; l2 ; : : : ; lp of nonnegative decreasing functions: " E
p Y
# li .Xi / E
i D1
"
p Y
# li .Yi /
(71)
i D1
Definition 12. (Lower orthant order for processes). A process .Xt ; t 0/ is smaller than a process .Yt ; t 0/ for the lower orthant order (notation: .Xt ; t 0/ .Yt ; t 0/) if, for every integer p and every 0 t1 < t2 < < tp :
l.o
.Xt1 ; Xt2 ; : : : ; Xtp / .Yt1 ; Yt2 ; : : : ; Ytp /:
(72)
l.o
If X and Y are two càdlàg processes, (72) is equivalent to: for every h W R ! R càdlàg: P.for every t 0; Xt h.t// P.for every t 0; Yt h.t//:
(73)
Remark 5. Observe that, if X D .Xt ; 0/ and Y D .Yt ; 0/ are two processes such that: X Y X; then l.o
u.o (1.d)
.Xt ; t 0/ D .Yt ; t 0/
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
363
Let .Xt ; t 0/ be a real-valued measurable process and, for t 0, let Ft denotes the distribution function of Xt . If U is uniformly distributed on Œ0; 1, then (1.d)
.Xt ; t 0/ D
Ft1 .U /; t 0 :
Moreover, we state the following: Proposition 2. Let .Xt ; t 0/ a real-valued process, and for t 0, let Ft be the distribution function of Xt .Then, if U is uniformly distributed on Œ0; 1, one has:
Ft1 .U /; t 0
.Xt ; t 0/ .Ft1 .U //: u.o.
l.o.
Proof (of Proposition 2). For every integer p, every p-tuple 1 ; 2 ; : : : ; p of reals and every 0 t1 < t2 < < tp : P.Xt1 > 1 ; Xt2 > 2 ; : : : ; Xtp > p /
min
i D1;2;:::;p
P.Xti > i /
D 1 max Fti .i / i D1;2;:::;p
DP U >
max Fti .i /
i D1;2;:::;p
D P U > Ft1 .1 /; U > Ft2 .2 /; : : : ; U > Ftp .p / 1 1 D P Ft1 .U / > ; F .U / > ; : : : ; F .U / > 1 2 p : t2 tp 1 On the other hand, one has: P.Xt1 1 ; Xt2 2 ; : : : ; Xtp p /
min
i D1;2;:::;p
P.Xti i /
1 1 : .U / ; F .U / ; : : : ; F .U / D P Ft1 1 2 p t t 1 2 p
t u
Let us introduce some definitions. Definition 13. For a given family of probability measures D .t ; t 0/, we denote by D the set of real-valued processes which admit the family as onedimensional marginals: D WD f.Xt ; t 0/I for every t 0; Xt t g: In particular, if the family increases in the convex order, then D is the set of peacocks associated to .
364
A.-M. Bogso et al.
The next corollary follows immediately from Proposition 2. Corollary 1. Let be a family of probability measures. Then, the process .Ft1 .U /; t 0/ is an absolute maximum of D for the upper orthant order and an absolute minimum of D for the lower orthant order. The following result is due to Cambanis et al. [4]. Theorem 9. Let .X1 ; X2 / and .Y1 ; Y2 / be two R2 -valued random vectors such that: (law)
(law)
X1 D Y1 ; X2 D Y2 and .X1 ; X2 / .Y1 ; Y2 /
(74)
l.o
Let k W R R ! R be right-continuous and quasi-monotone, i.e: k.x; y/ C k.x 0 ; y 0 / k.x; y 0 / k.x 0 ; y/ 0; for every x x 0 ; y y 0 :
(75)
Suppose that the expectations EŒk.X1 ; X2 / and EŒk.Y1 ; Y2 / exist (even if infinite valued) and either of the following conditions is satisfied: i. k is symmetric and the expectations EŒk.X1 ; X1 / and EŒk.X2 ; X2 / are finite ii. The expectations EŒk.X1 ; x1 / and EŒk.x2 ; X2 / are finite for some x1 and x2 . Then: EŒk.X1 ; X2 / EŒk.Y1 ; Y2 /: The next result is deduced from Proposition 2 and Theorem 9. Corollary 2. Let X WD .Xt ; t 0/ be a peacock and, for every t 0, let Ft be the distribution function of Xt . Let U be uniformly distributed on Œ0; 1. Then: (1.d)
1. For every real-valued process Y WD .Yt ; t 0/ such that Yt D Xt and every quasi-monotone function k W R R ! R satisfying the same conditions as in Theorem 9, one has:
8.s; t/ 2 RC RC ; E k Fs1 .U /; Ft1 .U / EŒk.Ys ; Yt /:
(76)
In particular, for every p 1 such that EŒjXu jp < 1, for every u 0 and every .s; t/ 2 RC RC , hˇ ˇp i E ˇFt1 .U / Fs1 .U /ˇ E jYt Ys jp ; 2. .Ft1 .U /; t 0/ is a strong peacock.
(77)
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
365
To prove Corollary 2 we may observe, for the first point, that for every p 1, the function k W .x; y/ 7! jx yjp is quasi-monotone and, for the second point, that if W R ! R is increasing, then k W .x; y/ 7! .x/.y x/ is a quasi-monotone function.
5.1.3 A Comparison Theorem for Peacocks Let .Xt ; t 0/ be a real-valued process which is square integrable and which satisfies: t 7! Xt is a.s. measurable: (78) For a probability measure on f.s; t/I 0 s tg, let us define the 2-variability of .Xt ; t 0/ with respect to by the quantity: “ ˘ .X / WD f0st g
EŒ.Xt Xs /2 .ds; dt/:
Definition 14. For a family of probability measures WD .t ; t 0/, let DC denotes the set of strong peacocks which admit as their one-dimensional marginals family: DC WD f.Xt ; t 0/I X is a strong peacock such that, 8t 0; Xt t g: Given a family of probability measures WD .t ; t 0/ which increases in the convex order, we wish to determinate for which processes in DC , ˘ attains its maximum (resp. its minimum). Theorem 10. Let be a probability measure on f.s; t/I 0 s tg. 1. The maximum of ˘ .X / in DC is equal to: ZZ max ˘ .X / D
X 2DC
f0st g
2 E Xt E Xs2 .ds; dt/
(79)
and is attained when .Xt ; t 0/ is a martingale. 2. The minimum of ˘ .X / in DC is equal to: ZZ min ˘ .X / D
X 2DC
f0st g
E
h
2 i Ft1 .U / Fs1 .U / .ds; dt/
and is attained by Xt D Ft1 .U /; t 0 .
(80)
366
A.-M. Bogso et al.
Proof (of Theorem 10). 1. Let .Xt ; t 0/ be a strong peacock. For every 0 s t, one has: E .Xt Xs /2 D E Xt2 C E Xs2 2EŒXt Xs D E Xt2 E Xs2 2EŒ.Xt Xs /Xs (from (62)): E Xt2 E Xs2 Hence, integrating against , we obtain: ZZ max ˘ .X /
X 2DC
f0st g
2 E Xt E Xs2 .ds; dt/ WD M.X /
and M.X / is clearly attained when .Xt ; t 0/ is a martingale. 2. This point is a consequence of Theorem 9 and Corollary 2.
t u
6 Very Strong Peacocks 6.1 Definition, Examples and Counterexamples Definition 15. An integrable real-valued process .Xt ; t 0/ is said to be a very strong peacock (VSP) if, for every n 2 N , every 0 < t1 < < tn < tnC1 and every 2 En , we have: E .Xt1 ; : : : ; Xtn / .XtnC1 Xtn / 0:
(81)
Remark 6. 1. The definition of a strong peacock involves its 2-dimensional marginals while the definition of a very strong peacock involves all its finite-dimensional marginals. 2. Every very strong peacock is a strong peacock. But, the converse is not true. Let us give two examples: (a) Let G1 and G2 be two independent, centered Gaussian r.v.’s such that EŒG12 D EŒG22 D 1, ˛, ˇ be two constants satisfying 1 C 2˛ 2 ˇ and .X1 ; X2 ; X3 / be the random Gaussian vector defined by: X1 D G1 ˛G2 ; X2 D ˇG1 ; X3 D ˇG1 C ˛G2 :
(82)
Then, .X1 ; X2 ; X3 / is a strong peacock (from Proposition 1) which is not a very strong peacock since EŒX1 .X3 X2 / D ˛ 2 E G12 < 0:
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
367
(b) Likewise, let G1 and G2 be symmetric, independent and identically two distributed r.v.’s such that E Gi2 D 1 (i D 1; 2). Then, for every ˇ 3, the random vector .X1 ; X2 ; X3 / given by: X1 D G1 G2 ; X2 D ˇG1 ; X3 D ˇG1 C G2 :
(83)
is a strong peacock for which (81) does not hold. Proof. Since G1 and G2 are independent and centered, we first observe that: E 1fX2 ag .X3 X2 / D E 1fˇG1 ag G2 D 0: Moreover, E 1fX1 ag .X2 X1 / D E 1fG1 G2 ag ..ˇ 1/G1 C G2 / D .ˇ 1/E 1fG1 G2 ag G1 C E 1fG1 G2 ag G2 D .ˇ 1/E 1fG2 G1 ag G2 CE 1fG1 G2 ag G2 „ ƒ‚ … (by interchanging G1 and G2 /
D .ˇ 2/E 1fG2 G1 ag G2 CE 1fjG1 G2 jag G2 ƒ‚ … „
0 by Lemma 2, since ˇ>2
E 1fjG1 G2 jag G2 D 0; (since G1 and G2 are symmetric) and similarly, E 1fX1 ag .X3 X1 / D E 1fG1 G2 ag ..ˇ 1/G1 C 2G2 / D .ˇ 1/E 1fG1 G2 ag G1 C 2E 1fG1 G2 ag G2 D .ˇ 1/E 1fG2 G1 ag G2 C 2E 1fG1 G2 ag G2 D .ˇ 3/E 1fG2 G1 ag G2 C2E 1fjG1 G2 jag G2 „ ƒ‚ … 0 by Lemma 2, since ˇ3
2E 1fjG1 G2 jag G2 D 0: Thus, .X1 ; X2 ; X3 / is a strong peacock. But, .X1 ; X2 ; X3 / is not a very strong peacock since EŒX1 .X3 X2 / D E G12 < 0: t u Let us give some examples of very strong peacocks.
368
A.-M. Bogso et al.
Example 13. 1. Each of the processes cited in Example 11 is a very strong peacock. We refer the reader to ([11], Chap. 8) for further examples. 2. Let .t ; t 0/ be an increasing process with independent increments (for example a subordinator) and f W R ! R be a convex and increasing (or concave and decreasing) function such that EŒjf .t /j < 1, for every t 0. Then, .Xt WD f .t / EŒf .t /; t 0/ is a very strong peacock. Proof. Let f be a convex and increasing function and let n 1, 0 < t1 < t2 < < tn < tnC1 and 2 En . We first note that: e W .x1 ; : : : ; xn / 7! .f .x1 / EŒf .t1 /; : : : ; f .xn / EŒf .tn // belongs to En (84) and, by setting cn WD EŒf .tnC1 / EŒf .tn /, E .Xt1 ; : : : ; Xtn /.XtnC1 Xtn / D E e .t1 ; : : : ; tn /.f .tnC1 / f .tn / cn / : (85) Let us prove by induction that, for every i 2 1; n, there exists a function 'i 2 Ei such that: E .Xt1 ; : : : ; Xtn /.XtnC1 Xtn / E 'i .t1 ; : : : ; ti /.f .tnC1 / f .tn / cn / : (86) . On the other hand, let us suppose We note that, for i D n, we may choose 'n D e that (86) holds for some i 2 1; n. Then, since ti is independent of tnC1 ti and tn ti , one has: E .Xt1 ; : : : ; Xtn /.XtnC1 Xtn / (by induction) E 'i .t1 ; : : : ; ti /.f .tnC1 / f .tn / cn /
D E 'i .t1 ; : : : ; ti / f .ti C tnC1 ti / f .ti C tn ti / cn
D E 'i .t1 ; : : : ; ti / EŒf .ti C tnC1 ti /jFti EŒf .ti C tn ti /jFti cn (where Fti WD .s ; 0 s ti // i h bi .t / ; D E 'i .t1 ; : : : ; ti /f i bi .x/ D EŒf .x C t t / EŒf .x C t t / cn /: (where f i n i nC1
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
369
bi is increasing since f is convex and t But, the function f tn . Hence, nC1 h i bi .t / E .Xt1 ; : : : ; Xtn /.XtnC1 Xtn / E 'i .t1 ; : : : ; ti 1 ; ti / f i i h b1 .0/ f bi .t / E 'i t1 ; : : : ; ti 1 ; f i i
h i b1 .0/ f .t / f .t / cn ; D E 'i t1 ; : : : ; ti 1 ; f n i nC1 i.e., (86) also holds for i 1 with b1 .0/ : 'i 1 W .x1 ; : : : ; xi 1 / 7! 'i x1 ; : : : ; xi 1 ; f i Thus, (86) holds for every i 2 1; n. In particular, for i D 1, there exists '1 2 E1 such that: h i b1 .t / E .Xt1 ; : : : ; Xtn /.XtnC1 Xtn / E '1 .t1 / f 1 h i b1 .0/ E f b1 .t / '1 f 1 1 b1 .0/ E f .t / f .t / cn D 0: D '1 f n 1 nC1 t u
6.2 Peacocks Obtained by Quotient Under the Very Strong Peacock Hypothesis Lemma 8. An integrable real-valued process is a very strong peacock if and only if, for every n 1, every 0 < t1 < < tn < tnC1 , every i n and every 2 En : E .Xt1 ; : : : ; Xtn / .XtnC1 Xti / 0:
(87)
Proof (of Lemma 8). For every n 1 and i n, we shall prove by induction the following condition: E .Xt1 ; : : : ; Xtn / .XtnC1 XtnC1i / 0
(88)
which, of course, is equivalent to (87). If i D 1, we recover (81). Now, let 1 i n 1 be fixed and suppose that (81) is satisfied and that (88) holds for i . Let us prove that (88) is also true for i C 1. One has:
370
A.-M. Bogso et al.
E .Xt1 ; : : : ; Xtn1 ; Xtn / .XtnC1 XtnC1.i C1/ / D E .Xt1 ; : : : ; Xtn1 ; Xtn / .XtnC1 Xtn / ƒ‚ … „ 0 (from (81))
C E Œ .Xt1 ; : : : ; Xtn1 ; Xtn / .Xtn Xtni / E Œ .Xt1 ; : : : ; Xtn1 ; Xtn / .Xtn Xtni / E Œ .Xt1 ; : : : ; Xtn1 ; Xtni / .Xtn Xtni / 0 (since belongs to En and (88) holds for 1 i n 1):
t u
The importance of very strong peacocks lies in the following result. Theorem 11. Let .Xt ; t 0/ be a right-continuous and centered very strong peacock such that for every t 0: #
"
sup jXs j < 1:
E
(89)
s2Œ0;t
Then, for every right-continuous and strictly increasing function ˛ W RC ! RC such that ˛.0/ D 0: Qt WD
1 ˛.t/
Z
t
Xs d˛.s/; t 0
is a peacock.
0
Remark 7. 1. Theorem 11 is a generalization of the case where .Xs ; s 0/ is a martingale (see Example 1). 2. Let .s ; s 0/ be a subordinator and f W RC ! R be increasing, convex and such that EŒjf .t /j < 1, for every t 0. Then, it follows from Theorem 11 and from the second point of Example 13 that, for every right-continuous and strictly increasing function ˛ W RC ! RC satisfying ˛.0/ D 0: Qt WD
1 ˛.t/
Z
t
.f .s / EŒf .s // d˛.s/; t 0
is a peacock.
0
Proof (of Theorem 11). Let T > 0 be fixed. 1. Let us first suppose that 1Œ0;T d˛ is a linear combination of Dirac measures and show that, for every r 2 2; 1, every a1 > 0; a2 > 0; : : : ; ar > 0 such that ˛.r/ WD
r X i D1
ai D ˛.T /
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
371
and every 0 < 1 < 2 < < n T : ! n 1 X Qn WD ai Xi ; n 2 1; r is a peacock. ˛.n/ i D1
(90)
2 CC and n 2. For every n 2 2; r, one has:
Let
EŒ .Qn / EŒ .Qn1 / EŒ 0 .Qn1 /.Qn Qn1 / " ! !# n1 n n1 X X X 1 1 1 0 DE ai Xi ai Xi ai Xi ˛.n 1/ i D1 ˛.n/ i D1 ˛.n 1/ i D1 X an ai E Œ .X1 ; : : : ; Xn1 / .Xn Xi / ; ˛.n/˛.n 1/ i D1 n1
D where
X 1 ai xi ˛.n 1/ i D1 n1
W .x1 ; : : : ; xn1 / 7!
0
! belongs to En1 :
Then, the result follows from Lemma 8. 2. Let us set D 1Œ0;T d˛ and, for every 0 t T ,
Qt WD
1 .Œ0; t/
Z
t
Xu .d u/: 0
Since the function 2 Œ0; T 7! X is right-continuous and bounded from above by sup jX j which is finite a.s., then there exists a sequence .n ; n 0/ of 0T R R measures of type used in (1), with supp n Œ0; T , n .d u/ D .d u/ and, for every 0 t T : Z
Z
t
lim
n!1 0
t
Xu n .d u/ D
Xu .d u/ a.s.
(91)
0
lim n .Œ0; t/ D .Œ0; t/:
n!1
(92)
Then, from (91) and (92), it follows that: .n /
lim Qt
n!1
But, using (1),
./
D Qt
a.s., for every 0 t T:
(93)
372
A.-M. Bogso et al.
Table 1 Table of the main peacocks studied in this paper Main hypothesis Peacocks R t .Xt ; t 0/ is conditionally Ct WD 0 q.s; Xs /ds .t /; t 0 h R i monotone and is positive, t with .t / D E q.s; X /ds s 0 convex and increasing .Xt ; t 0/ is a process with independent and log-concave increments; is continuous and x 7! t .x/ WD .t; x/ is increasing for every t 0
R 1 t exp 0 q.s; Xs / d˛.s/ @Nt WD h R i ; t 0A t E exp 0 q.s; Xs / d˛.s/ 0
.Lt ; t 0/ is a Lévy process suchthat R t the variable exp 0 Lu d u is integrable
Theorem 3
R 1 t t .Xt / exp 0 qs .Xs / ds @Nt WD h R i ; t 0A Theorem 8 t E t .Xt / exp 0 qs .Xs / ds Nt WD
.Xt / ;t 0 EŒ .Xt /
Theorem 4
.Ct WD .Xt / EŒ .Xt /; t 0/
.Xt ; t 0/ is a centered very strong peacock
Theorem 2
0
.Xt ; t 0/ solves an SDE, is positive and increasing
.Xt ; t 0/ is conditionally monotone and solves an SDE; is a positive Radon measure on RC
References
0 @Nt WD
Qt WD
Rt E
1 q.s; Xs /.ds/
hR t 0
Theorem 5
0
i ; t 0A q.s; Xs /.ds/
1 Rt X d˛.s/; t 0 s ˛.t / 0
Theorem 6
Theorem 11
R 1 t exp 0 Ls ds @Nt WD h R i ; t 0A t E exp 0 Ls ds 0
0 @N et WD
R 1 t
1
Example 7
exp t 0 Ls ds h R i ; t 0A t E exp 1t 0 Ls ds
In this table: • ˛ W RC ! RC is a right-continuous and increasing function such that ˛.0/ D 0 • q W RC R ! RC is a continuous and positive function such that, for every s 0, x 7! qs .x/ WD q.s; x/ is increasing
Peacocks Obtained by Normalisation: Strong and Very Strong Peacocks
373
. / Qt n ; 0 t T is a peacock for every n 0; i h i h . / . / i.e., for every 0 s < t T , E Qs n D E Qt n and, for every
2 CC :
h i . / .Qs.n / / D E .Qt n / :
(95)
ˇ ˇ ˇ . / ˇ sup sup ˇQt n ˇ sup jX j
(96)
E Moreover,
(94)
0t T n0
0T
which is integrable from (89). ./ Therefore, using (3), (93)–(95) and the dominated convergence Theorem, Qt ; 0 t T / is a peacock for every T > 0. t u
References 1. M.Y. An, Log-concave probability distributions: Theory and statistical testing. Papers 96-01. Centre for Labour Market and Social Research, Danmark (1996) 2. D. Baker, M. Yor, A Brownian sheet martingale with the same marginals as the arithmetic average of geometric Brownian motion. Electron. J. Probab. 14(52), 1532–1540 (2009) 3. A.-M. Bogso, C. Profeta, B. Roynette, in Some Examples of Peacocks in a Markovian Set-Up, ed. by Donati-Martin. Séminaire de Probabiliés XLIV (Springer, Berlin, 2012) 4. S. Cambanis, G. Simons, W. Stout, Inequalities for Ek.X; Y / when the marginals are fixed. Z. Wahrscheinlichkeitstheorie und Verw. Gebiete 36, 285–294 (1976) 5. P. Carr, C.-O. Ewald, Y. Xiao, On the qualitative effect of volatility and duration on prices of Asian options. Finance Res. Lett. 5(3), 162–171 (2008) 6. H. Daduna, R. Szekli, A queueing theoretical proof of increasing property of Pólya frequency functions. Stat. Probab. Lett. 26(3), 233–242 (1996) 7. B. Efron, Increasing properties of Pólya frequency functions. Ann. Math. Stat. 36, 272–279 (1965) 8. F. Hirsch, B. Roynette, M. Yor, Applying ItOo’s motto: “Look at the infinite dimensional picture” by constructing sheets to obtain processes increasing in the convex order. Period. Math. Hungar. 61(1–2), 195–211 (2010) 9. F. Hirsch, B. Roynette, M. Yor, Unifying constructions of martingales associated with processes increasing in the convex order, via Lévy and Sato sheets. Expo. Math. 4, 299–324 (2010) 10. F. Hirsch, B. Roynette, M. Yor, From an Itô type calculus for Gaussian processes to integrals 992 of log-normal processes increasing in the convex order. J. Mat. Soc. Jpn. 63(3), 887–917 (2011) 11. F. Hirsch, C. Profeta, B. Roynette, M. Yor, Peacocks and Associated Martingales, with explicit constructions. Bocconi & Springer Series, vol. 3. Springer, Milan; Bocconie University Press, Milan, 2011. xxxii+384 pp 12. J. Jacod, P. Protter, Time reversal on Lévy processes. Ann. Probab. 16(2), 620–641 (1988) 13. H.G. Kellerer, Markov-Komposition und eine Anwendung auf Martingale. Math. Ann. 198, 99–122 (1972)
374
A.-M. Bogso et al.
14. D. Revuz, M. Yor, in Continuous Martingales and Brownian Motion. Grundlehren der Mathematischen Wissenschaften, vol. 293, 3rd edn. (Springer, Berlin, 1999) 15. B. Roynette, P. Vallois, M. Yor, Limiting laws associated with Brownian motion perturbed by normalized exponential weights I. Studia Sci. Math. Hungar. 43(2), 171–246 (2006) 16. I.J. Schoenberg, On Pólya frequency functions I. The totally positive functions and their Laplace transforms. J. Analyse Math. 1, 331–374 (1951) 17. M. Shaked, J.G. Shanthikumar, in Stochastic Orders and Their Applications. Probability and Mathematical Statistics (Academic, Boston, 1994) 18. M. Shaked, J.G. Shanthikumar, in Stochastic Orders. Springer Series in Statistics (Springer, New York, 2007) 19. J.G. Shanthikumar, On stochastic comparison of random vectors. J. Appl. Probab. 24(1), 123–136 (1987)
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths Simon C. Harris and Matthew I. Roberts
Abstract We give a proof of a result on the growth of the number of particles along chosen paths in a branching Brownian motion. The work follows the approach of classical large deviations results, in which paths of particles in C Œ0; T , for large T , are rescaled onto C Œ0; 1. The methods used are probabilistic and take advantage of modern spine techniques.
1 Introduction and Statement of Result 1.1 Introduction Fix a positive real number r > 0 and a random variable A taking values in f2; 3; : : :g such that m WD EŒA 1 > 1 and EŒA log A < 1. We consider a branching Brownian motion (BBM) under a probability measure P, which is described as follows. We begin with one particle at the origin. Each particle u, once born, performs a Brownian motion independent of all other particles, until it dies, an event which occurs at an independent exponential time after its birth with mean 1=r. At the time of a particle’s death it is replaced (at its current position) by a random number Au of offspring where Au has the same distribution as A. Each of
S.C. Harris () Department of Mathematical Sciences, University of Bath, Claverton Down, Bath, Avon BA2 7AY, UK e-mail: [email protected] M.I. Roberts Laboratoire de Probabilités et Modèles Aléatoires, Université Paris VI, 4, Place Jussieu, 75005 Paris, France C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__17, © Springer-Verlag Berlin Heidelberg 2012
375
376
S.C. Harris and M.I. Roberts
these particles, relative to its initial position, repeats (independently) the stochastic behaviour of its parent. We let N.t/ be the set of particles alive at time t, and for u 2 N.t/ and s t let Xu .s/ be the position of particle u (or its ancestor) at time s. Fix a set D C Œ0; 1 and 2 Œ0; 1; then we are interested in the size of the sets NT .D; / WD fu 2 N.T / W 9f 2 D with Xu .t/ D Tf .t=T / 8t 2 Œ0; T g for large T .
1.2 The Main Result We define the class H1 of functions by Z s g.s/ds 8s 2 Œ0; 1 ; H1 WD f 2 C Œ0; 1 W 9g 2 L2 Œ0; 1 with f .s/ D 0
and to save on notation we set f 0 .t/ WD 1 if f 2 C Œ0; 1 is not differentiable at the point t. We then take integrals in the Lebesgue sense so that we may integrate functions that equal 1 on sets of zero measure. We let (
1 0 .f / WD inf 2 Œ0; 1 W rm 2
Z
) f 0 .s/2 ds < 0
2 Œ0; 1 [ f1g
0
(we think of 0 as the extinction time along f , the time at which the number of particles near f hits zero) and define our rate function K, for f 2 C Œ0; 1 and 2 Œ0; 1, as ( K.f; / WD
rm 1
1 2
R 0
f 0 .s/2 ds if f 2 H1 and 0 .f / otherwise.
We expect approximately exp.K.f; /T / particles whose paths up to time T (when suitably rescaled) look like f . This is made precise in Theorem 1. Theorem 1. For any closed set D C Œ0; 1 and 2 Œ0; 1, lim sup T !1
1 log jNT .D; /j sup K.f; / T f 2D
almost surely, and for any open set U C Œ0; 1 and 2 Œ0; 1,
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
lim inf T !1
377
1 log jNT .U; /j sup K.f; / T f 2U
almost surely. Sections 3 and 4 will be concerned with giving a proof of this theorem. An almost identical result was stated by Git in [2]. We would like to give an alternative proof for two reasons. Firstly, we believe that our proof of the lower bound is perhaps more intuitive, and certainly more robust, than that given in [2]. There are many more general setups for which our proofs will go through without too much extra work. One possibility is to allow particles to die without giving birth to any offspring (that is, to allow A to take the value 0): in this case the statement of the theorem would be conditional on the survival of the process, and we will draw attention to any areas where our proof must be adapted significantly to take account of this. There is work in progress on some further interesting cases and their applications, in particular the case where breeding occurs at the inhomogeneous rate rx p , p 2 Œ0; 2/, for a particle at position x. Secondly, there seems to be a slight oversight in the proof of Lemma 1 in [2], and that lemma is then used in obtaining both the upper and lower bounds. Although the gap seems minor at first, the complete lack of simple continuity properties of the processes involved means that almost all of the work involved in proving the upper bound is concerned with this matter. We give details of the oversight as an appendix. Our tactic for the proof is to first work along lattice times, and then upgrade to the full result using Borel-Cantelli arguments. We begin, in Sect. 2, by introducing a family of martingales and changes of measure which will provide us with intuitive tools for our proofs. We then apply these tools to give an entirely new proof of the lower bound for Theorem 1 in Sect. 3. Finally, in Sect. 4, we take the same approach as in [2] to gain the upper bound along lattice times, and then rule out some technicalities in order to move to continuous time. This work complements the article by Harris and Roberts [5]. Large deviation probabilities for the same model were given by Lee [6] and Hardy and Harris [3].
2 A Family of Spine Martingales 2.1 The Spine Setup We will need to use some modern “spine” techniques as part of our proof. We only need some of the most basic spine tools, and we do not attempt to explain the details of these rigorously, but rather refer the interested reader to the article [4]. We first embellish our probability space by keeping track of some extra information about one particular infinite line of descent or spine. This line of descent is
378
S.C. Harris and M.I. Roberts
defined as follows: our original particle is part of the spine; when this particle dies, we choose one of its offspring uniformly at random to become part of the spine. We continue in this manner: when a spine particle dies, we choose uniformly at random between its offspring to decide which becomes part of the spine. In this way at any time t 0 we have exactly one particle in N.t/ that is part of the spine. We refer to both this particle and its position with the label t ; this is an abuse of notation, but it should always be clear from the context which meaning is intended. It is not hard to see that the spatial motion of the spine, .t /t 0 , is a standard Brownian motion. The resulting probability measure (on the set of marked Galton-Watson trees with Q and we find need for four different filtrations to encode spines) we denote by P, differing amounts of this new information: • Ft contains the all the information about the marked tree up to time t. However, it does not know which particle is the spine at any point. Thus it is simply the natural filtration of the original branching Brownian motion. • FQt contains all the information about both the marked tree and the spine up to time t. • GQt contains all the information about the spine up to time t, including the birth times of other particles along its path, and how many particles were born at each of these times; it does not know anything about the rest of the tree. • Gt contains just the spatial information about the spine up to time t; it does not know anything about the rest of the tree. We note that Ft FQt and Gt GQt FQt , and also that PQ is an extension of P in Q F1 D P. All of the above is covered more rigorously in [4]. that Pj Lemma 1 (Many-to-one lemma). If g.t/ is Gt -measurable and can be written X
g.t/ D
gu .t/½ft Dug
u2N.t /
where each gu .t/ is Ft -measurable, then 2 E4
X
3 Q gu .t/5 D e rmt EŒg.t/:
u2N.t /
This lemma is extremely useful as it allows us to reduce questions about the entire population down to calculations involving just one standard Brownian motion—the spine. A proof may be found in [4].
2.2 Martingales and Changes of Measure For f 2 C Œ0; 1 and 2 Œ0; 1 define NT .f; "; / WD fu 2 N.T / W jXu .t/ Tf .t=T /j < "T 8t 2 Œ0; T g
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
379
so that NT .f; "; / D NT .B.f; "/; /. We look for martingales associated with these sets. For convenience, in this section we use the shorthand NT .t/ WD NT .f; "; t=T /: Q if Since the motion of the spine is simply a standard Brownian motion under P, 2 f 2 C Œ0; 1 then Itô’s formula shows that for t 2 Œ0; T , the process VT .t/ WD e
2 t =8"2 T 2
cos
Rt 0 R 1 t 0 2 .t Tf .t=T // e 0 f .s=T /d s 2 0 f .s=T / ds 2"T
Q By stopping this process at the first exit time of the is a Gt -martingale under P. Brownian motion from the tube f.x; t/ W jTf .t=T / xj < "T g, we obtain also that T .t/ WD VT .t/½fjTf .s=T /s j<"T
8st g
is a Gt -martingale on Œ0; T . As in [4], we may build from T a collection of FQt martingales QT on Œ0; T given by Y
QT .t/ WD
Av e rmt T .t/;
v<t
but these martingales will not be examined in this article—they are important only in changing measure below, and in that when we project QT .t/ back onto Ft we get a new set of mean-one Ft -martingales ZT . These processes ZT are the main objects of interest in this section, and can be expressed for t 2 Œ0; T as the sum X
ZT .t/ D
VT .t/e rmt .u/
u2NT .t /
where .u/
VT .t/ WD e
2 t =8"2 T 2
cos
Rt 0 R 1 t 0 2 .Xu .t/ Tf .t=T // e 0 f .s=T /dXu .s/ 2 0 f .s=T / ds : 2"T
Q T , via We now define new measures, Q Q T j Q D QT .t/Pj Q Q Q Ft Ft for t T —and note that Q T jFt D ZT .t/Pj Q Ft Q
and
Q T jGt D T .t/Pj Q Gt : Q
Q T , the spine moves as a Brownian motion with drift Lemma 2. Under Q f 0 .t=T /
tan .x Tf .t=T // 2"T 2"T
380
S.C. Harris and M.I. Roberts
when at position x at time t; in particular, Q T -almost surely. jt Tf .t=T /j "T 8t T Q Each particle u in the spine dies at an accelerated rate .m C1/r, to be replaced by a random number Au of offspring where Au is taken from the size-biased distribution Q T .Au D k/ D kP .A D k/.m C 1/1 , k D 0; 1; : : : (note relative to A, given by Q that this distribution does not depend on T ). All other particles, once born, behave exactly as they would under P: they move like independent standard Brownian motions, die at the usual rate r, and give birth to a number of particles that is distributed like A. Proof. Most of this is standard in the spine literature; for example proof can be found in [4]. We will not use the precise drift of the spine except for the fact that the spine remains within the tube: to see this note that since the event is GQT -measurable, Q T .9t T W jt Tf .t=T /j > "T / D EŒ Q T .T /½f9t T Wj Tf .t =T /j>"T g D 0 Q t t u
by the definition of T .T /. Another important tool in this section is the spine decomposition. Q T -almost surely, Lemma 3 (Spine decomposition). Q Q T ŒZT .t/jGQT D Q
X
.Au 1/VT .Su /e rmSu C VT .t/e rmt
u<t
where we recall that fu < t g is the set of ancestors of the spine particle at time t, and Su denotes the time at which particle u split into two new particles. A proof of the spine decomposition may be found in [4]. Lemma 4. If f 2 C 2 Œ0; 1 then for any u 2 NT .t/, almost surely under both PQ and Q T we have Q ˇZ t ˇ Z Z t ˇ ˇ 0 2 ˇ f 0 .s=T /dXu .s/ ˇ 2"T f .s=T / ds ˇ ˇ 0
0
t =T
jf 00 .s/jds C "T jf 0 .0/j:
0
Proof. From the integration by parts formula for Itô calculus (since for any particle Q we know that for any u 2 N.t/, .Xu .s/; 0 s t/ is a Brownian motion under P) Q g 2 C 2 Œ0; 1 with g.0/ D 0, under P, 0
Z
t
g .t/Xu .t/ D
00
Z
t
g .s/Xu .s/ds C
0
From ordinary integration by parts,
0
g 0 .s/dXu .s/:
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
Z
t
0
Z
0
t
g .s/ ds D g .t/g.t/ 2
0
381
g.s/g 00 .s/ds:
0
Now set g.t/ D Tf .t=T / for t 2 Œ0; T . We note that, if u 2 NT .t/ then jXu .s/ g.s/j < "T for all s t. Thus ˇZ t ˇ Z t ˇ ˇ 0 2 ˇ f 0 .s=T /dXu .s/ f .s=T / ds ˇˇ ˇ 0 0 ˇ ˇZ t Z t ˇ ˇ g 0 .s/2 ds ˇˇ D ˇˇ g 0 .s/dXu .s/ 0
0
ˇ ˇ Z t ˇ ˇ ˇˇ g 0 .t/.Xu .t/ g.t// g 00 .s/.Xu .s/ g.s//ds ˇˇ 0 Z t jg 0 .t/ g 0 .0/j"T C jg 0 .0/j"T jg 00 .s/j"T ds 0
Z
t
2"T
jg 00 .s/jds C "T jg 0 .0/j
0
Z
t =T
D 2"T
jf 00 .s/jds C "T jf 0 .0/j
0
Q T P, Q almost surely under Q Q T. almost surely under PQ and, since Q
t u
We now use this result to give approximations on ZT .t/ under certain conditions. One of these conditions involves the seemingly unnatural assumption f 0 .0/ D 0. This is caused by the fact that in this section we make no approximations to the Q T except for using that it always remains within "T of our path of the spine under Q T -rescaled path—hence we are left with a rather bad estimate on its path at small times, where it will not get anywhere near "T . This does not matter to us, however, precisely because of this freedom to move within the "-tube about f : if f 0 .0/ ¤ 0 then we may choose g near to f (in an appropriate way; certainly within the "-tube) such that g 0 .0/ D 0. This issue arises in Lemma 8 and rigorous details are given there. R Lemma 5. If f 2 C 2 Œ0; 1, f 0 .0/ D 0 and rm > 12 0 f 0 .s/2 ds for all 2 .0; , then for small enough " > 0 and any T > 0 and t T , there exists > 0 such that X 2 2 2 2 Q T ŒZT .t/jGQT Q .Au 1/e =8" T Su C e =8" T t u<T
Q T -almost surely. Q Proof. Since rm > 12 choose > 0 such that
R 0
f 0 .s/2 ds for all 2 .0; and f 0 .0/ D 0, we may
382
S.C. Harris and M.I. Roberts
2 rm
1 2
Z
f 0 .s/2 ds 8 2 Œ0; :
0
Then for any " > 0 satisfying Z
2"
jf 00 .s/jds 8 2 Œ0;
0
Q T the spine we have, by Lemma 4 (since f 0 .0/ D 0 and using the fact that under Q is always in NT .t/), VT .t/e rmt e
2 =8"2 T rmt C T 2
R t =T 0
f 0 .s/2 dsC2"T
R t =T 0
jf 00 .s/jds
e
2 =8"2 T t
for all t 2 Œ0; T . Plugging this into the spine decomposition, we get Q T ŒZT .t/jGQT Q
X
.Au 1/e
2 =8"2 T S
u
Ce
2 =8"2 T t
:
t u
u<T
R Proposition 1. If f 2 C 2 Œ0; 1, f 0 .0/ D 0 and rm > 12 0 f 0 .s/2 ds for all 2 .0; , then for small enough " > 0 the set fZT .t/ W T 1; t T g is uniformly integrable under P. Proof. Fix ı > 0. We first claim that there exists K such that sup T 1; t T
Q T ŒZT .t/jGQT > K/ < ı=2: Q T .Q Q
To see this, take an auxiliary probability space with probability measure Q, and on this space consider a sequence A1 ; A2 ; : : : of independent and identically distributed random variables satisfying Q.Ai D k/ D
kP.A D k/ mC1
QT so that the Ai have the same distribution as births Au along the spine under Q (recall that there is no dependence on T ). Take also a sequence e1 ; e2 ; : : : of independent random variables that are exponentially distributed with parameter r.m C 1/; then set Sn D e1 C : : : C en (so that the random variables Sn have Q T ). By Lemma 5 we the same distribution as the birth times along the spine under Q have 0 1 1 X Q T ŒZT .t/jGQT > K/ Q @ .Aj 1/e 2 =8"2 Sj C e 2 =8"2 > K A : Q T .Q sup Q T 1 t T
j D1
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
383
P Sj Hence our claim holds if the random variable 1 can be shown to j D1 .Aj 1/e be Q-almost surely finite. Now for any 2 .0; 1/, Q.
X .An 1/e Sn D 1/ Q.An e Sn > n infinitely often/ n
Q
Sn log An > log C infinitely often : n n
By the strong law of large numbers, Sn =n ! 1=r.m C 1/ almost surely under Q; so if 2 .exp.=r.m C 1//; 1/ then the quantity above is no larger than log An Q lim sup >0 : n n!1 But this quantity is zero by Borel-Cantelli: indeed, for any T , X n
Q
X log An >" D Q.log A1 > "n/ n n Z 1 log A1 Q.log A1 "x/dx D Q " 0
which is finite for any " > 0 since (by direct calculation from the distribution of Q log A < 1 (this was one of our assumptions at the A1 under Q) QŒlog A1 D PŒA beginning of the article). Thus our claim holds. Now choose M > 0 such that 1=M < ı=2; then for K chosen as above, and any T 1, t T , Q T .ZT .t/ > MK; Q Q T ŒZT .t/jGQT K/ Q T .ZT .t/ > MK/ Q Q Q T ŒZT .t/jGQT > K/ Q T .Q CQ Q T ZT .t/ ½ Q C ı=2 Q Q MK fQT ŒZT .t /jGT Kg " # Q T ŒZT .t/jGQT Q Q D QT ½fQQ T ŒZT .t /jGQT Kg C ı=2 MK 1=M C ı=2 ı: Thus, setting K 0 D MK, for any T 1, t T , Q T .ZT .t/ > K 0 / ı: PŒZT .t/½fZT .t />K 0g D Q Since ı > 0 was arbitrary, the proof is complete.
t u
384
S.C. Harris and M.I. Roberts
As our final result in this section we link explicitly the martingales ZT with the number of particles NT . Lemma 6. For any ı > 0, if f 2 C 2 Œ0; 1, f .0/ D 0 and " is small enough then T 2 rmT C ZT .T / jNT .f; "; /j exp 8"2 T 2
Z
!
0
f .s/ ds C ıT 2
:
0
Proof. Simply plugging the result of Lemma 4 into the definition of ZT .T / gives the desired inequality. t u We note here that, in fact, a similar bound can be given in the opposite direction, so that NT .f; "=2; / is dominated by ZT .T / multiplied by some deterministic function of T . We will not need this bound, but it is interesting to note that the study of the martingales ZT is in a sense equivalent to the study of the number of particles NT .
3 The Lower Bound 3.1 The Heuristic for the Lower Bound We want to show that NT .f; "; / cannot be too small for large T . For f 2 C Œ0; 1 and 2 Œ0; 1, define ( J.f; / WD
rm 1
1 2
R 0
f 0 .s/2 ds if f 2 H1 otherwise.
We note that J resembles our rate function K, but without the truncation at the extinction time 0 . We shall work mostly with the simpler object J , before deducing our result involving K at the very last step. We now give a short heuristic to describe our route through the proof of the lower bound. Step 1. Consider a small (relative to T ) time T . How many particles are in NT .f; "; /? If is much smaller than ", then (with high probability) no particle has had enough time to reach anywhere near the edge of the tube (approximately distance "T from the origin) before time T . Thus, with high probability, jNT .f; "; /j D jN.T /j exp.rmT /: Step 2. Given their positions at time T , the particles in NT .f; "; / act independently. Each particle u in this set thus draws out an independent branching Brownian motion. Let NT .u; f; "; / be the set of descendants of u that are in NT .f; "; /. How big is this set? Since is very small, each particle u is close to
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
385
the origin. Thus we may hope to find some q < 1 such that P .jNT .u; f; "; /j < exp.J.f; /T ıT // q: (Of course, in reality we believe that this quantity will be exponentially small—but to begin with, the constant bound can be shown more readily.) Step 3. If NT .f; "; / is to be small, then each of the sets NT .u; f; "; / for u 2 NT .f; "; / must be small. Thus P .jNT .f; "; /j < exp.J.f; /T ıT // . q
exp.rmT /
;
and we may apply Borel-Cantelli to deduce our result along lattice times (that is, times Tj , j 0 such that there exists > 0 with Tj Tj 1 D for all j 1). Step 4. We carry out a simple tube-reduction argument to move to continuous time. The idea here is that if the result were true on lattice times but not in continuous time, the number of particles in NT .f; "; / must fall dramatically at infinitely many non-lattice times. We simply rule out this possibility using standard properties of Brownian motion. The most difficult part of the proof is Step 2. However, the spine results of Sect. 2 will simplify our task significantly.
3.2 The Proof of the Lower Bound We begin with Step 1 of our heuristic, considering the size of NT .f; "; / for small . Lemma 7. For any continuous f with f .0/ D 0 and any " > 0, there exist > 0, k > 0 and T1 such that P.9u 2 N.T / W u 62 NT .f; "=2; // e kT
8T T1 :
Proof. Choose small enough that sups2Œ0; jf .s/j < "=4. Then, using the manyto-one lemma and standard properties of Brownian motion, P.9u 2 N.T / W u 62 NT .f; "=2; //
!
D P 9u 2 N.T / W sup jXu .sT / Tf .s/j "T =2 s
2 P4
3
X
½fsups jXu .sT /Tf .s/j"T =2g 5
u2N.T /
e
rmT
PQ sup jsT Tf .s/j "T =2 s
!
386
S.C. Harris and M.I. Roberts
! e
rmT
PQ sup jsT j "T =4 s
p 2 16 e rmT " T =32 p : " 2T A suitably small choice of gives the exponential decay required.
t u
We now move on to Step 2, using the results of Sect. 2 to bound the probability of having a small number of particles strictly below 1. The bound given is extremely crude, and there is much room for manoeuvre in the proof, but any improvement would only add unnecessary detail. Lemma 8. If f 2 C 2 Œ0; 1 and J.f; s/ > 0 8s 2 .0; , then for any " > 0 and ı > 0 there exists T0 0 and q < 1 such that
P jNT .f; "; /j < e J.f; /T ıT q
8T T0 :
Proof. Note that by Lemma 6 for small enough " > 0 and large enough T , jNT .f; "; /je J.f; /T CıT =2 ZT .T / and hence
P jNT .f; "; /j < e J.f; /T ıT P ZT .T / < e ıT =2 : Suppose first that f 0 .0/ D 0. Then, again for small enough ", by Proposition 1 the set fZT .T /; T 1; t 2 Œ1; T g is uniformly integrable. Thus we may choose K such that sup EŒZT .T /½fZT .T />Kg 1=4; T 1
and then 1 D EŒZT .T / D EŒZT .T /½fZT .T /1=2g C EŒZT .T /½f1=2Kg 1=2 C KP.ZT .T / > 1=2/ C 1=4 so that P.ZT .T / > 1=2/ 1=4K: Hence for large enough T ,
P jNT .f; "; /j < e J.f; /T ıT 1 1=4K:
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
387
This is true for all small " > 0; but increasing " only increases jNT .f; "; /j so the statement holds for all " > 0. Finally, if f 0 .0/ ¤ 0 then choose g 2 C 2 Œ0; 1 such that g.0/ D g 0 .0/ D 0, sups jf gj "=2, J.g; / > 0 for all and J.g; / > J.f; / ı=2 (for small , the function g.t/ WD
f .t/ C at C bt 2 C ct 3 C dt 4 if t 2 Œ0; / ; f .t/ if t 2 Œ; 1
with a D f 0 .0/, b D 3f 0 .0/=, c D 3f 0 .0/=2 and d D f 0 .0/=3 , will work). Then as above we may choose K such that P.jNT .f; "; /j < e J.f; /T ıT / P.jNT .g; "=2; /j < e J.g; /T ıT =2 / 1 1=4K t u
as required.
Our next result runs along integer times—these times are sufficient for our needs, although the following proof would in fact work for any lattice times. Proposition 2. Suppose that f 2 C 2 Œ0; 1 and J.f; s/ > 0 8s 2 .0; . Then lim inf j !1 j 2N
1 log jNj .f; "; /j J.f; / j
almost surely. Proof. For any particle u, define NT .u; f; "; / WD fv 2 N.T / W u v; jXv .t/ Tf .t=T /j < "T 8t 2 Œ0; T g D fv W u vg \ NT .f; "; /; the set of descendants of u that are in NT .f; "; /. Then for ı > 0 and 2 Œ0; , ˇ
P jNT .f; "; /j < e J.f; /T ıT ˇ FT Y ˇ
P jNT .u; f; "; /j < e J.f; /T ıT ˇ FT u2NT .f;"=2;/
Y
P jNT .g; "=2; /j < e J.f; /T ıT
u2NT .f;"=2;/
since fjNT .u; f; "; /j W u 2 NT .f; "=2; /g are independent random variables, and where g W Œ0; 1 ! R is any twice continuously differentiable extension of the function gN W Œ0; ! R t ! f .t C / f ./:
388
S.C. Harris and M.I. Roberts
If is small enough, then jJ.f; / J.g; /j < ı=2 and J.g; s/ > 0
8s 2 .0; :
Hence, applying Lemma 8, there exists q < 1 such that for all large T ,
P jNT .g; "=2; /j < e J.f; /T ıT
P jNT .g; "=2; /j < e J.g; /T ıT =2 q: Thus for large T , ˇ
P jNT .f; "; /j < e J.f; /T ıT ˇ FT q jNT .f;"=2;/j :
(1)
Now, recalling that N.t/ is the total number of particles alive at time t, it is wellknown (and easy to calculate) that for ˛ 2 .0; 1/, i h E ˛ jN.t /j
˛ ˛ C .1 ˛/e rt
(in fact this is exactly EŒ˛ jN.t /j in the case of strictly dyadic branching). Taking expectations in (1), and then applying Lemma 7, for small we can get
P jNT .f; "; /j < e J.f; /T ıT
i h P .9u 2 N.T / W u 62 NT .f; "=2; // C E q jN.T /j e kT C
q q C .1 q/e rT
for some k > 0 and all large enough T . The Borel-Cantelli lemma now tells us that 1 P lim inf log jNj .f; "; /j < J.f; / ı D 0; j !1 j and taking a union over ı > 0 gives the result. jN.t /j
t u
may not hold if we allowed the possibility We note that our estimate on EŒ˛ of death with no offspring. In this case a more sophisticated estimate is required, taking into account the probability that the process becomes extinct. We look now at moving to continuous time using Step 4 of our heuristic. For simplicity of notation, we break with convention by defining kf k WD sup jf .s/j s2Œ0;
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
389
for f 2 C Œ0; or f 2 C Œ0; 1 (on this latter space, k k is not a norm, but this will not matter to us). Proposition 3. Suppose that f 2 C 2 Œ0; 1 and J.f; s/ > 0 8s 2 .0; . Then lim inf T !1
1 log jNT .f; "; /j J.f; / T
almost surely. Proof. We claim first that for large enough j 2 N,
jNj .f; "; /j >
inf
t 2Œj;j C1
jNt .f; 2"; /j
(
"j 9u 2 N..j C 1// W sup jXu .t/ Xu .j /j > 2 t 2Œj;j C1
) :
Indeed, if v 2 Nj .f; "; /, t 2 Œj; j C 1 and s 2 Œ0; t then for any descendant u of v at time t, jXu .s/ tf .s=t/j jXu .s/ Xu .s ^ j /j C jXu .s ^ j / jf ..s ^ j /=j /j C jjf ..s ^ j /=j / jf .s=t/j C jjf .s=t/ tf .s=t/j jXu .s/ Xu .s ^ j /j C "j Cj
jf .x/ f .y/j C kf k
sup x;y2Œ0; jxyj1=j
jXu .s/ Xu .s ^ j /j C
3" j 2
for large j;
so that if any particle is in Nj .f; "; / but not in Nt .f; 2"; / then it must satisfy sup jXu .s/ Xu .j /j "j=2:
j st
This is enough to establish the claim, and we deduce via the many-to-one lemma and standard properties of Brownian motion that P.jNj .f; "; /j >
inf
t 2Œj;j C1
jNt .f; 2"; /j/
P 9u 2 N..j C 1// W Q sup D e rm.j C1/P.
t 2Œj;j C1
! sup t 2Œj;j C1
jXu .t/ Xu .j /j "j=2
jt j j "j=2/
390
S.C. Harris and M.I. Roberts
"j
8 p
2
exp.rm.j C 1/ "2 j 2 =8/:
Since these probabilities are summable we may apply Borel-Cantelli to see that P.jNj .f; "; /j >
inf
t 2Œj;j C1
jNt .f; 2"; /j infinitely often/ D 0:
Now, 1 P lim inf log jNT .f; "; /j < J.f; / T !1 T 1 P lim inf log jNj .f; 2"; /j < J.f; / j !1 j inft 2Œj;j C1 jNt .f; "; /j <1 C P lim inf j !1 jNj .f; 2"; /j which is zero by Proposition 2 and Borel-Cantelli.
t u
If we were including the possibility of death with no offspring then we would have to check that no particles in Nj .f; "; / managed to reach the outside of the slightly altered 2"-tube and then die before time j C 1. The only added difficulty would be in keeping track of notation. We are now in a position to give our lower bound in full. Corollary 1. For any open set U C Œ0; 1 and 2 Œ0; 1, we have lim inf T !1
1 log jNT .U; /j sup K.f; / T f 2U
almost surely. Proof. If supf 2U K.f; / D 1 then there is nothing to prove. Thus it suffices to consider the case when there exists f 2 U such that 0 .f /. Since U is open, in this case we can in fact find f 2 U such that J.f; s/ > 0 for all s 2 .0; (if J.f; / D 0 for some , just choose small enough that .1 /f 2 U ) and such that f is twice continuously differentiable on Œ0; 1 (twice continuously differentiable functions are dense in C Œ0; 1). Thus necessarily supg2U K.g; / > 0, and for any ı > 0 we may further assume (by a simple argument, for example by approximating with piecewise linear functions and then smoothing) that J.f; / > supg2U K.g; / ı. Again since U is open, we may take " such that B.f; "/ U ; then clearly for any T NT .f; "; / NT .U; /
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
391
so by Proposition 2 we have lim inf T !1
1 log NT .U; / sup K.g; / ı T g2U
almost surely, and by taking a union over ı > 0 we may deduce the result.
t u
4 The Upper Bound Our plan is as follows: we first carry out the simple task of obtaining a bound along lattice times (Proposition 4). We then move to continuous time in Lemma 9, at the cost of restricting to open balls about fixed paths, by a tube-expansion argument similar to the tube-reduction argument used in Proposition 3 of the lower bound. In Lemma 10 we then rule out the possibility of any particles following unusual paths, which allows us to restrict our attention to a compact set, and hence a finite number of small open balls about sensible paths. Finally we draw this work together in Proposition 5 to give the bound in continuous time for any closed set D. Our first task, then, is to establish an upper bound along integer times. As with the lower bound, these times are sufficient for our needs, although the following proof would work for any lattice times. In a slight abuse of notation, for D C Œ0; 1 and 2 Œ0; 1 we define J.D; / WD sup J.f; /: f 2D
Proposition 4. For any closed set D C Œ0; 1 and 2 Œ0; 1 we have lim sup j !1 j 2N
1 log jNj .D; /j J.D; / j
almost surely. Proof. From the upper bound for Schilder’s theorem (Theorem 5.1 of [7]) we have lim sup T !1
1 Q T 2 NT .D; // inf 1 log P. f 2D 2 T
Z
f 0 .s/2 ds:
0
Thus, by the many-to-one lemma, lim sup T !1
1 1 Q T 2 NT .D; // log E jNT .D; /j lim sup log e rmT P. T T !1 T Z 1 0 2 f .s/ ds rm inf f 2D 2 0 D J.D; /:
392
S.C. Harris and M.I. Roberts
Applying Markov’s inequality, for any ı > 0 we get
E jNT .D; /j 1 1 J.D; /T CıT lim sup log J.D; /T CıT ı lim sup log P jNT .D; /j e e T !1 T T !1 T so that
1 X
P jNj .D; /j e J.D; /j Cıj < 1 j D1
and hence by the Borel-Cantelli lemma 1 P lim sup log jNj .D; /j J.D; / C ı j j !1
! D 0: t u
Taking a union over ı > 0 now gives the result.
We note that the proof by Git [2] works up to this point; the rest of the proof of the upper bound will be concerned with plugging the gap in [2]. For D C Œ0; 1 and " > 0, let D " WD ff 2 C Œ0; 1 W inf kf gk "g: g2D
Recall that we defined NT .f; "; / WD NT .B.f; "/; /. Lemma 9. If D C Œ0; 1 and f 2 D, then lim sup T !1
1 log jNT .f; "; /j J.D 2" ; / T
almost surely. Proof. First note that 1 P lim sup log jNT .f; "; /j > J.D 2" ; / C ı T !1 T 1 P lim sup log jNj .f; 2"; /j > J.D 2" ; / j j !1
!
! 1 jNt .f; "; /j >ı : C P lim sup log sup j !1 j t 2Œj;j C1 jNj .f; 2"; /j Since f 2 D, the uniform closed ball of radius 2" about f is a subset of D 2" , so by Proposition 4,
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
393
! 1 2" P lim sup log jNj .f; 2"; /j > J.D ; / D 0 j !1 j and we may concentrate on the last term. We claim that for j large enough, for any t 2 Œj; j C 1 we have Nt .f; "; j=t/ Nj .f; 2"; /: Indeed, if u 2 Nt .f; "; j=t/ then for any s j , jXu .s/ jf .s=j /j jXu .s/ tf .s=t/j C jjf .s=j / tf .s=j /j C t jf .s=j / f .s=t/j t " C kf k C t
sup
jf .x/ f .y/j
x;y2Œ0; jxyj1=j
which is smaller than 2"j for large j since f is absolutely continuous. We deduce that for large j every particle in Nt .f; "; / for any t 2 Œj; j C 1 has an ancestor in Nj .f; 2"; /; thus, letting N.u; s; t/ be the set of all descendants (including, possibly, u itself) of particle u 2 N.s/ at time t, # jNt .f; "; /j E sup t 2Œj;j C1 jNj .f; 2"; /j " ˇ
# E supt 2Œj;j C1 jNt .f; "; /jˇ Fj E jNj .f; 2"; /j ˇ i3 2 h P ˇ E supt 2Œj;j C1 u2Nj .f;2"; / jN.u; j; t/jˇ Fj 5: E4 jNj .f; 2"; /j "
Since jN.u; j; t/j is non-decreasing in t, using the Markov property we get "
jNt .f; "; /j E sup t 2Œj;j C1 jNj .f; 2"; /j
#
"P u2Nj .f;2"; /
E
jNj .f; 2"; /j jNj .f; 2"; /jEŒjN./j DE jNj .f; 2"; /j
D exp.rm/: Hence by Markov’s inequality
ˇ
# E jN.u; j; .j C 1//jˇFj
394
S.C. Harris and M.I. Roberts
! jNt .f; "; /j > exp .ıj / exp .rm ıj / P sup t 2Œj;j C1 jNj .f; 2"; /j and applying Borel-Cantelli 1 jNt .f; "; /j >ı P lim sup log sup j jN j .f; 2"; /j j !1 t 2Œj;j C1
! D 0: t u
Again taking a union over ı > 0 gives the result.
If we were considering the possibility of particles dying with no offspring then N.u; j; t/ would not be non-decreasing in t, but considering instead the set of all descendants of u ever alive between times j and t would give us a slightly worse—but still good enough—estimate. We move now onto ruling out extreme paths, by choosing a “bad set” FN and showing that no particles follow paths in this set. There is a balance to be found between including enough paths in FN that C0 Œ0; 1 n FN is compact, but not so many that we might find some (rescaled) Brownian paths within FN at large times. For simplicity of notation we extend the definition of NT .D; / to sets D C Œ0; in the obvious way, setting NT .D; / WD fu 2 N.T / W 9f 2 D with Xu .t/ D Tf .t=T / 8t 2 Œ0; T g: Lemma 10. Fix 2 Œ0; 1. For N 2 N, let 1 1 : FN WD f 2 C Œ0; W 9n N; u; s 2 Œ0; with ju sj 2 ; jf .u/ f .s/j > p n n
Then for large N lim sup T !1
1 log jNT .FN ; /j D 1 T
almost surely. Proof. Fix T S 0; then for any t 2 ŒS; T ,
1 ft 2 Nt .FN ; /g D 9n N; u; s 2 Œ0; W ju sj 2 ; n 1 9n N; u; s 2 Œ0; W ju sj 2 ; n
ˇ ˇ ˇ ut st ˇ ˇ > p1 ˇ ˇ ˇ t n ˇ ˇ ˇ uT sT ˇ ˇ > p1 : ˇ ˇ ˇ S n
Since the right-hand side does not depend on t, we deduce that
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
f9t 2 ŒS; T W t 2 Nt .FN ; /g ˇ 1 ˇ uT sT 9n N; u; s 2 Œ0; W ju sj 2 ; ˇˇ n S
395
ˇ ˇ ˇ > p1 : ˇ n
Now, for s 2 Œ0; , define .n; s/ WD b2n2 sc=2n2 . Suppose p we have a continuous function f such that sups2Œ0; jf .s/ f ..n; s//j 1=4 n. If u; s 2 Œ0; satisfy ju sj 1=n2 , then jf .u/ f .s/j jf .u/ f ..n; u//j C jf .s/ f ..n; s//j C jf ..n; s// f ..n; u//j 1 1 2 1 p C p C p Dp : 4 n 4 n 4 n n Thus f9t 2 ŒS; T W t 2 Nt .FN ; /g 9n N; s
ˇ ˇ ˇ sT .n;s/T ˇ 1 ˇ> p W ˇˇ ˇ 4 n : S
Standard properties of Brownian motion now give us that
p Q P.9t 2 ŒS; T W t 2 Nt .FN ; // PQ 9n N; s W jsT .n;s/T j > S=4 n ! X p 2Q 2n P sup jsT j > S=4 n nN
s2Œ0;1=2n2
p X 8 n3 T S 2n : p exp 16T S nN Taking S D j and T D j C 1, we note that for large N , p X X 8 n3 T jN jn S 2n exp exp p exp 16T 32 64 S nN nN so that (again for large N ), Q P.9t 2 Œj; j C 1 W t 2 Nt .FN ; // exp.2rmj /: Applying Markov’s inequality and the many-to-one lemma,
396
S.C. Harris and M.I. Roberts
" P. sup
t 2Œj;j C1
jNt .FN ; /j 1/ E
# sup t 2Œj;j C1
2
jNt .FN ; /j
X
E4
3
½f9t 2Œj;j C1; vu W v2Nt .FN ; /g 5
u2N.j C1/
Q e rm.j C1/P.9t 2 Œj; j C 1 W t 2 Nt .FN ; // exp.rm.j C 1/ 2rmj /: Thus, by Borel-Cantelli, we have that for large enough N P.lim sup
sup
j !1 t 2Œj;j C1
jNt .FN ; /j 1/ D 0
and since jNT .FN ; /j is integer-valued, lim sup T !1
1 log jNT .FN ; /j D 1 T t u
almost surely.
Now that we have ruled out any extreme paths, we check that we can cover the remainder of our sets in a suitable way. Lemma 11. For 2 Œ0; 1, let C0 Œ0; WD ff 2 C Œ0; W f .0/ D 0g: For each N 2 N, the set C0 Œ0; n FN is totally bounded under k k (that is, it may be covered by open balls of arbitrarily small radius). Proof. Given " > 0 and N 2 N, choose n such that n N _ .1="2 /. p For any function f 2 C0 Œ0; n FN , if ju sj < 1=n2 then jf .u/ f .s/j 1= n ". Thus the set C0 Œ0; n FN is equicontinuous (and, since each function must start from 0, uniformly bounded) and we may apply the Arzelà-Ascoli theorem to say that C0 Œ0; n FN is relatively compact, which is equivalent to totally bounded since .C Œ0; ; k k / is a complete metric space. t u We are now in a position to give an upper bound for any closed set D in continuous time. This upper bound is not quite what we asked for in Theorem 1, but this issue—replacing J with K—will be corrected in Corollary 2. Proposition 5. If D C Œ0; 1 is closed, then for any 2 Œ0; 1 lim sup T !1
almost surely.
1 log jNT .D; /j J.D; / T
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
397
Proof. Clearly (since our first particle starts from 0) NT .D n C0 Œ0; 1; / D ; for all T , so we may assume without loss of generality that D C0 Œ0; 1. Now, for each , ( f 7!
1 2
R
1
0
f 0 .s/2 ds if f 2 H1 otherwise
is a good rate function on C0 Œ0; (that is, lower-semicontinuous with compact level sets): we refer to Sect. 5.2 of [1] but it is possible to give a proof by showing directly that the function is lower-semicontinuous, then applying Jensen’s inequality and the Arzelà-Ascoli theorem to prove that its level sets in C0 Œ0; 1 are compact. Hence we know that for any ı > 0, ff 2 C0 Œ0; W J.f; / J.D; / C ıg is compact, and since it is disjoint from ff 2 C0 Œ0; W 9g 2 D with f .s/ D g.s/ 8s 2 Œ0; g; which is closed, there is a positive distance between the two sets. Thus we may fix ı > 0 and choose " > 0 such that J.D 2" ; / < J.D; / C ı. Then, by Lemma 11, for any N we may choose a finite ˛ (depending on N ) and some fk , k D 1; 2; : : : ; ˛ such that balls of radius " about the fk cover C0 Œ0; n FN . Thus 1 P lim sup log jNT .D; /j > J.D; / C ı T !1 T 1 P lim sup log jNT .FN ; /j > J.D; / C ı T !1 T ˛ X 1 C P lim sup log jNT .fk ; "; /j > J.D 2" ; / : T !1 T kD1
By Lemmas 9 and 10, for large enough N the terms on the right-hand side are all zero. As usual we take a union over ı > 0 to complete the proof. u t Corollary 2. For any closed set D C Œ0; 1 and 2 Œ0; 1, we have lim sup T !1
1 log jNT .D; /j sup K.f; / T f 2D
almost surely. Proof. Since jNT .D; /j is integer valued, 1 1 log jNT .D; /j < 0 ) log jNT .D; /j D 1: T T
398
S.C. Harris and M.I. Roberts
Thus, by Proposition 4, if J.D; / < 0 then 1 P lim sup log jNT .D; /j > 1 D 0: T !1 T Further, clearly for and any T 0, if NT .D; / D ; then necessarily we have NT .D; / D ;. Thus if there exists with J.D; / < 0, then 1 P lim sup log jNT .D; /j > 1 D 0 T !1 T which completes the proof.
t u
Combining Corollary 1 with Corollary 2 completes the proof of Theorem 1.
Appendix: The Oversight in [2] In [2] it is written that under a certain assumption, setting 1 1 Wn D ! 2 ˝ W lim sup log jNT .D; /j > J.D; / C n T !1 T (it is not important what J.D; / is here) we have P.Wn / > 0 for some n. This is correct, but the article then goes on to say “It is now clear that lim sup T !1
1 1 log E jNT .D; /j J.D; / C ” T n
which does not appear to be obviously true. To see this explicitly, work on the probability space Œ0; 1 with Lebesgue probability measure P. Let XT , T 0 be the càdlàg random process defined (for ! 2 Œ0; 1 and T 0) by XT .!/ D
e 2T if T n 2 Œ! e 4T ; ! C e 4T / for some n 2 N e T otherwise.
Branching Brownian Motion: Almost Sure Growth Along Scaled Paths
399
Then for every !, lim sup but
1 log XT .!/ D 2 T
1 log EŒXT ! 1: T
Acknowledgements MIR was supported by an EPSRC studentship and by ANR MADCOF grant ANR-08-BLAN-0220-01.
References 1. A. Dembo, O. Zeitouni, Large Deviations Techniques and Applications. Applications of Mathematics (New York), vol. 38, 2nd edn. (Springer, New York, 1998) 2. Y. Git, Almost Sure Path Properties of Branching Diffusion Processes. In Séminaire de Probabilités, XXXII. Lecture Notes in Math., vol. 1686 (Springer, Berlin, 1998), pp. 108–127 3. R. Hardy, S.C. Harris, A conceptual approach to a path result for branching Brownian motion. Stoch. Process. Appl. 116(12), 1992–2013 (2006) 4. R. Hardy, S.C. Harris, A Spine Approach to Branching Diffusions with Applications to Lp Convergence of Martingales. In Séminaire de Probabilités, XLII. Lecture Notes in Math., vol. 1979 (Springer, Berlin, 2009) 5. S.C. Harris, M.I. Roberts. The unscaled paths of branching Brownian motion. Ann. Inst. Henri Poincaré Probab. Stat. (to appear) 6. T.-Y. Lee, Some large-deviation theorems for branching diffusions. Ann. Probab. 20(3), 1288– 1309 (1992) 7. S.R.S. Varadhan, Large Deviations and Applications. CBMS-NSF Regional Conference Series in Applied Mathematics, vol. 46 (Society for Industrial and Applied Mathematics (SIAM), Philadelphia, 1984)
On the Delocalized Phase of the Random Pinning Model Jean-Christophe Mourrat
Abstract We consider the model of a directed polymer pinned to a line of i.i.d. random charges, and focus on the interior of the delocalized phase. We first show that in this region, the partition function remains bounded. We then prove that for almost every environment of charges, the probability that the number of contact points in Œ0; n exceeds c log n tends to 0 as n tends to infinity. The proofs rely on recent results of Birkner, Greven, den Hollander (2010) and Cheliotis, den Hollander (2010).
1 Introduction Let D .i /i 2N be a sequence such that 0 D 0 and .i C1 i /i >0 are independent and identically distributed random variables with values in N D f1; 2; : : :g. Let P be the distribution of , E the associated expectation, and K.n/ D PŒ1 D n. We assume that there exists ˛ > 0 such that log K.n/ ! .1 C ˛/: log n n!1
(1)
As an example, one can think about the sequence as the sequence of arrival times at 0 of a one-dimensional simple random walk (and in this case, ˛ D 1=2). In a slight abuse of notation, we will look also at the sequence as a set, and write for instance n 2 instead of 9i W n D i . Let ! D .!k /k2N be independent and identically distributed random variables. We write P for the law of !, and E for the associated expectation. We will refer to J.-C. Mourrat () Ecole polytechnique fédérale de Lausanne, Institut de Mathématiques, Station 8, 1015 Lausanne, Switzerland e-mail: [email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__18, © Springer-Verlag Berlin Heidelberg 2012
401
402
J.-C. Mourrat
! as the environment. We assume that the !k are centred random variables, and that they have exponential moments of all order. Let ˇ > 0; h > 0, and n 2 N . We ˇ;h;! ˇ;h;! consider the probability measure Pn (expectation En ) which is defined as the following Gibbs transformation of the measure P: ! n1 ˇ;h;! X dPn 1 ./ D ˇ;h;! exp .ˇ!k h/1fk2 g 1fn2 g: dP Zn kD0 In the above definition, ˇ can be thought of as the inverse temperature, h as the ˇ;h;! disorder bias, and Zn is a normalization constant called the partition function, ! # n1 X D E exp .ˇ!k h/1fk2 g 1fn2 g : "
Znˇ;h;!
kD0
At the exponential scale, the asymptotic behaviour of the partition function is captured by the free energy F.ˇ; h/ defined as F .ˇ; h/
1 log Znˇ;h;! : n!C1 n
D lim
Superadditivity of the partition function implies that this limit is well defined almost surely, and that it is deterministic (see for instance [5, Theorem 4.1]). Assumption 1 implies that F.ˇ; h/ > 0. It is intuitively clear that the free energy can become strictly positive only if the set \ Œ0; n is likely to contain many points under the measure ˇ;h;! Pn . We thus say that we are in the localized phase if F.ˇ; h/ > 0, and in the delocalized phase otherwise. One can show [7, Theorem 11.3] that for every ˇ > 0, there exists hc .ˇ/ > 0 such that h < hc .ˇ/ ) localized phase, i.e. F.ˇ; h/ > 0; h > hc .ˇ/ ) delocalized phase, i.e. F.ˇ; h/ D 0; and moreover, the function ˇ 7! hc .ˇ/ is strictly increasing.
2 Statement of the Main Results We focus here on the interior of the delocalized phase, that is to say when h > hc .ˇ/. Note that, due to the strict monotonicity of the function hc ./, one sits indeed in the interior of the delocalized phase if one fixes h D hc .ˇ0 / and considers any inverse temperature ˇ < ˇ0 . By definition, the partition function is known to grow subexponentially in this region. In [2, Remark p. 417], the authors ask whether the partition function remains
On the Delocalized Phase of the Random Pinning Model
403
bounded there. We answer positively to this question, and can in fact be slightly more precise. Theorem 1. Let ˇ > 0 and h > hc .ˇ/. For almost every environment, one has C1 X
Znˇ;h;! < C1:
nD1
Remark 1. This result implies that, in the interior of the delocalized phase, the ˇ;h;! unconstrained (or free) partition function Zn;f is also almost surely bounded (in ˇ;h;!
fact, tends to 0) as n tends to infinity. Indeed, Zn;f
is defined by
"
ˇ;h;! Zn;f
n1 X D E exp .ˇ!k h/1fk2 g
!# ;
kD0
which is equal to C1 X n0 Dn
! # C1 n1 X X ˇ;h;! a.s. 0 0 E exp .ˇ!k h/1fk2 g I \ Œn; n D fn g 6 Zn0 ! 0: "
n0 Dn
kD0
n!1
Our second result concerns the size of the set \ Œ0; n, that we may call the ˇ;h;! set of contact points, under the measure Pn . Let us write En;N for the event that j \ Œ0; nj > N (where we write jAj for the cardinal of a set A). Theorem 2. Let ˇ > 0 and h > hc .ˇ/. For every " > 0 and for almost every environment, there exists N" ; C" > 0 such that for any N > N" and any n: .En;N / 6 Pˇ;h;! n
C" N.hhc .ˇ/"/ e : K.n/
In particular, for every constant c such that c>
1C˛ h hc .ˇ/
and for almost every environment, one has .En;c log n / ! 0: Pˇ;h;! n n!1
To my knowledge, results of this kind were known only under the averaged measure ˇ;h;! PPn , and with some restrictions on the distribution of ! due to the use of concentration arguments (see [6] or [5, Sect. 8.2]). In particular, in the interior of
404
J.-C. Mourrat
the delocalized phase and for almost every environment, the polymer intersects the pinning line less that the simple random walk does. It is worth comparing this result with the case when randomness of the medium is absent, that is, when ˇ D 0. In this context, the distribution of the number of contact points of the polymer forms a tight sequence as n varies (see for instance [7, Theorem 7.3]). It is only natural to expect that a similar result holds true in the disordered case as well. Interestingly, boundedness of the number of contact points in the delocalized phase was recently obtained for a specific model of pinning on a random interface with long-range correlations in [1], even at criticality (h D hc .ˇ/). In this work, the specific structure of the environment enables the authors to identify the critical point explicitly, a feature which makes the subsequent analysis more tractable.
3 Proofs In this section, we present the proofs of Theorems 1 and 2. Although one might think at first that such an approach cannot be of much help as far as the delocalized phase is concerned, we will rely on recent results obtained in [3, 4], where the authors develop a large deviations point of view of the problem. Let us define ˇ;h;!
FN
X
D
N 1 Y
K.li C1 li /e .ˇ!li h/ :
(2)
0Dl0
Our results are based on the following fact, due to [4], that holds both in the delocalized and in the localized phases. Lemma 1. For almost every environment, one has 1 ˇ;h;! log FN D hc .ˇ/ h: N N !C1
lim sup
Proof. Although this result is not stated as a proposition in [4], the authors give all the necessary elements to prove it. Indeed, we can start from [4, (3.11)], which reads 1 ˇ;h;! log FN D h C S que .ˇI 1/; N !C1 N
lim sup
where S que .ˇI z/ is defined in [4, (3.10)]. We then learn from [4, (3.13)] that hc .ˇ/ D S que .ˇI 1 /; so what remains to see is that
On the Delocalized Phase of the Random Pinning Model
405
S que .ˇI 1 / D S que .ˇI 1/: The proof of this fact is first obtained assuming further that the support of the distribution of !0 is finite, see [4, Lemma 3.2]. The general case, where one assumes only finiteness of all exponential moments, is considered in [4, Sect. 3.3]. One can start by observing that, since z 7! S que .ˇ; z/ is increasing (in the wide sense), one has S que .ˇI 1 / 6 S que .ˇI 1/. On the other hand, it is shown in step 1 of the proof of [4, Lemma 3.3] that S que .ˇI 1/ 6 A.ˇ/, where A.ˇ/ is defined in [4, (3.21)]. Moreover, steps 2–4 of the proof of [4, Lemma 3.3] are devoted to the justification of the fact that A.ˇ/ 6 S que .ˇI 1 /. It thus follows that S que .ˇI 1/ 6 S que .ˇI 1 /, which finishes the proof. t u Proof of Theorem 1. The proof is close to [4, Sect. 3.2]. We can decompose Zn the following way: Znˇ;h;! D
C1 X
X
N 1 Y
K.li C1 li /e .ˇ!li h/ :
N D1 0Dl0
An interversion of sums then leads to C1 X
Znˇ;h;! D
nD1
C1 X
ˇ;h;!
FN
;
N D1
and Lemma 1 ensures the almost sure convergence of the second series when h > hc .ˇ/. t u ˇ;h;!
For an event A, let us write Zn
.A/ for the quantity
# ! n1 X .ˇ!k h/1fk2 g 1fn2 g I A : E exp "
kD0 ˇ;h;!
In words, Zn .A/ is a partition function in which one integrates with respect to P only on the event A. In order to prove Theorem 2, we first give a refined version of Theorem 1, which goes as follows. Proposition 1. Let ˇ > 0 and h > hc .ˇ/. For every " > 0 and for almost every environment, there exist N" ; C" such that for any N > N" : C1 X
Znˇ;h;! .En;N / 6 C" e N.hhc .ˇ/"/ :
nD1
Proof. We can assume that " < h hc .ˇ/. Note that, for any n and N0 ,
406
J.-C. Mourrat
Znˇ;h;! .En;N0 / D
C1 X
N 1 Y
X
K.li C1 li /e .ˇ!li h/ :
N DN0 0Dl0
By an interversion of sums, we obtain that C1 X
Znˇ;h;! .En;N0 /
C1 X
D
nD1
ˇ;h;!
FN
:
N DN0
By Lemma 1, there exists N" such that for every N > N" , ˇ;h;!
FN
6 e N.hhc .ˇ/"=2/ ;
and as a consequence, for every N0 > N" , one has C1 X
C1 X
Znˇ;h;! .En;N0 / 6
nD1
e N.hhc .ˇ/"=2/ ;
N DN0
t u
which implies the announced claim. Proof of Theorem 2. Note that Pˇ;h;! .En;N / D
ˇ;h;!
Zn
.En;N / : ˇ;h;! Zn
The numerator can be bounded from above using Proposition 1. For the denominator, one can use the bound Znˇ;h;! > K.n/e ˇ!0 h ; which proves the desired result.
t u
References 1. Q. Berger, H. Lacoin, Sharp critical behavior for pinning model in random correlated environment. Preprint, To appear in Stochastic Processes and their Applications http://dx.doi.org/10. 1016/j.spa.2011.12.007, arXiv:1104.4969v1 (2011) 2. M. Birkner, R. Sun, Annealed vs quenched critical points for a random walk pinning model. Ann. Inst. Henri Poincaré Probab. Stat. 46(2), 414–441 (2010) 3. M. Birkner, A. Greven, F. den Hollander, Quenched large deviation principle for words in a letter sequence. Probab. Theory Related Fields 148 (3–4), 403–456 (2010) 4. D. Cheliotis, F. den Hollander, Variational characterization of the critical curve for pinning of random polymers. Preprint, To appear at Annals of Probability, arXiv:1005.3661v1 (2010) 5. G. Giacomin, Random Polymer Models (Imperial College Press, London, 2007)
On the Delocalized Phase of the Random Pinning Model
407
6. G. Giacomin, F.L. Toninelli, Estimates on path delocalization for copolymers at selective interfaces. Probab. Theory Related Fields 133(4), 464–482 (2005) 7. F. den Hollander, in Random Polymers. Ecole d’été de probabilités de Saint Flour XXXVII. Lecture Notes in Mathematics, vol. 1974 (Springer, Berlin, 2009)
Large Deviations for Gaussian Stationary Processes and Semi-Classical Analysis Bernard Bercu, Jean-François Bony, and Vincent Bruneau
Abstract In this paper, we obtain a large deviation principle for quadratic forms of Gaussian stationary processes. It is established by the conjunction of a result of Roch and Silbermann on the spectrum of products of Toeplitz matrices together with the analysis of large deviations carried out by Gamboa, Rouault and the first author. An alternative proof of the needed result on Toeplitz matrices, based on semi-classical analysis, is also provided. Keywords Large deviations • Gaussian processes • Distribution of eigenvalues • Toeplitz matrices
1 Introduction For any bounded measurable real function f on the torus T D Œ; Œ, the `2 .N/ Toeplitz and Hankel operators are respectively defined as bi j T .f / D f
i;j 0
and
bi Cj C1 H.f / D f
i;j 0
(1)
bn / stands for the sequence of Fourier coefficients of f . We refer the reader where .f to the books of Böttcher and Silbermann [2,3] for a general presentation of Toeplitz operators. A well-known identity between the product T .f /T .g/ and T .fg/ is T .fg/ T .f /T .g/ D H.f /H.e g/
(2)
B. Bercu () J.-F. Bony V. Bruneau Institut de Mathématiques de Bordeaux, Université Bordeaux 1, UMR CNRS 5251, 351 cours de la libération, 33405 Talence cedex, France e-mail: [email protected]; [email protected]; [email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__19, © Springer-Verlag Berlin Heidelberg 2012
409
410
B. Bercu et al.
where e g .x/ D g.x/. The analogue of identity (2) for finite section Toeplitz matrices is given by the formula of Widom [14] e/H.g/Qn Tn .fg/ Tn .f /Tn .g/ D Pn H.f /H.e g /Pn C Qn H.f
(3)
where the projection Pn and the operator Qn are given by Pn .x0 ; x1 ; x2 ; : : :/ D .x0 ; x1 ; : : : ; xn ; 0; : : :/; Qn .x0 ; x1 ; x2 ; : : :/ D .xn ; xn1 ; : : : ; x0 ; 0; : : :/; and Tn .f / is the finite section of order n 1 of T .f / which means that Tn .f / is identified with Pn T .f /Pn . In other words, our operators will be considered as operators on Im P and Im Pn where P stands for the projection operator on `2 .N/. e/. We clearly have Qn2 D Pn , Pn Qn D Qn Pn D Qn , and Qn Tn .f /Qn D Tn .f The classical Szegö theorem deals with the asymptotic behavior of the spectrum of a single Toeplitz matrix. It states that if f is a bounded measurable real function on T, the limiting set of eigenvalues of the sequence .Tn .f // is exactly .T .f // D Œessinff; esssupf ; where .T .f // denotes the spectrum of the operator T .f /. Moreover, the empirical spectral measure of .Tn .f // converges to Pf which is the image probability of the uniform measure on T by the application f . In other words, if n0 ; : : : ; nn are the eigenvalues of Tn .f /, then for any bounded continuous real function ' 1 1X '.nk / D n!1 n 2 n
Z '.f .x// dx:
lim
kD0
(4)
T
In particular, the maximum eigenvalue of Tn .f / converges to esssupf while the minimum eigenvalue of Tn .f / converges to essinff . One can find more details in Sect. 5.2 of [8] or in Sect. 5.4 of [2]. Our purpose is to make use of similar results for the spectrum of the product of two Toeplitz matrices Tn .f /Tn .g/. Several authors have investigated the asymptotic behavior of the spectrum of Tn .f /Tn .g/. More precisely, it was shown in Lemma 5 of [1] or Lemma 2.6 of [13] that if f and g are two bounded measurable real functions on T, then the empirical spectral measure associated with the sequence .Tn .f /Tn .g// converges to the limiting measure Pfg . However, the limiting set of eigenvalues of .Tn .f /Tn .g// is much more difficult to understand. Via a theorem of Roch and Silbermann, we shall see that, as soon as f and g 0 are bounded piecewise continuous real functions, the limiting set of eigenvalues of .Tn .f /Tn .g// coincides with the spectrum of the limiting operator T .f /T .g/. In particular, the maximum and the minimum eigenvalues of Tn .f /Tn .g/ both converge to the maximum and minimum of the spectrum of T .f /T .g/. In this paper, we make use of the previous results on Toeplitz operators to obtain a large deviation principle (LDP) for quadratic forms of Gaussian stationary
Large Deviations for Gaussian Stationary Processes and Semi-Classical Analysis
411
processes. More precisely, consider a centered stationary real Gaussian process .Xn / with bounded piecewise continuous spectral density g. It was shown in [1] an LDP for subsequences of the empirical periodogram .Wn .f // integrated over a bounded piecewise continuous real function f . We can now deduce a full LDP for the sequence .Wn .f //. We also give an alternative proof of the theorem of Roch and Silbermann in the particular case of Toeplitz operators with continuous symbols. Our approach is based on semi-classical analysis and scattering theory by construction of quasimodes which are approximative eigenvectors. We hope that this microlocal approach can be used in other situations. The paper is organized as follows. In Sect. 2, we recall a theorem of Roch and Silbermann. Sect. 3 is devoted to the application in probability. An enlightening example is treated in Sect. 4. Then, we give our alternative proof of the result of Roch and Silbermann in the case of Toeplitz operators with continuous symbols. This result and our functional point of view on Toeplitz operators are given in Sect. 5. The convergence of the spectrum is proved in Sect. 6. Finally, in Sect. 7, we propose an alternative proof of Coburn’s theorem dealing with the essential spectrum of products of Toeplitz operators.
2 Results on Toeplitz Operators Denote by A the Banach algebra of all sequences .An / of uniformly bounded linear operators on Im Pn endowed with the sum and the composition term by term, and the supremum of the operator norm of the elements. Let B be the collection of all sequences .An / of A for which one can find two bounded linear operators A and e A in Im P such that An ! A;
An ! A ;
Qn An Qn ! e A;
Qn An Qn ! e A ;
where stands the adjoint operator and ! stands for the strong convergence. Finally, denote by C the smallest closed subalgebra of A containing the collection of all sequences .Tn .f // where f are bounded piecewise continuous real functions. In fact, C is a subalgebra of B and Tn .f / ! T .f /;
e/: Qn Tn .f /Qn ! T .f
We refer to Sect. 2.5 of [2] for more details on B. We are now in position to state a theorem of Roch and Silbermann. Theorem 1 (Roch–Silbermann). Let .Tn / be a sequence of selfadjoint operators of C . Moreover, denote the strong limits of Tn and Qn Tn Qn by T and e T, respectively. For 2 R, the following properties are equivalent: e/, (i) 2 .T / [ .T (ii) is the limit of a sequence .n / where n 2 .Tn /, (iii) is the limit of a subsequence .nk / where nk 2 .Tnk /.
412
B. Bercu et al.
Theorem 1 was established in [12] together with several examples of application. It is given, in its present form, in Theorem 4.16 of [2]. A direct application of this result is as follows. First of all, let us introduce some notations. Let f and g be two bounded piecewise continuous real functions with g 0. From Lemma 3 below, the sequence .Tn .g/1=2 / as well as .Tn .g/1=2 Tn .f /Tn .g/1=2 / belong to C , Tn .g/1=2 Tn .f /Tn .g/1=2 ! T .g/1=2 T .f /T .g/1=2 ; e/T .e g /1=2 T .f g /1=2 : Qn Tn .g/1=2 Tn .f /Tn .g/1=2 Qn ! T .e On Im Pn , we clearly have Tn .f /Tn .g/ D Tn .g/1=2 Tn .f /Tn .g/1=2 ; with the same multiplicity. Moreover, by Lemma 7, we also have on Im P e/T .e T .g/1=2 T .f /T .g/1=2 D T .f /T .g/ D T .f g/ e/T .e D T .e g /1=2 T .f g /1=2 : Denote the maximum and minimum eigenvalues of Tn .f /Tn .g/ by nmax .f; g/ D max Tn .f /Tn .g/ ; nmin .f; g/ D min Tn .f /Tn .g/ : In addition, denote the extrema of the spectrum of T .f /T .g/ by max .f; g/ D max T .f /T .g/ ; min .f; g/ D min T .f /T .g/ : One can observe that, in general, we do not know if max .f; g/ and min .f; g/ are eigenvalues. Corollary 1. Assume that f and g are two bounded piecewise continuous real functions on T with g 0. Then, the limiting sets of eigenvalues of the sequence .Tn .f /Tn .g// are given by .T .f /T .g//. In particular, lim nmax .f; g/ D max .f; g/;
(5)
lim nmin .f; g/ D min .f; g/:
(6)
n!1
n!1
In Sect. 4, we shall show via an example related to Gaussian autoregressive process that it is not true in general that for two bounded continuous real functions f and g, max .f; g/ D sup.fg/ or min .f; g/ D inf.fg/. One can also observe that the
Large Deviations for Gaussian Stationary Processes and Semi-Classical Analysis
413
norm of T .g/1=2 T .f /T .g/1=2 is not always equal to kfgk1 or kf k1 kgk1 . The situation is totally different from the case of a single Toeplitz operator T .f / with bounded continuous real function as max .f; 1/ D sup.f / and min .f; 1/ D inf.f /.
3 Application in Probability Let .Xn / be a centered stationary real Gaussian process with bounded piecewise continuous spectral density g 0 which means that Z 1 exp.i.j k/x/g.x/ dx: EŒXj Xk D 2 T We assume in all the sequel that g is not the zero function. For any bounded piecewise continuous real function f on the torus T, we are interested in the asymptotic behavior of ˇ n ˇ2 Z ˇX ˇ 1 Wn .f / D f .x/ˇˇ Xj exp.ijx/ˇˇ dx: (7) 2 n T j D0 The purpose of this section is to provide the last step in the analysis of the large deviation properties of .Wn .f // by establishing an LDP for .Wn .f // in the spirit of the original work of [1] or of Bryc and Dembo [4]. We refer the reader to the book of Dembo and Zeitouni [6] for the general theory on large deviations. The covariance matrix associated with the vector X .n/ D .X0 ; : : : ; Xn /t is Tn .g/. Consequently, it immediately follows from (7) that Wn .f / D
1 .n/t 1 X Tn .f /X .n/ D Y .n/t Tn .g/1=2 Tn .f /Tn .g/1=2 Y .n/ n n
(8)
where the vector Y .n/ has a Gaussian N .0; In / distribution. In order to investigate the large deviation properties of .Wn .f //, it is necessary to calculate the normalized cumulant generating function given, for all t 2 R, by Ln .t/ D
1 log E exp.ntWn .f // : n
For convenience and in all the sequel, we use of the notation that log t D 1 if t 0. We deduce from (8) and standard Gaussian calculation that for all t 2 R 1 log det In 2tTn .g/1=2 Tn .f /Tn .g/1=2 2n n 1 X D log.1 2tnk /; 2n
Ln .t/ D
kD0
where n0 ; : : : ; nn are the eigenvalues of Tn .g/1=2 Tn .f /Tn .g/1=2 . For all t 2 R, let
414
B. Bercu et al.
Lfg .t/ D
1 4
Z log.1 2tf .x/g.x// dx; T
and denote by Ifg its Fenchel-Legendre transform ˚ Ifg .x/ D sup xt Lfg .t/ : t 2R
Furthermore, for all x 2 R, let 8 1 ˆ ˆ Ifg .a/ C .x a/ ˆ ˆ 2min .f; g/ ˆ ˆ < Jfg .x/ D Ifg .x/ ˆ ˆ ˆ ˆ 1 ˆ ˆ .x b/ : Ifg .b/ C 2max .f; g/
if x 2 1; a if x 2a; bŒ
(9)
if x 2 Œb; C1Œ
where a and b are the extended real numbers given by aD
L0fg
1 2min .f; g/
if min .f; g/ < 0 and min .f; g/ < inf.fg/, a D 1 otherwise, while bD
L0fg
1 2max .f; g/
if max .f; g/ > 0 and max .f; g/ > sup.fg/, b D C1 otherwise. We immediately deduce from Theorem 1 of [1] together with Corollary 1, that an LDP holds for .Wn .f //. Theorem 2. The sequence .Wn .f // satisfies an LDP with good rate function Jfg . More precisely, for any closed set F R lim sup n!1
1 log P.Wn .f / 2 F / inf Jfg .x/; x2F n
while for any open set G R lim inf n!1
1 log P.Wn .f / 2 G/ inf Jfg .x/: x2G n
Remark 1. Denote by the derivative of Lfg at point zero D
1 2
Z f .x/g.x/dx: T
Large Deviations for Gaussian Stationary Processes and Semi-Classical Analysis
415
Then, we have Jfg ./ D 0 and it follows from Theorem 2 that for all x > lim
n!1
1 log P.Wn .f / x/ D Jfg .x/; n
whereas for all x < 1 log P.Wn .f / x/ D Jfg .x/: n!1 n lim
4 An Illustrative Example Let a and be two real numbers with jj < 1 and consider the two bounded continuous real functions f and g given by f .x/ D a C cos.x/
g.x/ D
and
1C
2
1 : 2 cos.x/
The goal of this section is to study the limiting set of eigenvalues of the sequence .Tn .f /Tn .g//. We clearly have kf k1 D jaj C 1 and kgk1 D .1 jj/2 . The function g is simply the spectral density of a Gaussian autoregressive process [1]. If D 0, g D 1 and the product Tn .f /Tn .g/ reduces to Tn .f /. Consequently, max .f; 1/ D a C 1 and min .f; 1/ D a 1. If ¤ 0, denote a D
.1 C / 2
and
b D
.1 / : 2
It is more convenient to work with the inverse of Tn .g/. As a matter of fact, Tn .g/1 is a tridiagonal matrix quite similar to Tn .g 1 / except that, at the two diagonal corners of Tn .g 1 /, the coefficient 1 C 2 is replaced by 1 0
Tn .g/1
1 0 B 1 C 2 B B D B ::: ::: ::: B @ : : : 1 C 2 ::: 0
1 ::: ::: C C C ::: C: C A 1
It is not hard to see that det.Tn .g/1 / D 1 2 . In order to find the eigenvalues of the product Tn .f /Tn .g/, it is equivalent to calculate the zeros of its characteristic polynomial which correspond also to the zeros of det.Mn .t// where Mn .t/ D tTn .f / Tn .g/1
416
B. Bercu et al.
with t D 1=. As Tn .f / and Tn .g/1 are both tridiagonal matrices, we can easily compute det.Mn .t//. Via the same lines than in Lemma 11 of [1], we find that for n large enough, Mn .t/ is negative definite only on the domain D D D1 [ D2 with n o D1 D 2 2 < p 2 and q 2 < 4 2 .p C 2 / ; n o D2 D p < 2 2 and p < jqj ; where p D at .1C 2 / and q D t C2. In term of the variable , the inverses of the boundaries of D give the extrema of .T .f /T .g// that is max .f; g/ and min .f; g/. After some tedious but straightforward calculations, we obtain three inverses of the boundaries a1 ; .1 C /2
aC1 ; .1 /2
1 : 4.1 C a/
Two of them coincide with inf.fg/ and sup.fg/. It only depends on the location of a with respect to .1 C 2 /=.2/. The last one can be max .f; g/ > sup.fg/ or min .f; g/ < inf.fg/. It only depends on the sign of as well as on the location of a with respect to the interval Œa ; b . More precisely, if > 0 then max .f; g/ D sup.fg/ while min .f; g/ D
a1 1 aC1 < inf.fg/ D min ; 4.1 C a/ .1 C /2 .1 /2
if a 2a ; b Œ and min .f; g/ D inf.fg/ otherwise. Moreover, if < 0 then min .f; g/ D inf.fg/ while max .f; g/ D
a1 1 aC1 > sup.fg/ D max ; 4.1 C a/ .1 C /2 .1 /2
if a 2a ; b Œ and max .f; g/ D sup.fg/ otherwise.
5 Toeplitz Operators and Functional Calculus We will prove the following result which implies Corollary 1 for continuous functions. Theorem 3. Let f and g be two bounded continuous real functions with g 0. For 2 R, the following properties are equivalent: (i) 2 .T .f /T .g//, (ii) is the limit of a sequence .n / where n 2 .Tn .f /Tn .g//, (iii) is the limit of a subsequence .nk / where nk 2 .Tnk .f /Tnk .g//.
Large Deviations for Gaussian Stationary Processes and Semi-Classical Analysis
417
First, let us interpret the projection operators Pn and P as spectral projectors of the derivation operator and introduce the main ingredients of the proofs.
5.1 A Functional Point of View We consider the Toeplitz operators T .f / and Tn .f / as the cut-off, in frequencies, of the operator of multiplication by f . To be more precise, let us introduce the Fourier transform, F W L2 .T/ ! `2 .Z/, defined by .F u/k D b uk D
1 2
Z
u.x/e i kx dx:
The operator F is an isomorphism. We denote by F 1 its inverse, and we introduce b and P b n as the projections P b Wb P u 2 `2 .Z/ 7! .: : : ; 0; 0;b u0 ;b u1 ; : : :/ 2 `2 .Z/ b n Wb P u 2 `2 .Z/ 7! .: : : ; 0; 0;b u0 ;b u1 ; : : : ;b un ; 0; 0; : : :/ 2 `2 .Z/: On the other hand, if we identify f 2 L1 .T/ with L.f /, the bounded operator defined on L2 .T/ by u 2 L2 .T/ 7! f u 2 L2 .T/; we have T .f / D P f P
and
Tn .f / D Pn f Pn ;
b F and Pn D F 1 P b n F . In the following, we will systematically with P D F 1 P identify f with the operator L.f /. Since 1 d i kx .e / D ke i kx ; i dx the derivation operator D defined on d H 1 .T/ D u 2 L2 .T/I u 2 L2 .T/ D fu 2 L2 .T/I .k uO k /k 2 `2 .Z/g dx by D W u 2 H 1 .T/ 7!
1 d u 2 L2 .T/ i dx
is self-adjoint on L2 .T/ and F DF 1 is the diagonal operator .kık;j /k;j 2Z . For any bounded Borel function ', the bounded operator '.D/ is defined with the help of the functional calculus for self-adjoint operators. It satisfies '.D/ D F 1 M.'/F ;
418
B. Bercu et al.
where M.'/ is the operator b u 2 `2 .Z/ 7! .: : : ; '.k/b uk ; : : :/ 2 `2 .Z/: In particular, if 1I denotes the indicator function of the interval I , we have 1Œ0;C1Œ .D/ D P
and
1Œ0;n .D/ D 1Œ0;1 .n1 D/ D Pn :
Moreover, note that if supp.'/ Œa; b, we have the trivial properties 1Œa;b .D/ '.D/ D '.D/
and
1Œa;b .D/ e i kx D e i kx 1Œak;bk .D/:
In the rest of the paper, a function is a oca!b .1/ if, for each c fixed, the function goes to 0 as a tends to b. In the same way, a function is a O c .1/ if, for each c fixed, the function is a O.1/.
5.2 A Commutator Estimate In this subsection, we recall a standard result of the functional analysis. For 2 R, we denote by S .R/ the class of functions ' in C 1 .R/ such that j@ks '.s/j Ck hsik ; for k 0. Here hxi D .1 C jxj2 /1=2 . Lemma 1 (Lemma C.3.2 of [7]). Let A; B be self-adjoint operators on a Hilbert space with B and ŒA; B bounded. If ' 2 S .R/ with < 1, then kŒ'.A/; Bk C' kŒA; Bk: Here, ŒA; B D AB BA denotes the commutator. The constant C' only depends on '. Applying this lemma, we immediately obtain Lemma 2. Let f 2 C 0 .T/ and ' 2 S .R/ with 0. Then Œ'."D/; f D o"!0 .1/: Proof. By Weierstrass’s theorem, there exist fk 2 C 1 .T/ satisfying fk ! f in L1 .T/. Then, viewed as operators, we have fk ! f . Remark that Œ"D; fk D "ifk0 . From Lemma 1, we obtain kŒ'."D/; fk k "C' kfk0 k1 :
Large Deviations for Gaussian Stationary Processes and Semi-Classical Analysis
419
Then, using the assumption that ' is bounded, Œ'."D/; f D Œ'."D/; fk C ok!1 .1/ D O k ."/ C ok!1 .1/ D o"!0 .1/; since Œ'."D/; f does not depend on k.
t u
5.3 Essential Spectrum of the Product of Toeplitz Operators Here, we recall, in our setting, a consequence of a theorem of Coburn [5] concerning the essential spectrum of the product of Toeplitz operators. This result has been extended by Douglas to a more general framework (see [10, Theorem 4.5.10]). We shall give in Sect. 7 an alternative proof of the following theorem, more related to our approach. Theorem 4 (Coburn). Let f and g be two bounded continuous real functions with g 0. The bounded self-adjoint operator T .g/1=2 T .f /T .g/1=2 satisfies on Im P ess T .g/1=2 T .f /T .g/1=2 D inf.fg/; sup.fg/ : Here, ess .A/ denotes the essential spectrum of A. In Theorem 4, the operator T .g/1=2 T .f /T .g/1=2 is viewed as an operator on Im P . On L2 .T/, this operator is a block diagonal operator with respect to the orthogonal sum L2 D Im P ˚? Im.1 P / and is equal to 0 on Im.1 P /. In particular, we have Remark 2. If the operator T .g/1=2 T .f /T .g/1=2 is viewed on L2 .T/, we have ess T .g/1=2 T .f /T .g/1=2 D inf.fg/; sup.fg/ [ f0g:
6 Proof of Theorem 3 The goal of this section is to prove Theorem 3. First of all, one can observe that part (ii) clearly implies (iii). In the next subsection, we first show that (i) implies (ii).
6.1 The Implication (i) Gives (ii) Lemma 3. Let f and g be two bounded piecewise continuous real functions with g 0. Then,
420
B. Bercu et al.
Tn .g/1=2 Tn .f /Tn .g/1=2 ! T .g/1=2 T .f /T .g/1=2 strongly on L2 .T/. If belongs to the spectrum of T .f /T .g/ on Im P , then there exists an eigenvalue n of Tn .f /Tn .g/ on Im Pn such that n ! . Proof. Since Pn ! P , it follows from Lemma III.3.8 of [9] that for all f 2 L1 .T/, Tn .f / ! T .f /. In particular, from Problem VI.14 of [11] (see also Theorem VI.9 of [11]), Tn .g/1=2 ! T .g/1=2 . Consequently, we deduce from Lemma III.3.8 of [9] that Tn .g/1=2 Tn .f /Tn .g/1=2 ! T .g/1=2 T .f /T .g/1=2 ;
(10)
on L2 .T/. In particular, we obtain on Im P Tn .g/1=2 Tn .f /Tn .g/1=2 C M.P Pn / ! T .g/1=2 T .f /T .g/1=2 ; for all M 2 R. We choose D kf k1 kgk1 and M D C 1. Therefore, it follows from Corollary VIII.1.6 together with Theorem VIII.1.14 of [9] that, for each belonging to the spectrum .T .f /T .g// D .T .g/1=2 T .f /T .g/1=2 / on Im P , there exists an eigenvalue n of the matrix Tn .g/1=2 Tn .f /Tn .g/1=2 C M.P Pn /; on Im P such that n ! . As kT .g/1=2 T .f /T .g/1=2 k , we necessarily have 2 Œ; and then M jj C 1. In particular, for n large enough, M > jn j C 1=2. Therefore, n is an eigenvalue of Tn .g/1=2 Tn .f /Tn .g/1=2 on Im Pn because Tn .g/1=2 Tn .f /Tn .g/1=2 C M.P Pn / D Tn .g/1=2 Tn .f /Tn .g/1=2 ˚? M.P Pn /; is a block diagonal operator with respect to the orthogonal sum Im P D Im Pn ˚? Im.P Pn /: t u
6.2 The Implication (iii) Gives (i) Let N be a sequence of eigenvalues of TN .f /TN .g/ such that N ! 2 R. Here N is a subsequence of N and we have to show that is in the spectrum of T .f /T .g/. From Theorem 4, we know that Œinf.fg/; sup.fg/ is always inside the spectrum of T .f /T .g/. Thus, we can assume that … inf.fg/; sup.fg/ :
(11)
By Weierstrass’s theorem, there exists a sequence of functions .fk / 2 C 1 .T/ such that fk ! f in L1 .T/ and supp fbk Œk; k. We also consider .gk / a sequence corresponding to g with the same properties mutatis mutandis. In
Large Deviations for Gaussian Stationary Processes and Semi-Classical Analysis
421
particular, for all n 2 N, Tn .f / D Tn .fk / C ok!1 .1/
and
T .f / D T .fk / C ok!1 .1/:
(12)
Recall that, by definition, a ok!1 .1/ is uniform with respect to n. Finally, let uN 2 Im PN be an eigenvector of TN .f /TN .g/ associated with N and satisfying kuN k D 1. From (12), TN .f /TN .g/uN D N uN D uN C oN !1 .1/ TN .fk /TN .gk /uN D uN C oN !1 .1/ C ok!1 .1/:
(13) (14)
In the following, we denote Dn D n1 D.
6.2.1 Localization of the Eigenvectors Lemma 4. Let ' 2 C01 .0; 1Œ/. Then, in L2 .T/ norm, '.DN /uN D oN !1 .1/: Proof. From Lemma 2, we have '.DN /TN .f / D '.DN /1Œ0;1 .DN /f 1Œ0;1 .DN / D '.DN /f 1Œ0;1 .DN / D f '.DN /1Œ0;1 .DN / C oN !1 .1/ D f '.DN / C oN !1 .1/:
(15)
Applying two times this estimate, we obtain '.DN /TN .f /TN .g/uN D f '.DN /TN .g/uN C oN !1 .1/ D fg'.DN /uN C oN !1 .1/: Then, (13) gives .fg /'.DN /uN D oN !1 .1/: Since … Œinf.fg/; sup.fg/, the function .fg /1 belongs to L1 .T/ and the lemma follows from the last equation. t u Now, we take ' 2 C01 .0; 1Œ; Œ0; 1/ such that ' D 1 near Œ"; 1 " for " > 0 small enough (we choose " D 1=8). Let ' 2 C01 .Œ"; 2"; Œ0; 1/ and ' C 2 C01 .Œ1 2"; 1 C "; Œ0; 1/ be two functions such that ' C ' C ' C D 1; in the neighborhood of Œ0; 1. Set
422
B. Bercu et al. ˙ ˙ u˙ N D ' .DN /uN D ' .DN /1Œ0;1 .DN /uN :
(16)
As kuN k D 1, it follows from Lemma 4 that C ku N C uN k D 1 C oN !1 .1/:
(17)
In particular, we can assume, up to the extraction of a subsequence, that 8N
ku N k 1=3
or
8N
kuC N k 1=3:
In the next section, we will suppose that ku N k 1=3:
(18)
The case kuC N k 1=3 follows essentially the same lines and is treated in Sect. 6.2.3. C But before, we show that u N and uN are both quasimodes of TN .f /TN .g/ (this means that they are eigenvectors modulo a small term). Lemma 5. We have ˙ k TN .fk /TN .gk /u˙ N D uN C ok!1 .1/ C oN !1 .1/:
Proof. As in (15), using Lemma 2, we get ˙ TN .fk /TN .gk /u˙ N D 1Œ0;1 .DN /fk 1Œ0;1 .DN /gk 1Œ0;1 .DN /' .DN /uN
D 1Œ0;1 .DN /fk 1Œ0;1 .DN /gk ' ˙ .DN /1Œ0;1 .DN /uN D 1Œ0;1 .DN /fk 1Œ0;1 .DN /' ˙ .DN /gk 1Œ0;1 .DN /uN C okN !1 .1/ D 1Œ0;1 .DN /' ˙ .DN /fk 1Œ0;1 .DN /gk 1Œ0;1 .DN /uN C okN !1 .1/ D ' ˙ .DN /TN .fk /TN .gk /uN C okN !1 .1/: The lemma follows from (14), (16) and the last identity.
(19) t u
6.2.2 Concentration Near the Low Frequencies Here, we assume (18) and we prove that u N , viewed as an element of Im P , is a quasimode of T .fk /T .gk /. Lemma 6. For 4k N , we have T .fk /T .gk /u N D TN .fk /TN .gk /uN :
Remark 3. In fact, for 4k N n, we have
Large Deviations for Gaussian Stationary Processes and Semi-Classical Analysis
423
Tn .fk /Tn .gk /u N D TN .fk /TN .gk /uN :
Proof. Recall that, if u; v are two functions of L2 .T/ such that suppb u Œa; b and suppb v Œc; d , then supp uv b Œa C c; b C d . By definition, T .fk /T .gk /u N D Pfk P gk P ' .DN /uN D Pfk P gk PN ' .DN /uN :
(20)
Since supp gbk Œk; k and supp F .PN ' .DN /uN / Œ0; N=4, the Fourier transform of the function gk PN ' .DN /uN is supported inside Œk; N=4 C k Œk; N . In particular, P gk PN ' .DN /uN D PN gk PN ' .DN /uN ;
(21)
and the Fourier transform of this function is supported inside Œ0; N=4Ck. As before, the Fourier transform of fk PN gk PN ' .DN /uN is supported inside Œk; N=4 C 2k Œk; N . Then Pfk PN gk PN ' .DN /uN D PN fk PN gk PN ' .DN /uN : The lemma follows from (20) to (22).
(22) t u
From (12), Lemmas 5 and 6, we get k T .f /T .g/u N D uN C ok!1 .1/ C oN !1 .1/;
(23)
for 4k N . If … .T .f /T .g//, the operator T .f /T .g/ is invertible and then k u N D ok!1 .1/ C oN !1 .1/: From (18), we obtain 1=3 ok!1 .1/CokN !1 .1/. Taking k large enough and then N large enough, it is clear that this is impossible. Thus, 2 .T .f /T .g//; which implies Theorem 3 under Assumption (18).
6.2.3 Concentration Near the High Frequencies e We replace the Assumption (18) by kuC N k 1=3. Let J be the isometry f 7! f in L2 .T/. One can observe that J.uv/ D J.u/J.v/. Using the notation PŒa;b D 1Œa;b .D/, we have PŒa;b J D JPŒb;a and PŒa;b e i cx D e i cx PŒac;bc . Combining these identities with Lemma 5, we get
424
B. Bercu et al. iN x TN .Jfk /TN .Jgk /e iN x .J uC .J uC N / D PŒ0;N .Jfk /PŒ0;N .Jgk /PŒ0;N e N/
D e iN x PŒN;0 .Jfk /PŒN;0 .Jgk /PŒN;0 .J uC N/ D e iN x JPŒ0;N fk PŒ0;N gk PŒ0;N uC N k D e iN x .J uC N / C ok!1 .1/ C oN !1 .1/:
(24)
iN x In particular, e u .J uC u N / satisfies ke N De N k 1=3, k u u TN .Jfk /TN .Jgk /e N D e N C ok!1 .1/ C oN !1 .1/;
and the support of the Fourier transform of e u N is inside Œ0; N=4. Hence, we can apply the method developed in the case ku k N 1=3. The unique difference is that e f; g are replaced by f ;e g . Then, we obtain e/T .e 2 T .f g/ : Theorem 3 follows from the following lemma and .T .f /T .g// D .T .g/T .f // (the spectrum of T .f /T .g/ is real and .T .f /T .g/ z/ D T .g/T .f / z). Lemma 7. Let f; g 2 L1 .T/. Then e/T .e T .f g / D T .g/T .f / : Proof. For A a bounded linear operator on L2 , we define At by .At u; v/ D .u; Av/; t D PŒb;a , .AB/t D B t At for all u; v 2 L2 . Simple calculi give f t D f , PŒa;b and then t T .f /t D PŒ0;C1Œ f PŒ0;C1Œ D P1;0 f P1;0 :
By the same way, since J D J D J 1 , e/: ePŒ0;C1Œ D T .f JP1;0 f P1;0 J D PŒ0;C1Œ f Combining these identities concerning t and J , we get e /T .e eP1;0 J J.T .f g//t J 1 D J P1;0e g P1;0 P1;0 f D T .g/T .f /:
(25)
Since JAt J z D J.A z/t J , A and JAt J have the same spectrum and the lemma follows. t u
Large Deviations for Gaussian Stationary Processes and Semi-Classical Analysis
425
7 Proof of Theorem 4 We give here an alternative proof of Coburn’s theorem. Let 2 C 1 .R/ satisfying D 1 near Œ2; C1Œ and D 0 near 1; 1. For " > 0, we have on Im P e" T .g/1=2 T .f /T .g/1=2 D T .g/1=2 ."D/T .f / ."D/T .g/1=2 C R D T .g/1=2 ."D/f
e" ; ."D/T .g/1=2 C R
(26)
where e" D T .g/1=2 .1 ."D//T .f / ."D/T .g/1=2 CT .g/1=2 T .f /.1 ."D//T .g/1=2 ; R is a self-adjoint operator of finite rank. Recall that if A 0 is a bounded operator with kAk 1, then C1 X cj .1 A/j ; A1=2 D where k1 Ak 1 and implies
P
j D0 j 0 jcj j
2 < C1. On the other hand, Lemma 2
T .g/ ."D/ D P gP ."D/ D P g ."D/ D P ."D/g C o"!0 .1/ D
."D/g C o"!0 .1/;
(27)
Then, for a fixed ı > 0 such that kT .g/k kgk1 < ı 1 , we have T .g/1=2 ."D/ D ı 1=2 T .ıg/1=2 ."D/ D ı 1=2
C1 X
cj .1 T .ıg//j ."D/
j D0
D ı 1=2
J X
cj .1 T .ıg//j ."D/ C oJ !1 .1/
j D0
D ı 1=2 ."D/
J X
cj .1 ıg/j C oJ !1 .1/ C oJ"!0 .1/
j D0
D ı 1=2 ."D/.ıg/1=2 C oJ !1 .1/ C oJ"!0 .1/ D
."D/g 1=2 C o"!0 .1/;
(28)
since these quantities do not depend on J . Using this identity and its adjoint, (26) becomes
426
B. Bercu et al.
T .g/1=2 T .f /T .g/1=2 D
e" C o"!0 .1/ ."D/fg ."D/ C R
D T .fg/ C R" C e" ;
(29)
where e" D o"!0 .1/ and e" C . ."D/ 1/T .fg/ ."D/ C T .fg/. ."D/ 1/; R" D R is a self-adjoint operator of finite rank. In particular, e" is a self-adjoint operator. Since, on Im P inf.fg/ T .fg/ sup.fg/; we get .T .fg/ C e" / Œinf.fg/ o"!0 .1/; sup.fg/ C o"!0 .1/. As R" is of finite rank, we obtain, from Weyl’s theorem [11, Theorem S.13], ess T .g/1=2 T .f /T .g/1=2 D ess .T .fg/ C e" / Œinf.fg/ o"!0 .1/; sup.fg/ C o"!0 .1/: As the essential spectrum of T .g/1=2 T .f /T .g/1=2 does not depend on ", we get ess T .g/1=2 T .f /T .g/1=2 Œinf.fg/; sup.fg/;
(30)
which is the first inclusion of Coburn’s theorem. Now, let ' 2 C 1 .Œ1; 1; Œ0; 1/ with k'kL2 D 1. For x0 2 T and ˛; ˇ 2 N, we set u D ˛ 1=2 ' ˛.x x0 / e iˇx and v D P u 2 Im P; which satisfies kuk D 1. We have .1 P /u D ˛ 1=2 11;0 .D/e iˇx ' ˛.x x0 / D ˛ 1=2 e iˇx 11;ˇ .D/' ˛.x x0 /
D ˛ 1=2 e iˇx 11;ˇ .D/.D C i /M .D C i /M ' ˛.x x0 / (31) D O ˇ M ˛ M ;
in L2 norm for any M 2 N. Moreover, for a continuous function `, we have `u D `.x0 /˛ 1=2 ' ˛.x x0 / e iˇx C o˛!1 .1/;
(32)
in L2 norm. Using that kT .`/1=2 k k`k1 , for all function ` 2 L1 with ` 0, we get 1=2
Large Deviations for Gaussian Stationary Processes and Semi-Classical Analysis
427
T .f /T .g/v D Pf P gP u D Pf P gu C O ˛ˇ 1 D g.x0 /Pf P u C O ˛ˇ 1 C o˛!1 .1/ D g.x0 /Pf u C O ˛ˇ 1 C o˛!1 .1/ D .fg/.x0 /P u C O ˛ˇ 1 C o˛!1 .1/ D .fg/.x0 /v C O ˛ˇ 1 C o˛!1 .1/:
(33)
Taking ˇ D ˛ 2 ! C1, (31) implies kvk D 1 C o˛!1 .1/. On the other hand, (33) leads to T .f /T .g/v D .fg/.x0 /v C o˛!1 .1/: Then, .fg/.x0 / 2 .T .f /T .g// D .T .g/1=2 T .f /T .g/1=2 /. Therefore,
inf.fg/; sup.fg/ T .g/1=2 T .f /T .g/1=2 :
(34)
Recall that the essential spectrum of a self-adjoint bounded operator on an infinite Hilbert space is never empty. Therefore, if inf.fg/ D sup.fg/, (30) implies the theorem. Assume now that inf.fg/ < sup.fg/. Then Œinf.fg/; sup.fg/ is an interval with non empty interior. From the definition of the essential spectrum, this interval is necessarily inside the essential spectrum of T .g/1=2 T .f /T .g/1=2 . This achieves the proof of the second inclusion of Coburn’s theorem. Acknowledgements The authors would like to thanks A. Böttcher for providing the reference of Roch and Silbermann. They also thank the anonymous referee for his careful reading of the paper.
References 1. B. Bercu, F. Gamboa, A. Rouault, Large deviations for quadratic forms of stationary Gaussian processes. Stoch. Process. Appl. 71(1), 75–90 (1997) 2. A. Böttcher, B. Silbermann, Introduction to Large Truncated Toeplitz Matrices (Springer, New York, 1999) 3. A. Böttcher, B. Silbermann, Analysis of Toeplitz Operators, 2nd edn. Springer Monographs in Mathematics (Springer, Berlin, 2006); Prepared jointly with Alexei Karlovich 4. W. Bryc, A. Dembo, Large deviations for quadratic functionals of Gaussian processes. J. Theor. Probab. 10(2), 307–332 (1997); Dedicated to Murray Rosenblatt 5. L.A. Coburn, The C -algebra generated by an isometry. Bull. Am. Math. Soc. 73, 722–726 (1967) 6. A. Dembo, O. Zeitouni, Large Deviations Techniques and Applications, 2nd edn. Applications of Mathematics (New York), vol. 38 (Springer, New York, 1998) 7. J. Derezi´nski, C. Gérard, Scattering Theory of Classical and Quantum N -Particle Systems. Texts and Monographs in Physics (Springer, Berlin, 1997) 8. U. Grenander, G. Szegö, Toeplitz Forms and Their Applications. California Monographs in Mathematical Sciences (University of California Press, Berkeley, 1958)
428
B. Bercu et al.
9. T. Kato, Perturbation Theory for Linear Operators. Classics in Mathematics (Springer, Berlin, 1995); Reprint of the 1980 edition 10. N. Nikolski, Operators, Functions, and Systems: An Easy Reading, vol. 1, Hardy, Hankel, and Toeplitz (translated from the French by A. Hartmann), Mathematical Surveys and Monographs, vol. 92 (American Mathematical Society, RI, 2002) 11. M. Reed, B. Simon, Methods of Modern Mathematical Physics. I, 2nd edn. Functional Analysis (Academic, New York, 1980) 12. S. Roch, B. Silbermann, Limiting sets of eigenvalues and singular values of Toeplitz matrices. Asymptotic Anal. 8, 293–309 (1994) 13. S. Serra-Capizzano, Distribution results on the algebra generated by Toeplitz sequences: A finite-dimensional approach. Linear Algebra Appl. 328(1–3), 121–130 (2001) 14. H. Widom, Asymptotic behavior of block Toeplitz matrices and determinants. II. Adv. Math. 21(1), 1–29 (1976)
Girsanov Theory Under a Finite Entropy Condition Christian Léonard
Abstract This paper is about Girsanov’s theory. It (almost) doesn’t contain new results but it is based on a simplified new approach which takes advantage of the (weak) extra requirement that some relative entropy is finite. Under this assumption, we present and prove all the standard results pertaining to the absolute continuity of two continuous-time processes on Rd with or without jumps. We have tried to give as much as possible a self-contained presentation. The main advantage of the finite entropy strategy is that it allows us to replace martingale representation results by the simpler Riesz representations of the dual of a Hilbert space (in the continuous case) or of an Orlicz function space (in the jump case). Keywords Stochastic processes • Relative entropy • Girsanov’s theory • Diffusion processes • Processes with jumps AMS Classification: 60G07, 60J60, 60J75, 60G44
1 Introduction This paper is about Girsanov’s theory. It (almost) doesn’t contain new results but it is based on a simplified new approach which takes advantage of the (weak) extra requirement that some relative entropy is finite. Under this assumption, we present and prove all the standard results pertaining to the absolute continuity of two continuous-time processes on Rd with or without jumps. This article intends to look like lecture notes and we have tried to give as much as possible a self-contained presentation of Girsanov’s theory. The author hopes that it C. Léonard () Modal-X. Université Paris Ouest. Bât.G, 200 av. de la République, 92001 Nanterre, France e-mail: [email protected] C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9__20, © Springer-Verlag Berlin Heidelberg 2012
429
430
C. Léonard
could be useful for students and also to readers already acquainted with stochastic calculus. The main advantage of the finite entropy strategy is that it allows us to replace martingale representation results by the simpler Riesz representations of the dual of a Hilbert space (in the continuous case) or of an Orlicz function space (in the jump case). The gain is especially interesting in the jump case where martingale representation results are not easy, see [1]. Another feature of this simplified approach is that very few about exponential martingales is needed. Girsanov’s theory studies the relation between a reference process R and another process P which is assumed to be absolutely continuous with respect to R: In particular, it is known that if R is the law of an Rd -valued semimartingale, then P is also the law of a semimartingale. In its wide meaning, this theory also provides us with a formula for the Radon-Nikodým density dP dR : In this article, we assume that the probability measure P has its relative entropy with respect to R: H.P jR/ WD
EP log C1
dP dR
2 Œ0; 1 if P R otherwise,
which is finite, i.e. H.P jR/ D ER
dP dP < 1: log dR dR
(1)
In comparison, requiring P R only amounts to assume that ER
dP dR
<1
(2)
since P has a finite mass. We are going to take advantage of the only difference between (1) and (2) which is the stronger integrability property carried by the extra term log dP : dR A key argument of this approach is the variational representation of the relative entropy which is stated at Lemma 1. A clear exposition of the general Girsanov theorems, with no explicit expression of dP in terms of the characteristics of the processes, is given in Protter’s dR textbook [4]. The most complete results about Girsanov’s theory for Rd -valued processes, including explicit formulas for dP dR ; are available in Jacod’s textbook [2]. An alternate presentation of this realm of results is also given in the later book by Jacod and Shiryaev [3]. A good standard reference in the continuous case is Revuz and Yor’s textbook [6] about continuous martingales. Next Sect. 2 is devoted to the statement of the main results. At Sect. 3, we state and prove the above mentioned variational representation of the relative entropy. At Sects. 4 and 5, we present the proofs of Theorems 1 and 2 which correspond to the continuous case. At Sect. 6, we give the proofs of Theorems 3 and 4 which correspond to the jump case.
Girsanov Theory Under a Finite Entropy Condition
431
2 Statement of the Results We separate the cases when the sample paths are continuous and when they exhibit jumps.
2.1 Continuous Processes in Rd The paths to be considered are built on the time interval Œ0; 1. An Rd -valued continuous stochastic process is a random variable taking its values in the set ˝ D C.Œ0; 1; Rd / of all continuous paths from Œ0; 1 to Rd : The canonical process .Xt /t 2Œ0;1 is defined by Xt .!/ D !t ; t 2 Œ0; 1; ! D .!s /s2Œ0;1 2 ˝: In other words, X D .Xt /t 2Œ0;1 is the identity on ˝ and Xt W ˝ ! Rd is the tth projection. The set ˝ is endowed with the -field .Xt I t 2 Œ0; 1/ which is generated by the canonical projections. We also consider the canonical filtration
.XŒ0;t /I t 2 Œ0; 1 where for each t; XŒ0;t WD .Xs /s2Œ0;t : Let us give ourselves a reference probability measure R on ˝ such that X admits the R-semimartingale decomposition X D X0 C B R C M R ;
R-a.s.
(3)
This means that B R is an adapted process with bounded variation sample paths R-a.s. and M R is a local R-martingale, i.e. there exists an increasing sequence of stopping times .k /k1 which converges to infinity R-a.s. and such that for each k 1; the stopped process t 7! MtR^k is a uniformly integrable R-martingale. As a typical example, one may think of the solution to the SDE (if it exists) Z
Z
Xt D X0 C
bs .XŒ0;s / ds C
s .XŒ0;s / d Ws ;
Œ0;t
0t 1
(4)
Œ0;t
where W is a Wiener process on Rd and b W Œ0; 1 ˝ ! Rd and W Œ0; 1 ˝ ! Md d are locally bounded. In this situation, a natural localizing sequence .k /k1 isR the sequence of exit times from the Euclidean balls of radius t k; BtR D 0 bs .XŒ0;s / ds has absolutely continuous sample paths R-a.s. and Rt R Mt D 0 s .XŒ0;s / d Ws has the quadratic variation Z
t
ŒM R ; M R t D
as ds 0
R-a.s.
(5)
432
C. Léonard
where at WD t t .XŒ0;t / takes its values in the set SC of all positive semi-definite d d matrices. More generally, we assume that the quadratic variation of M R is a process which is R-a.s. equal to a random element of the set MSnaC .Œ0; 1/ of all bounded measures on Œ0; 1 with no atoms and taking their values in SC : ŒM R ; M R .dt/ D A.dt/ 2 MSnaC .Œ0; 1/;
R-a.s.
(6)
and also that t 2 Œ0; 1 7! ŒM R ; M R t WD A.Œ0; t/ D A.t; XŒ0;t I Œ0; t/ 2 SC ;
R-a.s.
is an adapted process. The quadratic variation given at (6) might have an atomless singular part in addition to its absolutely continuous component at dt: This notation is concise: A.dt/ is random and for any Rd -valued processes ˛; Rˇ; ˛t A.dt/ˇt is the infinitesimal element of a measure on Œ0; 1: In particular, t 7! Œ0;t A.ds/ˇs 2 Rd ; R R t 7! Œ0;t ˛s A.ds/ˇs 2 R and the process t 7! Œ0;t ˇs A.ds/ˇs 2 R is increasing. Summing up, R is a solution to the martingale problem MP.B R ; A/. This means that the canonical process X is the sum (3) of a bounded variation adapted process B R and a local R-martingale M R whose quadratic variation is specified by A and (6). We write R 2 MP.B R ; A/ for short. Theorem 1 (Girsanov’s theorem). Let R and P be as above, satisfying in particular the finite entropy condition (1). Then, P is the law of a semimartingale. More precisely, there exists an Rd -valued adapted process ˇ which is defined P -a.s. and such that Z EP ˇt A.dt/ˇt < 1 (7) Œ0;1
and defining bt WD B
Z A.ds/ˇs ;
0 t 1;
P -a.s.
(8)
Œ0;t
one obtains
b C MP; X D X0 C B R C B
P -a.s.
(9)
where M P is a local P -martingale such that ŒM P ; M P D ŒM R ; M R ; P -a.s. b A/: In other words, P 2 MP.B R C B; Remark 1. (a) The process ˇ only needs to be defined P -a.s. (and not R-a.s.) for the statement of Theorem 1 to be meaningful. In fact, its proof only gives the “construction” of a process ˇ, P -almost everywhere. b is well defined. Indeed, by Cauchy-Schwarz inequality, for any (b) The process B d R -valued process ;
Girsanov Theory Under a Finite Entropy Condition
1=2 Z
Z
Z jt A.dt/ˇt j Œ0;1
433
1=2
ˇt A.dt/ˇt Œ0;1
t A.dt/t Œ0;1
2 Œ0; 1; P -a.s. R Looking at A.!/ with ! fixed as a matrix of measures, we see that supf Œ0;1 t A.dt/t I W jt j D 1; 8tg is bounded above by the sum of the total variations of the entriesR of A. Consequently, this supremum is finite R P -a.s. On the other hand, as EP Œ0;1 ˇt A.dt/ˇt < 1; we see that a fortiori Œ0;1 ˇt A.dt/ˇt < 1; R b is well P -a.s. It follows that Œ0;1 jA.dt/ˇt j < 1; P -a.s. which means that B defined. (c) When the quadratic variation is given by (5), one retrieves the standard representation Z b Bt D as ˇs ds: Œ0;t
It is then known that under the minimal assumption (2), Theorem 1 still holds true with Z ˇt at ˇt dt < 1; P -a.s. instead of EP [3, Chap. III].
R Œ0;1
Œ0;1
ˇt at ˇt dt < 1 under the assumption (1), see for instance
We introduce a slight modification of the standard definitions of stochastic integral and local martingale. Definition 1 (P -locality). Let P R 2 P.˝/: i) A process M is said to be a P -local R-martingale if there exists an increasing sequence of Œ0; 1 [ f1g-valued stopping times .k /k1 such that limk!1 k D 1; P -a.s. (and not R-a.s.) such that the stopped processes M k are R-martingales, for all k 1: ii) A process Y is said to be a P -local R-stochastic integral if there exists an increasing sequence of Œ0; 1 [ f1g-valued stopping times .k /k1 such that limk!1 k D 1; P -a.s. (and not R-a.s.) such that the stopped processes Y k are L2 .R/-stochastic integrals, for all k 1: For any probability Q on ˝; let us denote Q0 D .X0 /# Q the law of the initial position X0 under Q: Definition (Condition (U)). One says that R 2 MP.B R ; A/ satisfies the uniqueness condition (U) if for any probability measure R0 on ˝ such that the initial laws R00 D R0 are equal, R0 R and R0 2 MP.B R ; A/, we have R D R0 : It is known [1] that if the SDE (4) admits a unique solution, for instance if the coefficients b and are locally Lipschitz, then its law R satisfies (U). Theorem 2 (The density dP =dR). Let R and P be as above, satisfying in particular the finite entropy condition (1). Keeping the notation of Theorem 1, we have
434
C. Léonard
1 H.P0 jR0 / C EP 2
Z ˇt A.dt/ˇt H.P jR/: Œ0;1
If in addition it is assumed that R satisfies the uniqueness condition (U), then 1 H.P0 jR0 / C EP 2
Z ˇt A.dt/ˇt D H.P jR/; Œ0;1
b AI P0 / in the class P is the unique solution of the martingale problem MP.B R C B; of all probability measures which have a finite relative entropy with respect to R and Z Z dP dP0 1 R D 1f dP >0g .X0 / exp ˇt dMt ˇt A.dt/ˇt dR dR dR0 2 Œ0;1 Œ0;1 Z Z dP0 1 P D 1f dP >0g .X0 / exp ˇt dMt C ˇt A.dt/ˇt dR dR0 2 Œ0;1 Œ0;1 where M R and M P are defined at (3) and (9) respectively and R R (i) Œ0;1 ˇt dMtR D Œ0;1 ˇt .dXt dBtR / is a P -local R-stochastic integral and Ra P -local R-martingale; R (ii) Œ0;1 ˇt dMtP D Œ0;1 ˇt .dXt dBtR A.dt/ˇt / is a well-defined P -stochastic R
2 integral such that EP Œ0;1 ˇt dMtP < 1: Because of the prefactor 1f dP >0g , the formula for dP is meaningful when dR dR R R considering Œ0;1 ˇt dMtR as a P -local R-stochastic integral and Œ0;1 ˇt dMtP R as a P -stochastic integral. Note that (7) implies that Œ0;1 ˇt A.dt/ˇt < 1; P -a.s. Let us denote dPŒ0;t Zt WD ; t 2 Œ0; 1: dRŒ0;t Since RŒ0;t and PŒ0;t are the push-forward of R and P by XŒ0;t ; we have Zt D ER . dP dR j XŒ0;t /; which also shows that .Zt /t 2Œ0;1 is a uniformly integrable R-martingale. Corollary 1. Suppose that the condition (U) is satisfied. Then, the R-martingale .Zt /t 2Œ0;1 satisfies: dZt D Zt ˇt dMtR ; R-a.s.
R R Proof. By Theorem 2, Zt D 1fZt >0g exp Œ0;t ˇs dMsR 12 Œ0;t ˇs A.ds/ˇs : It follows with Itô’s formula applied to the exponential function that, when Zt > 0; dZt D Zt ˇt dMtR ; P -a.s. On the other hand, for all 0 s t 1; we have 0 D PŒ0;s .Zs D 0/ D PŒ0;t .Zs D 0/ D ERŒ0;t .Zt 1fZs D 0g / which implies that Zt D 0 if Zs D 0 for some 0 s t; R-a.s. It follows that dZt D 0 when Zt D 0; R-a.s. Hence, choosing any version of ˇ on the P -negligible set where it is
Girsanov Theory Under a Finite Entropy Condition
435
unspecified, dZt D Zt ˇt dMtR ; R-a.s., which is meaningful R-a.s. because of the prefactor Zt : t u
2.2 Processes with Jumps in Rd The law of a process with jumps is a probability measure P on the canonical space ˝ D D.Œ0; 1; Rd / of all left limited and right continuous (càdlàg) paths, endowed with its canonical filtration. We denote X D .Xt /t 2Œ0;1 the canonical process, Xt D Xt Xt the jump at time t and Rd WD Rd n f0g the set of all effective jumps. A Lévy kernel is a random -finite positive measure L! .dtdq/ D .dt/L! .t; dq/;
!2˝
on Œ0; 1 Rd where is assumed to be a -finite positive atomless measure on Œ0; 1: As a definition, any Lévy kernel is assumed to be predictable, i.e. L! .t; dq/ D L.XŒ0;t / .!/I t; dq/ for all t 2 Œ0; 1: Let B be a bounded variation continuous adapted process. Definition 2 (Lévy kernel and martingale problem). We say that a probability measure P on ˝ solves the martingale problem MP.B; L/ if the integrability assumption Z EP
.jqj2 ^ 1/ L.dtdq/ < 1
(10)
Œ0;1Rd
holds and for any function f in Cb2 .Rd /; the process f .XQt /f .XQ0 /
Z .0;t Rd
Œf .XQs Cq/f .XQs /rf .XQs / q 1fjqj1g L.dsdq/ Z .0;t Rd
Œf .XQs C q/ f .XQs / 1fjqj>1g L.dsdq/
is a local P -martingale, where XQ WD X B. We write this P 2 MP.B; L/
436
C. Léonard
for short. In this case, we also say that P admits the Lévy kernel L and we denote this property P 2 LK.L/ for short. If P 2 MP.B; L/; the canonical process is decomposed as L ; Xt D X0 C Bt C .1.0;t 1fjqj>1g q/ ˇ X C .1.0;t 1fjqj1g q/ ˇ e where X WD
X
P -a.s.
(11)
ı.t;Xt /
t 2Œ0;1IXt 6D0
R P is the canonical jump measure, '.q/ ˇ X D Œ0;1Rd ' dX D t 2Œ0;1IXt 6D0 '.Xt / and '.q/ ˇ e L is the P -stochastic integral with respect to the compensated sum of jumps X e L ! .dtdq/ WD ! .dtdq/ L! .dtdq/: For short, we rewrite formula (11): X D X0 C B C .1fjqj>1g q/ ˇ X C .1fjqj1g q/ ˇ e L ;
P -a.s.
(12)
Definition 3 (Class Hp;r .P; L/). Let P be a probability measure on ˝ and L a Lévy kernel such that P 2 LK.L/: R We say that a predictable integrand h! .t; q/ is in the class Hp;r .P; L/ if EP Œ0;1Rd 1fjqj1g jht .q/jp L.dtdq/ < 1 and R EP Œ0;1Rd 1fjqj>1g jht .q/jr L.dtdq/ < 1: We denote Hp;p .P; L/ D Hp .P; L/. We take our reference law R such that R 2 MP.B R ; L/ for some adapted continuous bounded variation process B R : The integrability assumption (10) means that the integrand jqj is in H2;0 .R; L/: This will be always assumed in the future. We introduce the function .a/ D log Ee a.N 1/ D e a a 1;
a 2 R:
where N is a Poisson(1) random variable. Its convex conjugate is 8 < .b C 1/ log.b C 1/ b if b > 1 .b/ D 1 if b D 1 ; : 1 otherwise
b 2 R:
Note that and are respectively equivalent to a2 =2 and b 2 =2 near zero.
Girsanov Theory Under a Finite Entropy Condition
437
Theorem 3 (Girsanov’s theorem. The jump case). Let R and P be as above: R 2 MP.B R ; L/ and H.P jR/ < 1: Then, there exists a predictable nonnegative process ` W ˝ Œ0; 1 Rd ! Œ0; 1/ which is defined P -a.s., satisfying Z EP
.j` 1j/ d L < 1;
(13)
Œ0;1Rd
b` ; `L/ where such that P 2 MP.B R C B b`t D B
Z Œ0;t Rd
1fjqj1g .`s .q/ 1/q L.dsdq/;
t 2 Œ0; 1
is well-defined P -a.s. It will appear that, in several respects, log ` is analogous to ˇ in Theorem 1. Again, ` only needs to be defined P -a.s. and not R-a.s. for the statement of Theorem 3 to be meaningful. And indeed, its proof only provides us with a P -a.s.construction of `: Corollary 2. Suppose that in addition to the assumptions of Theorem 3, there exist some ao ; bo ; co > 0 such that Z ER exp ao
Œ0;1Rd
1fjqj>co g e bo jqj L.dtdq/ < 1:
(14)
It follows immediately that 1fjqj>co g jqj is R ˝ L-integrable so that the stochastic integral q ˇ e L is well-defined R-a.s. and we are allowed to rewrite (12) as X D X0 C B C q ˇ e L ;
R-a.s.;
for some adapted continuous bounded variation process B: Then, there exists a predictable nonnegative process ` W ˝ Œ0; 1 Rd ! Œ0; 1/ satisfying (13) such that `
`L ; X D X0 C B C B C q ˇ e Z
where ` Bt
P -a.s.;
D Œ0;t Rd
.`s .q/ 1/q L.dsdq/;
t 2 Œ0; 1
is well-defined P -a.s. and the P -stochastic integral q ˇe `L with respect to the Lévy kernel `L is a local P -martingale. Remark 2. (a) The energy estimate (13) is equivalent to: 1f0`2g .` 1/2 and 1f`2g ` log ` are integrable with respect to P ˝ L:
438
C. Léonard `
(b) Together with (13), (14) implies that the integral for B is well-defined since Z EP
Œ0;1Rd
.`t .q/ 1/jqj L.dtdq/ < 1:
(15)
In the present context of processes with jumps, the uniqueness condition (U) becomes: Definition (Condition (U)). One says that R 2 MP.B R ; L/ satisfies the uniqueness condition (U) if for any probability measure R0 on ˝ such that the initial laws R00 D R0 are equal, R0 R and R0 2 MP.B R ; L/, we have R D R0 : Theorem 4 (The density dP =dR). Suppose that R and P verify R 2 MP.B; L/ and H.P jR/ < 1: With ` given at Theorem 3, we have Z H.P0 jR0 / C EP
.` log ` ` C 1/ d L H.P jR/ Œ0;1Rd
with the convention 0 log 0 0 C 1 D 1: If in addition it is assumed that R satisfies the uniqueness condition (U), then Z H.P0 jR0 / C EP
.` log ` ` C 1/ d L D H.P jR/; Œ0;1Rd
b` ; `LI P0 / in the P is the unique solution of the martingale problem MP.B R C B class of all probability measures which have a finite relative entropy with respect to R and Z dP dP0 L D 1f dP >0g .X0 / exp log ` ˇ e .log `/ d L (16) dR dR dR0 Œ0;1Rd Z dP0 `L .X0 / exp log ` ˇ e C .` log ` ` C 1/ d L : D 1f dP >0g dR dR0 Œ0;1Rd
e e
e
In formula (16), exp indicates a shorthand for the rigorous following expression 8 dP dP0 ˆ ˆ D .X0 /Z C Z with ˆ ˆ ˆ dR dR 0 Z ˆ ˆ ˆ ˆ L C ˆ Z D 1 exp Œ1 log ` ˇ e .` log ` 1/d L dP f`1=2g ˆ f dR >0g < Zf`1=2g `L ˆ D 1f dP >0g exp Œ1f`1=2g log ` ˇ e C .` log ` ` C 1/d L ˆ ˆ dR ˆ f`1=2g ˆ Z ˆ Y ˆ ˆ ˆ ˆ Z D 1f dP >0; D1g exp Œ` 1d L `.t; Xt / ˆ dR : f0`<1=2g 0t 1I0<`.t;Xt /<1=2
(17)
Girsanov Theory Under a Finite Entropy Condition
439
where : L in Z C is a P -local R-martingale and a – The inner term Œ1f`1=2g log ` ˇ e P -local R-stochastic integral; – The inner term Œ1f`1=2g log `ˇe `L in Z C is a well-defined P -stochastic integral ˇ ˇ 2 such that EP Œ1f1=2`2g log ` ˇ e `L < 1 and EP ˇŒ1`2g log ` ˇ e `L ˇ < 1I – The product in Z contains finitely many terms, P -a.s.; (and is equal to 1 when it doesn’t contain any term); – We defined WD sup inf ft 2 Œ0; 1I `.t; Xt / 1=ng 2 Œ0; 1 [ f1g; n1
with the convention that inf ; D 1: Note that although ` is only defined P -a.s.; Z C , Z and are well-defined since they contain the prefactor 1f dP >0g : dR
Remark 3.
R (a) Because of (13), the integral f`1=2g .` log ` 1/d L inside Z C is finite P -a.s. Q (b) Similarly, the product 0t 1I0<`.t;Xt /<1=2 `.t; Xt / doesn’t vanish P -a.s. because it is proved at Lemma 8 that P . D 1/ D 1: (c) Note that this product is well-defined in Œ0; 1 since it contains P -a.s. at most countably many terms in .0; 1=2: But, if it contains infinitely many such terms, it vanishes. Therefore, it contains P -a.s. finitely many terms. (d) Since inf ft 2 Œ0; 1I `.t; Xt / D 0g ; if `.t; Xt / D 0 for some t 2 Œ0; 1; then dP D 0: Therefore, ` > 0; P -a.s. dR is (e) If 1f`1=2g log ` is R ˝ L-integrable, an alternate expression of dP dR Z Y dP0 dP D 1f dP >0; D1g .X0 / exp .` 1/ d L `.t; Xt /: dR dR dR0 Œ0;1Rd 0t 1 (f) If 1f0`<1=2g log ` is not R ˝L-integrable, then log `ˇe L is undefined and (16) with exp instead of exp is meaningless andRmust be replaced by (17). (g) On the other hand, if ` > 0; R-a.s. and ER Œ0;1Rd .log `/ d L < 1; then (16) with exp instead of exp. gives the rigorous expression for dP dR
e
e
For more details about the relationship between (16) and (17), see the discussion below Proposition 1 at the Appendix. Corollary 3. Suppose that the condition (U) is satisfied. Then, the R-martingale dP Zt D dRŒ0;t satisfies: dZt D Zt Œ.` 1/ ˇ de L t ; R-a.s. Œ0;t Its proof is similar to Corollary 1’s one. For more details about the martingale Z of Corollary 3, see Proposition 1 at the Appendix.
440
C. Léonard
3 Variational Representations of the Relative Entropy Theorems 1 and 3’s proofs rely on some variational representation of the relative entropy which is stated and proved below. Lemma 1 (Variational representations of the relative entropy). Let R be a probability measure on some space ˝: For any probability measure P 2 P.˝/ such that H.P jR/ < 1 we have H.P jR/
Z
Z
D sup
u dP log Z
Z
D sup
u dP log
Z e u dRI u W ˝ ! Œ1; 1/;
e u dR < 1
(18)
e u dRI u bounded
where we use the convention e 1 D 0 and the functions u are assumed to be measurable. R R About the first identity, H.P jR/ < 1 and e u dR < 1 imply that u dP 2 Œ1; 1/: Proof. The proof is based on Fenchel’s inequality ab .a log a a C 1/ C .e b 1/;
8a 0; b 2 Œ1; 1/
and its case of equality ab D .a log a a C 1/ C .e b 1/ , e b D a where as a convention 0 log 0 D 0; e 1 D 0 and 0 .1/ D 0: dP RChoosing Ra Du dR .!/, b D u.!/ and integrating with respect to R; we obtain u dP .e 1/ dR H.P jR/ < 1 which proves the last statement of the lemma. With the case of equality, we also see that Z H.P jR/ D sup
Z
Z .e u 1/ dRI u W ˝ ! Œ1; 1/;
u dP
e u dR < 1 : (19)
Now, let us take advantage of the unit mass of R and P W Z
Z
Z .u C ˇ/ dP
.e
.uCˇ/
1/ dR D
Z u dP e
e u dR C ˇ C 1;
ˇ
8ˇ 2 R:
Thanks to the elementary identity log ˛ D infˇ2R f˛e ˇ ˇ 1g; we obtain Z sup ˇ2R
Z .u C ˇ/ dP
.e
.uCˇ/
1/ dR D
Z
Z u dP log
e u dR:
Girsanov Theory Under a Finite Entropy Condition
441
Hence, Z
Z u dP
sup
Z .e u 1/ dRI u W ˝ ! Œ1; 1/;
Z
e u dR < 1
Z
D sup
.u C ˇ/ dP
Z
.e .uCˇ/ 1/ dRI ˇ 2 R; u W ˝ ! Œ1; 1/;
e u dR < 1 Z D sup
Z u dP log
Z e u dRI u W ˝ ! Œ1; 1/;
e u dR < 1 :
With (19), this proves the first identity of the lemma. The second one follows by monotone convergence, considering the approximadP tion .un /n1 of log dP t u dR which is given by: un WD .n/ ^ log dR _ n:
4 Proof of Theorem 1 For the proof of Theorem 1 we need to exhibit a large enough family of exponential supermartingales. Lemma 2 (Exponential supermartingales). Let M be a local martingale, then 1 ZtM D exp Mt ŒM; M t ; 2
0 t 1;
is also a local martingale and a supermartingale. In particular, 0 ER Z1M 1: Proof. Recall Itô’s formula 1 df .Yt / D f 0 .Yt / d Yt C f 00 .Yt / d ŒY; Y t 2 which is valid for any C 2 function f and any continuous semimartingale Y . Applying it to Yt D Mt 12 ŒM; M t and f .y/ D e y ; we obtain 1 1 dZtM D ZtM dMt d ŒM; M t C d ŒM; M t D ZtM dMt 2 2 which proves that Z M is a local martingale. Since Z M 0; Fatou’s lemma applied to the localized sequence ZtM^k as k tends to infinity tells us that Z M is a R-supermartingale, with .k /k1 an increasing sequence of stopping times which tends almost surely to infinity and localizes the local martingale M . In particular, E.Z1M / E.Z0M / D 1: t u
442
C. Léonard
The standard notation for the supermartingale of Lemma 2 is 1 E .M / WD exp M ŒM; M : 2 We are now ready for the proof of Theorem 1. Proof (of Theorem 1). We start with some useful notation. Let Q be a probability measure on ˝I later we shall take Q D R or Q D P: For any measurable function g on Œ0; 1 ˝; let us denote Z .g; g/A .!/ WD
gt .!/ At .!I dt/gt .!/ 2 Œ0; 1 Œ0;1
and introduce the function space ˚ G .Q/ WD g W Œ0; 1 ˝ ! Rd I g measurable; EQ .g; g/A < 1 endowed with the seminorm kgkG .Q/ WD .EQ .g; g/A /1=2 : Identifying the functions with their equivalence classes when factorizing by the kernel of this seminorm, turns G .Q/ into a Hilbert space. These equivalence classes are called G .Q/-classes and with some abuse, we say that two elements of the same class are equal G .Q/-almost everywhere. The relevant space of integrands for the stochastic integral is H
Q
WD fh 2 G .Q/I h adaptedg:
Identity (3) says that M R D X X0 B R is a local R-martingale. For all h 2 H R ; let us denote the stochastic integral Z h MtR WD
t 0
hs dMsR ;
t 2 Œ0; 1:
R
By Lemma 2, 0 < ER Z1hM 1 for all h 2 H R and because of (18), for any probability measure P such that H.P jR/ < 1; we have 1 R R R EP h M1 Œh M ; h M 1 H.P jR/; 8h 2 H R : 2
(20)
Note that, as P R; h M1R and Œh M R ; h M R 1 which are defined R-a.s.; are a fortiori defined P -a.s. With (6) and (20), we see that 1 EP .h M R / H.P jR/ C EP .h; h/A ; 2
8h 2 G .P / \ H R :
(21)
Girsanov Theory Under a Finite Entropy Condition
443
The notation G .P / \ H R is a little bit improper. Indeed, G .P / is a set of equivalence classes with respect to the equality G .P /-a.e., while H R is a set of G .R/-classes. But since P R; keeping in mind that any G .P /-class is the union of some G .R/-classes, one can interpret G .P / \ G .R/ as a set of G .P /-classes. It is also clear that G .P / \ H R D H P \ H R which is a set of G .P /-classes. Considering h in (21), we obtain for all > 0; ˇ ˇ ˇ ˇ ˇEP h M R ˇ H.P jR/ C 1 EP .h; h/A ; ˇ ˇ
2 2 Let
( S WD h W Œ0; 1 ˝ ! R I h D d
k X
8h 2 H P \ H R : ) hi 1Si ;Ti
i D1
denote the set of all simple adapted processes h where k is finite and for all i; hi 2 Rd and Si Ti are stopping times. As S H P \ H R ; taking D khkG .P / in previous inequality, we obtain the keystone of the proof: jEP .h M R /j ŒH.P jR/ C 1=2 khkG .P / ;
8h 2 S:
This estimate still holds when khkH .P / D 0: Indeed, for all real ˛, by (21) we see that ˛EP .h M R / H.P jR/ C ˛ 2 =2 EP .h; h/A D H.P jR/: Letting j˛j tend to infinity, it follows that EP .h M R / D 0: Under the assumption that H.P jR/ is finite, this means that h 7! h M R is continuous on S with respect to the Hilbert topology of H P : As S is dense in H P ; this linear form extends uniquely as a continuous linear form on H P : It also appears that this extension is again a stochastic integral with respect to P: We still denote this extension by h M R : As h 7! h M R is a continuous linear form on H P ; we know by the Riesz representation theorem that there exists a unique ˇ 2 H P such that EP .h M R / D EP .ˇ; h/A ;
8h 2 S:
Z
In other words, EP
Œ0;1
ht dMtP D 0;
Z
where MtP WD MtR
Œ0;t
8h 2 S
bt ; A.ds/ˇs D Xt X0 BtR B
which means that M P is a local P -martingale.
t u
444
C. Léonard
5 Proof of Theorem 2 It relies on a transfer result which is stated below at Lemma 3. But we first need to introduce its framework and some additional notation. Let P be a probability measure on ˝ such that ŒX; X D A; P -a.s. and X D X0 C B C M P ;
P -a.s.;
P where B is a bounded variation R process and M is a local P -martingale. Let be an adapted process such that Œ0;1 t A.dt/t < 1; P -a.s. We define
Z Zt D exp Œ0;t
s dMsP
1 2
s A.ds/s ;
Z
0t 1
Œ0;t
and for all k 1;
Z
WD inf t 2 Œ0; 1I
s A.ds/s k 2 Œ0; 1 [ f1g;
k
Œ0;t
with the convention inf ; D 1: We use the standard notation Yt D Y ^t for the process Y stopped at a stopping time : For all k; P k WD X k # P is the push-forward of P with respect to the stopping procedure X k : Note that P k and P match on the -field which is generated by XŒ0;k : Lemma 3. Let P and as above. Then, for all k 1; Z k is a genuine P -martingale and the measure Qk WD Z1k P k k bk k bk is R a probability measurek on ˝ which ksatisfies Q 2 MP.B ; A / where B t D Œ0;t ^k A.ds/s and M is a local Q -martingale.
Proof. Let us first show that Z k is a P k -martingale.1 The local martingale Z k is of the form Z k D E .N / WD exp.N 12 ŒN; N / with N a local P k -martingale such that ŒN; N k; P k -a.s. For all p 0; since E .N /p D exp.pN p2 ŒN; N / and E .pN / D exp.pN
p2 2 ŒN; N /
e pN e kp
2 =2
E .N /p e pN e kp
; we obtain
2 =2
E .pN /:
As a nonnegative local martingale, E .pN / is a supermartingale. We deduce from this that EP k E .pN / 1 and
1
It is a direct consequence of Novikov’s criterion, but we prefer presenting an elementary proof which will be a guideline for a similar result with jump processes.
Girsanov Theory Under a Finite Entropy Condition
EP k E .N /p e kp
2 =2
445
EP k E .pN / e kp
2 =2
< 1:
Choosing p > 1; it follows that E .N / is uniformly integrable. In particular, this implies that EP k E .N /1 D EP k E .N /0 D 1 and proves that Qk is a probability measure. Suppose now that the supermartingale E .N / is not a martingale. This implies that there exists 0 t < 1 such that on a subset with positive measure, EP k .E .N /1 j XŒ0;t / < E .N /t : Integrating, we get 1 D EP k E .N /1 < EP k E .N /t ; which contradicts EP k E .N /s EP k E .N /0 D 1; 8s: a consequence of the supermartingale property of E .N /. Therefore, E .N / is a genuine P k -martingale. bk ; Ak /. First of all, Let us fix k 1 and show that Qk is a solution to MP.B as it is assumed that ŒX; X D A; P -a.s.; we obtain ŒX; X D Ak ; P k -a.s. With Qk P k ; this implies that ŒX; X D Ak ; Qk -a.s. Now, we check bk C M k X D X0 C B k C B (22) where M k is a Qk -martingale. Let be a stopping time and denote Ft D Xt with 2 Rd : The martingale Z k is the stochastic exponential E .N / of Nt D R P k Œ0;t 1Œ0;k .s/s dMs : Hence, denoting Z D Z ; we have dZt D Zt 1Œ0;k .t/t P P dMt ; dFt D 1Œ0; .t/ .dBt C dMt / and d ŒZ; F t D Zt 1Œ0; ^k .t/ A.dt/t ; P k -a.s. Consequently, .a/
EQk Œ .X X0 / D EP k ŒZ F Z0 F0 Z .b/ D EP k .Ft dZt C Zt dFt C d ŒZ; F t / Z
Œ0;
Z Ft dZt C
D EP k Œ0;
Z C
Œ0;
Zt A.dt/t Œ0;
.c/
Zt .dBt C dMtP /
Z
Z
D EP k
Zt dBt C Œ0;
Z .d / D EQ k
Zt A.dt/t Œ0;
.dBt C A.dt/t / :
Œ0;
In order that all the above terms are meaningful, we choose such that it localizes F; B; M P and A: This is possible, taking for any n 1; n D F B M F B Rmin.n ; n ; n ; n / where n DR infft 2 Œ0; 1I jXt j ng; n MD infft 2 Œ0; 1I Œ0;t jdBs j ng; n D infft 2 Œ0; 1I Œ0;t s A.ds/s ng; and n is a localizing sequence of the local martingale M P : We have
446
C. Léonard
lim n D 1;
P k -a.s.
n!1
(23)
We used the definition of Qk and the martingale property of Z at (a) and (d), (b) is Itô’s formula and (c) relies on the martingale property of Z and .M P / . Finally, taking D & ^ n ; we see that for any stopping time &; any n 1 and any 2 Rd EQk Œ
.X&n
X0n /
Z
D EQ k
.dBt C A.dt/t / : Œ0;&^n
b is a local Qk -martingale. Taking (23) into account, this means that X X0 B B k We conclude remarking that for any process Y; we have Y D Y ; Qk -a.s. This leads us to (22). t u Let us denote P D X # P the law under P of the process X which is stopped at the stopping time . Lemma 4. If R fulfills the condition (U), then for any stopping time ; R also fulfills it. Proof. Let us fix the stopping time : Our assumption on R implies that X D X0 C B C M;
R -a.s.
where M D M R is a local R-martingale and we denote B D B R . Let Q R be given such that Q0 D R0 and X D X0 C B C M Q ;
Q-a.s.
where M Q is a local Q-martingale. We wish to show that Q D R : The disintegration R D RŒ0; ˝ R. j XŒ0; / means that for any bounded measurable function F on ˝; denoting F D F .X / D F .XŒ0; ; X.;1 /; Z ER .F / D
ER ŒF . ; X.;1 / j XŒ0; D RŒ0; .d /: ˝
Similarly, we introduce the probability measure R0 WD QŒ0; ˝ R. j XŒ0; /: To complete the proof, it is enough to show that R0 satisfies X D X0 C B C M 0 ;
R0 -a.s.
(24)
Girsanov Theory Under a Finite Entropy Condition
447
with M 0 a local R0 -martingale. Indeed, the condition (U) tells us that R0 D R; which implies that R0 D R : But R0 D Q; hence Q D R : Let us show (24). Let 2 Rd and a stopping time be given. We denote .n /n1 a localizing sequence of M D M R and B D B R : Then, ER0 Œ .Xn X0n / D ER0 Œ1f g .Xn Xn / C EQ Œ .Xn X0n / Z ER Œ1f g .Xn Xn / j XŒ0; D Q.d / C EQ Œ .Xn X0n / D Z
˝
D ˝
ER Œ1f g .Bn Bn / j XŒ0; D Q.d / C EQ Œ .Bn B0n /
D ER0 Œ .Bn B0n / This means that (24) is satisfied (with the localizing sequence .n /n1 ) and completes the proof of the lemma. t u For all k 1; we consider the stopping time Z k D inf t 2 Œ0; 1I
ˇs A.ds/ˇs k 2 Œ0; 1 [ f1g Œ0;t
where ˇ is the process which is associated with P in Theorem 1 and as a convention inf ; D 1: We are going to use this stopping time R-a.s. Since ˇ is only defined P -a.s.; we assume for the moment that P and R are equivalent measures: P R: Lemma 5. Assume that P R and suppose that R satisfies the condition (U). Then, for all k 1; on the stochastic interval 0; k ^ 1 we have, R-almost everywhere dP dP0 D10;k ^1 10;k ^1 .X0 / exp dR dR0
Z Œ0;k ^1
ˇt
1 dMtR
2
Z Œ0;k ^1
ˇt A.dt/ˇt : (25)
Proof. By conditioning with respect to X0 ; we see that we can assume without loss dP0 of generality, that R0 WD .X0 /# R D .X0 /# P DW P0 ; i.e. dR .X0 / D 1: Let k 1: 0 Denote Rk D Rk ; P k D P k : Applying Lemma 3 with D ˇ and remarking that bˇ D B bˇ ; we see that B bˇ / C B bˇ ; 10;k A/ Qk WD E .ˇ M P /k ^1 P k 2 MP.10;k Œ.B R C B D MP.10;k B R ; 10;k A/: But, it is known with Lemma 4 that Rk satisfies the condition (U). Therefore, Qk D Rk :
(26)
448
C. Léonard
Applying twice Lemma 3, we observe on the one hand that e k WD E .ˇ M R /k ^1 Rk 2 MP.10;k .B R C B bˇ /; 10;k A/; P
(27)
and on the other hand that ek e k WD E .ˇ M P /k ^1 P Q bˇ / B bˇ ; 10;k A/ D MP.10;k B R ; 10;k A/: 2 MP.10;k Œ.B R C B As for the proof of (26), the condition (U) which is satisfied by Rk leads us to e k D Rk : Therefore, we see with (26) that Qk D Q e k ; i.e. E .ˇ M P /k ^1 P k D Q P k P e e k which E .ˇ M /k ^1 P : And since E .ˇ M /k ^1 > 0; we obtain P k D P is (25). t u Remark 4. As a by-product of this proof, we obtain that for all k 1; P k is the bˇ /; 10;k AI P0 / in unique solution of the martingale problem MP.10;k .B R C B the class of all probability measures which are absolutely continuous with respect to Rk : We are ready to complete the proof of Theorem 2. Proof (of Theorem 2. Derivation of dP dR ). Provided that R satisfies the condition (U), when P R we obtain the announced formula Z Z dP dP0 1 R D (28) .X0 / exp ˇt dMt ˇt A.dt/ˇt ; dR dR0 2 Œ0;1 Œ0;1 Rletting k tend to infinity in (25), remarking that WD limk!1 k D infft 2 Œ0; 1I Œ0;t ˇs A.ds/ˇs D 1g and that (7) implies D 1; P -a.s.
(29)
and, since P R; we also have D 1; R-a.s. Indeed, since .!/ D 1; there 5 with k D ko W is some ko 1 such
R that ko .!/ D 1 Rand applying Lemma dP0 dP 1 R .!/ > 0: .!/ D .! / exp ˇ dM ˇ A.dt/ˇ 0 t t Œ0;1 t dR dR0 2 Œ0;1 t b AI P0 / It also follows with Remark 4 that P is the unique solution of MP.B R CB; in the class of all probability measures which are absolutely continuous with respect to R: Now, we consider the general case when P might not be equivalent to R: The main idea is to approximate P by a sequence .Pn /n1 such that Pn R for all n 1; and to rely on our previous intermediate results. We consider 1 1
P C R; Pn WD 1 n n
n 1:
Girsanov Theory Under a Finite Entropy Condition
449
Clearly, Pn R and by convexity H.Pn jR/ .1 n1 /H.P jR/ C n1 H.RjR/ H.P jR/ < 1: More precisely, the function x 2 Œ0; 1 7! H.xP C .1 x/RjR/ 2 Œ0; 1 is a finitely valued convex continuous and increasing. It follows that limn!1 H.Pn jR/ D H.P jR/: As regards the uniqueness statement, suppose that P and Q are two solutions of b AI P0 / such that H.P jR/; H.QjR/ < 1: With n D 2; by linearity MP.B R C B; P2 D .P C R/=2 and Q2 D .Q C R/=2 are again solutions to the same martingale problem. But we already saw a few lines above that this implies that P2 D Q2 : Therefore, P D Q: It is clear that limn!1 Pn D P in total variation norm. Let us prove that the stronger convergence lim H.P jPn / D 0 (30) n!1
also holds. It is easy to check that 1f dP 1g dP =dPn and 1f dP 1g dP =dPn are respecdR dR tively decreasing and increasing sequences of functions. It follows by monotone convergence that Z lim H.P jPn / D lim
n!1
n!1
log.dP =dPn / dP Z
D lim
n!1 f dP 1g dR
log.dP =dPn / dP
Z
C lim
n!1 f dP <1g dR
log.dP =dPn / dP D 0:
By Theorem 1, there exist two vector fields ˇ n and ˇ which are R R respectively defined R-a.s. and P -a.s. such that EPn Œ0;1 ˇtn A.dt/ˇtn < 1, EP Œ0;1 ˇt A.dt/ˇt < 1 and dXt D dBtR CA.dt/ˇtn CdMtPn ; R-a.s.I
dXt D dBtR CA.dt/ˇt CdMtP ; P -a.s.
where M Pn and M P are respectively a local Pn -martingale and a local P -martingale. Therefore, dMtPn D dMtP C A.dt/.ˇt ˇtn /;
P -a.s.
(31)
Extending ˇ arbitrarily by ˇ D 0 on the P -null set where it is unspecified, we know that Z Z 1 .ˇs ˇsn / dMsPn .ˇs ˇsn / A.ds/.ˇs ˇsn / exp 2 Œ0;t Œ0;t is a P n -supermartingale. It follows with Lemma 1, (31) and a standard monotone convergence argument that
450
C. Léonard
Z H.P jPn / EP
.ˇs Œ0;1
ˇsn /
Z 1 n n .ˇs ˇs / A.ds/.ˇs ˇs / 2 Œ0;1 Z 1 .ˇs ˇsn / A.ds/.ˇs ˇsn /: D EP 2 Œ0;1 dMsPn
With (30), this shows the key estimate Z lim EP
n!1
Œ0;1
.ˇs ˇsn / A.ds/.ˇs ˇsn / D 0:
(32)
Since H.Pn jR/ < 1 and Pn R; under the condition (U) we can invoke (28) to write Z Z dPn;0 dPn 1 n R n n D .X0 / exp ˇt dMt ˇ A.dt/ˇt : dR dR0 2 Œ0;1 t Œ0;1 As limn!1 Pn D P in total variation norm, up to the extraction of a R-a.s.-convergent subsequence we have limn!1 dPn =dR D dP =dR and limn!1 dPRn;0 =dR D dP0 =dR0 :R On the other hand, (32) implies that P -a.s.; limn!1 21 Œ0;1 ˇtn A.dt/ˇtn D 12 Œ0;1 ˇt A.dt/ˇt : It follows that dP dP0 D 1f dP >0g .X0 / exp dR dR dR0
Z Œ0;1
ˇt dMtR
1 2
ˇt A.dt/ˇt :
Z Œ0;1
where (32) also implies that the limit of the stochastic integrals Z
Z lim
n!1 Œ0;1
ˇtn dMtR D
exists P -a.s.
Œ0;1
ˇt dMtR ; P -a.s. t u
It remains to compute H.P jR/: Proof (End of the proof of Theorem 2. Computation of H.P jR/). Let us first compute H.P jR/ when R satisfies (U). Remark that in the proof of Lemma 5, for b which is behind (27) is a genuine e k -martingale N k D M R B all k 1 the local P ek D P k ; martingale. It is a consequence of the first statement of Lemma 3. As P N k is a genuine P k -martingale. This still holds when P R fails. Indeed, this hypothesis has only been invoked to insure that k is well-defined R-a.s. But in the present situation, k only needs to be defined P -a.s. With (25), we have dP k H.P k jRk / D EP k log dRk Z Z dP0 1 (25) D EP log .X0 / CEP k ˇt dMtR ˇt A.dt/ˇt dR0 2 Œ0;1 Œ0;1
Girsanov Theory Under a Finite Entropy Condition
451
Z
Z 1 b D H.P0 jR0 / C EP k ˇt C d Bt / ˇt A.dt/ˇt 2 Œ0;1 Œ0;1 Z Z 1 (8) k D H.P0 jR0 / C EP k ˇt A.dt/ˇt C EP k ˇt dNt 2 Œ0;1 Œ0;1 Z 1 D H.P0 jR0 / C EP ˇt A.dt/ˇt 2 Œ0; k ^1
(27)
.dNtk
where the last equality comes from the P k -martingale property of N k . It remains to let k tend to infinity to see that 1 H.P jR/ D H.P0 jR0 / C EP 2
Z
ˇt A.dt/ˇt : Œ0;1
Indeed, because of (29) and since the sequence .k /k1 is increasing, we obtain by monotone convergence that Z lim EP
k!1
Œ0; k ^1
ˇt A.dt/ˇt
1 D EP 2
Z
ˇt A.dt/ˇt : Œ0;1
As regards the left hand side of the equality, with Lemma 1 and (29), we see that H.P jR/ D supfEP u.X / log ER e u.X / I u 2 L1 .P /g k
D sup supfEP u.X / log ER e u.X
k /
I u 2 L1 .P /g
k
D lim H.P k jRk /: k!1
It remains to check that, without the condition (U), we have 1 H.P jR/ H.P0 jR0 / C EP 2
Z
ˇt A.dt/ˇt :
(33)
Œ0;1
Let us extend ˇ by ˇ D 0 on the P -null set where it is unspecified and define uQ .X / WD log
dP0 .X0 / C dR0
Z Œ0; k ^1
ˇt dMtR
.i/
1 2
Z Œ0; k ^1
ˇt A.dt/ˇt :
Choosing uQ .X / at inequality below, thanks to an already used supermartingale .ii/
argument, we obtain the inequality below and
452
C. Léonard
Z
(18)
H.P jR / D sup k
k
.i/
Z
.ii/
Z
Z u dP log Z
uQ dP k log
Z e dR I u W
k
u
e dR < 1
k
k
u
e uQ dRk
uQ dP k Z
.iii/
D H.P0 jR0 / C EP k 1 (8) D H.P0 jR0 / C EP k 2
Œ0; k ^1
bt ˇt d B
Z
Œ0; k ^1
1 2
Z Œ0; k ^1
ˇt A.dt/ˇt
ˇt A.dt/ˇt :
Equality (iii) is a consequence of uQ .X / D log
dP0 .X0 /C dR0
Z
bt / 1 ˇt .dMtP Cd B 2 Œ0; k ^1
Z Œ0; k ^1
ˇt A.dt/ˇt ; P k -a.s.
which comes from Theorem 1. It remains to let k tend to infinity, to obtain as above with (29) that (33) holds true. This completes the proof of the theorem. t u
6 Proofs of Theorems 3 and 4 We begin recalling Itô’s formula. Let P be the law of a semimartingale dXt D bt .dt/ C dMtP with M P a local P -martingale such that M P D q ˇe K , P -a.s. That is P 2 LK.K/ 2 d for some Lévy kernel K: For any f in C .R / which satisfies: ( ) When localizing with an increasing sequence .k /k1 of stopping times tending P -almost surely to infinity, for each k 1 the truncated process 1fjqj>1g 1ft k g Œf .Xt C q/ f .Xt / is a H1 .P; K/ integrand, Itô’s formula is df .Xt / D
hZ Rd
i Œf .Xt C q/ f .Xt / rf .Xt / q Kt .dq/ .dt/
Crf .Xt / bt .dt/ C dMt ;
P -a.s.
(34)
where M is a local P -martingale. This identity would fail if was not assumed to be atomless.
Girsanov Theory Under a Finite Entropy Condition
453
6.1 Proof of Theorem 3 Based on Itô’s formula, we start computing a large family of exponential local martingales. Recall that we denote a 7! .a/ WD e a a 1 D
X
an =nŠ;
a 2 R:
n2
Lemma 6 (Exponential martingale). Let h W ˝ Œ0; 1Rd ! R be a real valued predictable process which satisfies Z Œht .q/ L.dtdq/ < 1:
ER
(35)
Œ0;1R
Then, h and e h 1 belong to H1;2 .R; L/. In particular, h ˇ e L is a R-martingale. Moreover, Z
h L Zt WD exp h ˇ e t Œhs .q/ L.dsdq/ ; t 2 Œ0; 1 .0;t Rd
is a local R-martingale and a positive R-supermartingale which satisfies dZth D Zth Œ.e h.q/ 1/ ˇ de L t : Proof. The function is nonnegative, quadratic near zero, linear near 1 and it grows exponentially fast near C1: Therefore, (35) implies that h and e h 1 belong h to H1;2 .R; L/. In particular, L is a R-martingale. R M WD h ˇ e R h Let us denote Yt D Mt .0;t ˇs .ds/ where ˇt D Rd Œht .q/ Lt .dq/: Remark that (35) implies that these integrals are almost everywhere well-defined. Applying (34) with f .y/ D e y and d Yt D ˇt .dt/ C dMth , we obtain Z h de Yt D e Yt ˇt C
Rd
i Œht .q/ Lt .dq/ .dt/ C dMt D dMt
where M is a local martingale. We are allowed to do this because ( ) is satisfied. Indeed, with f .y/ D e y ; f .Yt C ht .q// f .Yt / f 0 .Yt /ht .q/ D e Yt Œht .q/ and if Yt WD Yt ^ is stopped at WD infft 2 Œ0; 1I Yt 62 C g 2 Œ0; 1[f1g for some compact subset C with the convention inf ; D 1; we see with (35) and the fact that any path in ˝ is bounded, that exp.Yt / Œht .q/ is in H1 .R; L/: Now, choosing the compact set C to be the ball of radius k and letting k tend to infinity, we obtain an increasing sequence of stopping times .k /k1 which tends almost surely to infinity. This proves that Z h WD e Y is a local martingale. We see that dMt D e Yt d Œ.e h.q/1 / ˇ e L t ; keeping track of the martingale terms in the above differential formula:
454
C. Léonard
de Yt D e Yt .Yt / C d Yt
L D e Yt Œht .q/ ˇ de t C
Z Rd
Œht .q/ Lt .dq/ .dt/
L ˇt .dt/ C h.q/ ˇ de t
D e Yt Œht .q/ ˇ de L L t C ht .q/ ˇ de t
L D e Yt .e ht .q/ 1/ ˇ de t : By Fatou’s lemma, any nonnegative local martingale is also a supermartingale. Proof (of Theorem 3). It follows the same line as the proof of Theorem 1. By Lemma 6, 0 < ER Z1h 1 for all h satisfying the assumption (35). By (18), for any probability measure P such that H.P jR/ < 1; we have EP
Z hˇe L 1
.h/ d L H.P jR/: Œ0;1Rd
As in the proof of Theorem 1, see that jEP .h ˇ e L 1 /j .H.P jR/ C 1/khk ; where
khk WD inf a > 0I EP
8h
Z
.h=a/ d L 1 2 Œ0; 1
(36)
Œ0;1Rd
is the Luxemburg norm of the Orlicz space Z n L W D h W Œ0; 1 Rd ˝ ! RI measurable s.t. EP
Œ0;1Rd
o for some bo > 0 :
.bo jhj/ d L < 1;
It differs from the corresponding small Orlicz space Z n S W D h W Œ0; 1Rd ˝!RI measurable s.t.EP
.bjhj/d L<1; 8b0
o
Œ0;1Rd
because the function .jaj/ grows exponentially fast. R We introduce the space B of all the bounded processes such that EP Œ0;1Rd jhj
d L < 1; and its subspace H B which consists of the processes in B which are predictable. We have B S and any h in H satisfies (35), which is the hypothesis of Lemma 6. Hence, (36) holds for all h 2 H and, as H.P jR/ < 1; it tells us that the linear mapping h 7! EP .h ˇ e L 1 / is continuous on H equipped with the
Girsanov Theory Under a Finite Entropy Condition
455
norm k k : Since the convex conjugate of the Young function .jaj/ is .jbj/; the dual space of .S ; k k /2 (see [5]), is isomorphic to Z n L WD k W Œ0; 1 Rd ˝ ! RI measurable s.t. EP
o .jkj/ d L < 1 :
Œ0;1Rd
Therefore, there exists some k 2 L such that Z L 1 D EP kh d L; EP h ˇ e Œ0;1Rd
8h 2 H :
(37) pr
Let us introduce the predictable projection k pr of k which is defined by kt WD EP .k j XŒ0;t / /; t 2 Œ0; 1: As the space B is dense in S ,3 H is dense in the subspace of all predictable processes in S and it follows that any g and k in L which both satisfy (37), share the same predictable projection: g pr D k pr : Consequently, there is a unique predictable process k in the space n
K .P / WD k W
Z Œ0; 1Rd ˝
! RI predictable s.t. EP
.jkj/ d L < 1
o
Œ0;1Rd
which verifies (37). As H is included in H1 .P; L/; we have for all h 2 H ; h ˇ e L h ˇ kL D X X h ˇ . L h ˇ kL D h ˇ . `L/ with ` WD k C 1: Consequently, (37) is equivalent to
EP h ˇ .X `L/ D 0; 8h 2 H ; (38) which is the content of the theorem. It remains however to note that, being an expectation of the positive measure X , `L is also a positive measure. Therefore, ` is nonnegative. This completes the proof of the theorem. t u
6.2 Proof of Corollary 2 It is mainly a remark based on Hölder’s inequality in Orlicz spaces. Proof (of Corollary 2). We are under the exponential integrability assumption (14) and we denote Z D dP dR : The finite entropy assumption (1) is equivalent to: Z belongs to the Orlicz space L .R/; i.e. kZk ;R < 1: Hölder’s inequality in Orlicz spaces4 expressed with the Luxemburg norms (see (36)) gives us for any nonnegative random variable U : EP .U / D ER .ZU / 2kZk ;R kU k ;R : This quantity is finite if kU k ;R < 1; and this is equivalent to ER .e ao U / < 1 for some
2
This doesn’t hold with L instead of S . In general, it is not dense in L : 4 It is an easy consequence of Fenchel’s inequality: jabj .jaj/ C .jbj/; for all a; b 2 R: 3
456
C. Léonard
R ao > 0: As a consequence, (14) implies that EP Œ0;1Rd 1fjqj1g e bo jqj L.dtdq/ < 1 for some bo : But this is equivalent to: 1fjqj1g jqj belongs to the Orlicz space L .P ˝ L/: With (13) we see that .`1/ is in L .P ˝L/ and by Hölder’s inequality again, we obtain Z EP 1fjqj1g jqjj`.t; q/ 1j L.dtdq/ < 1: Œ0;1Rd
R The small jump part: EP Œ0;1Rd 1fjqj<1g jqjj`.t; q/ 1j L.dtdq/ < 1; is a direct consequence of Hölder’s inequality in L2 : This proves (15). We write symbolically C .` 1/L: e L D L D `L C .` 1/L D b R Hence, q ˇ e L D q ˇ b C .` 1/q d L provided that all these terms are well defined. L is well-defined and we have just proved R But, we have assumed that q ˇ e that .`1/q d L is well-defined. Therefore, the remaining term is also well-defined and the proof is complete. t u
6.3 Proof of Theorem 4 It is similar to the proof of Theorem 2. We begin with a transfer result in the spirit of Lemma 3. Let P be a probability measure on ˝ such that P 2 MP.B; K/ where B is a continuous bounded variation adapted process and K is some Lévy kernel K.dtdq/ WD .dt/K.tI dq/: Let
be a Œ1; 1/-valued predictable process on Œ0; 1 Rd such that R f 1g . / d K < 1 and K.1 < 1/ < 1; P -a.s. We define for all t 2 Œ0; 1; Z K . / d K WD ZtC Zt Zt D exp ˇ e t
e
Œ0;t Rd
with Z 8 ˆ C C K ˆ Z D exp
ˇ e ˆ t t < ˆ ˆ ˆ : Zt D 1ft < g exp where
C
. /d K Z X
.s; Xs /
0st
!
.0;t Rd
.0;t Rd
.e
1/ d K
Girsanov Theory Under a Finite Entropy Condition
C D 1f ˛g ;
457
D 1f1 <˛g
with ˛ > 0; e 1 D 0 and D infft 2 Œ0; 1; .t; Xt / D 1g: Remark that, although Z C and Z both depend on the choice of ˛; their product Z D Z C Z doesn’t depend on ˛ > 0: For all j; k 1; we define Z jk WD inf t 2 Œ0; 1I
Œ0;t Rd
. C / d K kor .t; Xt / 62 Œj; k 2Œ0; 1 [ f1g
k
and Pjk WD X j # P: k
Lemma 7. Let P and be as above. Then, for all j; k 1; Z j is a genuine P -martingale and the measure k
Qjk WD Z1 j P k is a probability measure on ˝ which satisfies k
k bj ; 1 k e K Qjk 2 MP B j C B 0; j
where bt D B
Z Œ0;t Rd
1fjqj1g .e 1/q d K;
t 2 Œ0; 1:
(39)
bt might not be well defined in the general case. Only the stopped Note that B jk b processes B are asserted to be meaningful. R k Proof. Let us fix j; k 1: We have Z j D exp. kj ˇ e K Œ0;1Rd . kj / d K/ with kj D 10; k which is predictable since is predictable and 10; k is left
e
j
j
continuous. We drop the subscripts and superscripts j; k and write D kj ; C D k
. kj /C ; D . kj / ; Z j D Z for the remainder of the proof. By the definition of jk ; we obtain with this simplified notation Z
. C / d K k;
j k;
Œ0;1Rd
Pjk -a.s.
(40)
Let us first prove that Z is a Pjk -martingale. Since it is a local martingale, it is enough to show that p
EP k Z1 < 1; j
for some p > 1:
458
C. Léonard k
k
k
Choosing ˛ D j in the definition of .Z j /C and .Z j / ; we see that Z j D k C K /: For all p 0; .Z j /C D Z C D E ..e 1/ ˇ e Z C K p .Z / D exp p ˇ e C p
C
. / d K Œ0;1Rd
exp.p C ˇ e K /
and Z C E ..e p 1/ ˇ e K / D exp p C ˇ e K
.p C / d K
Œ0;1Rd
e =C.k; p/
C ˇK
e p
for some finite deterministic constant C.k; p/ > 0: To derive C.k; p/; we must take account of (40) and rely upon the inequality .pa/ c.k; p/ .a/ which holds for all a 2 .1; k and some 0 < c.k; p/ < 1: With this in hand, we obtain e C.k; p/E ..e p C 1/ ˇ e K /:
C ˇK
.Z C /p e p
C
We know with Lemma 6 that E ..e p 1/ ˇe K / is a nonnegative local martingale. C Therefore, it is a supermartingale. We deduce from this that EP k E ..e p 1/ ˇ j
e K / 1 and C
EP k .Z C /p C.k; p/EP k E ..e p 1/ ˇ e K / C.k; p/ < 1: j
j
C
K / is uniformly integrable. We Choosing p > 1; it follows that E ..e 1/ ˇ e
conclude as in Lemma 3’s proof that E ..e 1/ ˇ e K / is a genuine Pjk -martingale. Now, let us show that
Qjk 2 LK 10; k e K : j
Let be a finitely valued stopping time and f aP measurable function on Œ0; 1 Rd which will be specified later. We denote Ft D 0st ^ f .s; Xs / with the convention that f .t; 0/ D 0 for all t 2 Œ0; 1: By Lemma 6, the martingale Z satisfies dZt D 10; k .t/Zt Œ.e 1/ ˇ e K . We have also dFt D 10; .t/f .t; Xt / and j
d ŒZ; F t D 10; k ^ .t/Zt .e .Xt / 1/f .t; Xt /; Pjk -a.s. Consequently, j
EQ k j
X
f .t; Xt /
0t
D EP k .Z F Z0 F0 / j Z D EP k .Ft dZt C Zt dFt C d ŒZ; F t / j
Œ0;
Girsanov Theory Under a Finite Entropy Condition
"Z D EP k
Ft dZt C
j
D EP k j
Œ0;
X
X
459
X
Zt f .t; Xt /C
0t
# Zt .e
.t;Xt /
1/f .t; Xt /
0t
Zt e .t;Xt / f .t; Xt /
0t
Z D EP k j
Œ0; Rd
Zt f .t; q/e .t;q/ K.dtdq/
Z D EQ k
f .t; q/e .t;q/ K.dtdq/:
j
Œ0; Rd
We are going to choose such that P the above terms are meaningful. For each n 1; consider n WD infft 2 Œ0; 1I 0st ^ jf .s; Xs /j ng and take f in L1 .Pjk ˝ K/ to obtain limn!1 n D 1; Pjk -a.s. and a fortiori Qjk -a.s. It remains to take D ^ n with any stopping time to see that the Lévy kernel of Qjk is k
e K D e j K: P It remains to compute the drift term. Let us denote Xt WD 0st 1fjXs j>1g Xs the cumulated sum of large jumps of X; and X 4 WD X X its complement. Let be a finitely valued stopping time and take Gt D Xt4^ with 2 Rd : We have dGt D 10; .t/ .dBt C .1fjqj1g q/ ˇ de K t / and d ŒZ; Gt D
.X / k t 10; k ^ .t/Zt .e 1/1fjXt j1g Xt ; Pj -a.s. Therefore, j
EQk Œ .X4 X04 / j
D EP k ŒZ G Z0 G0 j Z D EP k .Gt dZt C Zt dGt C d ŒZ; Gt / j
D EP k
hZ
j
C
X
Œ0;
Z Gt dZt C
Œ0;
Œ0;
Zt .dBt C .1fjqj1g q/ ˇ de K t /
Zt 1fjXt j1g .e .t;Xt / 1/ Xt
i
0t
"Z
D EP k
Zt dBt C
j
Œ0;
Z j
Œ0;
nZ dBt C
Œ0;
Zt 1fjXt j1g .e
Z Zt dBt C
j
#
0t
Z D EP k D EQ k
X
Zt Œ0;
Rd
nZ Rd
.t;Xt /
1/ Xt
o 1fjqj1g .e .t;q/ 1/ qKt .dq/ .dt/
o 1fjqj1g .e .t;q/ 1/q Kt .dq/ .dt/
460
C. Léonard
where we take D n WD infft 2 Œ0; 1I jXt j ng which tends to 1 as n tends b k where B b is to infinity. This shows that the drift term of X under Qjk is .B C B/ bk is well-defined. given at (39) and the stopped process B As a first step, it is assumed that P R for the stopping times jk , j and to be defined (below) R-a.s. and not only P -a.s. Following the proofs of Lemmas 4 and 5, except for minor changes (but we skip the details), we arrive at analogous results: (i) If R fulfills the uniqueness condition (U), then for any stopping time ; R also fulfills (U). (ii) If P R; then for any j; k 1; we have 10; k ^1 j
dP dP0 D 10; k ^1 .X0 / exp .1.0; k ^1 log `/ ˇ e L j j dR dR0 ! Z .log `/ d L .0;jk ^1Rd
where Z jk W D inf t 2 Œ0; 1I
Œ0;t Rd
1f`>1=2g .log `/ d L k or log `.t; Xt /
62 Œj; k 2 Œ0; 1 [ f1g: For the proof of (ii), we use Lemma 6 where D log ` plays the same role as ˇ in Lemma 5, and we go backward with which corresponds to `1 . We fix j; and let k tend to infinity to obtain with (13) that ˚ lim jk D j WD inf t 2 Œ0; 1I `.t; Xt / < e j 2 Œ0; 1 [ f1g;
k!1
P -a.s.
and therefore R-a.s. also. More precisely, this increasing sequence is stationary after some time: there exists K.!/ < 1 such that jk .!/ D j .!/; for all k K.!/: It follows that for all j 1; dP dP0 D 10;j ^1 .X0 / exp 1.0;j ^1 log ` ˇ e L 10;j ^1 dR dR0 ! Z .log `/ d L :
(41)
.0;j ^1Rd
Lemma 8. We do not assume that P R and we extend ` by ` D 1 on the P -negligible subset where it is unspecified. Defining WD supj 1 j ; we have P . D 1/ D 1:
Girsanov Theory Under a Finite Entropy Condition
Proof. For all j 1; we have 1 ) X
P . 1/ P X
P t 1
1f`.t;Xt /ej g 1: Therefore,
! 1f`.t;Xt /ej g 1 EP
t 1
Z
D EP
461
Œ0;1Rd
X
1f`.t;Xt /ej g
t 1
1f`ej g `d L e j EP L.` e j / e j EP L.` 1=2/
where we used (38) at the marked equality. The result will follow letting j tend to infinity, provided that we show that R EP L.` 1=2/ < 1: But, we know with (13) that EP Œ0;1Rd .j` 1j/ d L < 1: Hence, EP L.` R 1=2/ EP Œ0;1Rd .j` 1j/ d L= .1=2/ < 1 and the proof is complete. u t Lemma 9. Assume P R: Let Rj and Pj be the laws of the stopped process X j ^1 under R and P respectively. Then, under the condition (U) we have for all j 1 Z H.Pj jRj / D H.P0 jR0 / C EP
.` log ` ` 1/ d L: .0;j ^1Rd
Proof. We denote Rjk and Pjk the laws of the stopped process X j ^1 under R and on 0; jk ^ 1 we see that P respectively. With the expression of dP dR k
H.Pjk jRjk /
!
Z D H.P0 jR0 / C EP k .1.0; k ^1 log `/ ˇ e L
j
.log `/ d L
j
.0;jk ^1Rd
!
Z D H.P0 jR0 /CEP k .1.0; k ^1 log `/ ˇ e
`L
j
j
Z D H.P0 jR0 / C EPj
C
Œ.` 1/ .log `/d L .0;jk ^1Rd
.` log ` ` 1/ d L .0;jk ^1Rd
where we invoke Lemma 7 at the last equality. We complete the proof letting k tend to infinity. t u Proof (Conclusion of the proof of Theorem 4). When P R; by Lemma 8, P -almost surely there exists jo large enough such that for all j jo ; j D 1 and (41) tells us that Z dP dP0 L D .X0 / exp .log `/ ˇ e .log `/ d L dR dR0 Œ0;1Rd
462
C. Léonard
and also that the product appearing in Z contains P -almost surely a finite number of terms which are all positive. Note that we do not use any limit result for stochastic or standard integrals; it is an immediate !-by-! result with a stationary sequence. This is the desired expression for dP dR when P R: Let us extend this result to the case when P might not be equivalent to R: We proceed exactly as in Theorem 2’s proof and start from (30): limn!1 H.P jPn / D 0 where Pn WD .1 1=n/P C R=n; n 1: Let us write D log ` and n D log `n which are well-defined P -a.s. Thanks to Theorem 3, we see that Z n H.P jPn / EP . n / ˇ e ` L . n / `n d L Z D EP . n / ˇ e `L C Z
Œ`=` log.`=` / `=` C 1 ` d L n
n
n
n
Œ0;1Rd
Œ`n =` log.`n =`/ 1 d `L
D EP Z D EP
Œ0;1Rd
Œ0;1Rd
. n / d `L Œ0;1Rd
which leads to the entropic estimate analogous to (32): Z lim EP
n!1
. n / d `L D 0:
(42)
Œ0;1Rd
R Taking the difference between log.dPn =dR/ D n ˇ e L Œ0;1Rd . n / d L and > 0g; the logarithm of the announced formula (16) for dP =dR on the set f dP dR we obtain Z . n / ˇ e `L . n / d `L; P -a.s. Œ0;1Rd
and the desired convergence follows from (42). Note that .a/ D a2 =2 C oa!0.a2 /: This completes the proof of (16). As in the proof of Theorem 2, we obtain the announced formula for H.P jR/ under the condition (U) with Lemmas 8 and 9, and the corresponding general inequality follows from choosing dP0 uQ .X / WD log .X0 / C .1.0; k ^1 log `/ ˇ e L j dR0
Z
.log `/ d L
.0;jk ^1Rd
in the variational representation formula (18), and then letting k and j tend to infinity. u t
Girsanov Theory Under a Finite Entropy Condition
463
Appendix. An Exponential Martingale with Jumps Next proposition is about exponential martingale with jumps. We didn’t use it during the proofs of this paper. But we give it here for having a more complete picture of the Girsanov theory. In this result, integrands h are considered which may attain the value 1: This is because with h D log `, h D 1 corresponds to ` D 0: Proposition 1 (Exponential martingale). Let h W ˝ Œ0; 1 Rd ! Œ1; 1/ be an extended real valued predictable process which may take the value 1 and satisfies Z ER Z
Œ0;1R
ER
1fht .q/1g Œht .q/ L.dtdq/ < 1;
(43)
1fht .q/<1g L.dtdq/ < 1:
(44)
Œ0;1R
Let us introduce the stopping time h WD infft 2 Œ0; 1I h.Xt / D 1g 2 Œ0; 1 [ f1g and the convention e 1 D 0: Then, e h 1 is in H1;2 .R; L/ and Z Zth WD 1ft < h g exp h ˇ e L t
e
.0;t Rd
Œhs .q/ L.dsdq/ ;
t 2 Œ0; 1
(45)
is a local R-martingale and a nonnegative R-supermartingale which satisfies dZth D 1ft h g Zth Œ.e h.q/ 1/ ˇ de L t :
(46)
The standard notation is Z h WD E .Œe h 1 ˇ e L /; the stochastic exponential of L Œe 1 ˇ e : Some details are necessary to make precise the sense of the inner h stochastic integral h ˇ e L t in the expression of Zt : We denote h
hC WD 1fh1g h 2 R h WD 1fh<1g h 2 Œ1; 0: Under the assumption (43), hC ˇ e L is well defined as a stochastic integral. On the other hand, implies that h .t; Xt / has R-a.s. finitely many jumps. P (44) It follows that 0st h .s; Xs / is meaningful for all t < h . But the integral R ˇ e L D t .0;t Rd hs .q/ L.dsdq/ might not be defined under (44) and h R P h .s; X / h .q/ L.dsdq/ is meaningless in this case. s 0st .0;t Rd s R Nevertheless, the full expression in the exponential .h/ WD h ˇ e L .h/ d L
464
C. Léonard
R P is defined as follows. We put .h / WD 0st h .s; Xs / .0;t Rd Œe hs .q/ 1 L.dsdq/ the R which is well defined under (44) and is obtained by cancelling C terms .0;t Rd h .q/ L.dsdq/. As .0/ D 0; we have
.h/ D
.h C h / D s C
.h / C .h / and for all t 2 Œ0; 1; 8 C ˆ Z h D Zth Zth with ˆ
< Ct R h C C Zt WD exp h ˇ e L t .0;t Rd Œhs .q/ L.dsdq/ ;
ˆ R ˆ h : Z h WD 1 h exp P s .q/ 1 L.dsdq/ : h .s; X / Œe d s ft < g t 0st .0;t R (47) This is what is meant by the concise expression (45). Proof. Now, we consider the general case where h may attain the value 1 and (35) is weakened by (43) and (44). We use the decomposition (47) and write C Z C D Z h and Z D Z h for short. Clearly, Z C and Z do not jump at the same times and d ŒZ C ; Z D Z C Z D 0: Hence, dZt D ZtC dZt C Zt dZtC :
(48)
The hC -part enters the framework of Lemma 6 and we have C
dZtC D ZtC Œe h 1 ˇ e L :
(49)
Let us look at the h -part. We need to compute dZt : For all t < h ; put Yt D
X 0st
h .s; Xs /
Z
.q/
Œe hs
1 L.dsdq/:
.0;t Rd
Then, with convention that h .t; 0/ D 0; d Yt D h .t; Xt / t .dt/ with R the h .q/ t t D Rd Œe 1 Lt .dq/; Yt D h .t; Xt / and with Itô’s formula, we arrive at
de Yt D e Yt Œe Yt 1 C d Yt Yt D e Yt Œe h .t;Xt / 1 t .dt/
L D e Yt Œe h 1 ˇ de t : It follows that
dZt D Zt Œe h 1 ˇ de L t ;
t < h:
At t D h ; by the Definition (47) of Z ; we have 1 dZjtD h D Z. 1 h / D Z. h / Œe
(50)
Girsanov Theory Under a Finite Entropy Condition
465
which is (50) at t D h with the convention e 1 D 0: This provides us with
dZt D 1ft h g Zt Œe h 1 ˇ de L t : Together with (48) and (49), this proves (46) which implies that Z h is a local R-martingale. By Fatou’s lemma, any nonnegative local martingale is also a supermartingale.
References 1. J. Jacod, Multivariate point processes: Predictable representation, Radon-Nikodým derivatives, representation of martingales. Z. Wahrsch. verw. Geb. 31, 235–253 (1975) 2. J. Jacod, Calcul stochastique et problèmes de martingales. Lecture Notes in Mathematics, vol. 714 (Springer, Berlin, 1979) 3. J. Jacod, A.N. Shiryaev, Limit Theorems for Stochastic Processes. Grundlehren der mathematischen Wissenshaften, vol. 288 (Springer, Berlin, 1987) 4. P.E. Protter, Stochastic Integration and Differential Equations. Applications of Mathematics. Stochastic Modelling and Applied Probability, vol. 21, 2nd edn. (Springer, Berlin, 2004) 5. M.M. Rao, Z.D. Ren, Theory of Orlicz Spaces. Pure and Applied Mathematics, vol. 146 (Marcel Dekker, New York, 1991) 6. D. Revuz, M. Yor, Continuous Martingales and Brownian Motion. Grundlehren der Mathematischen Wissenschaften, vol. 293, 3rd edn. (Springer, Berlin, 1999)
Erratum to Séminaire XXVII Michel Émery and Marc Yor
As pointed out by Vilmos Prokaj, the contribution On the Lévy transformation of Brownian motions and continuous martingales in Séminaire XXVII, by the two of us and our late friend Lester E. Dubins, contains a serious error: the proposition page 128 is false. The computation (not given in the article) which leads to this proposition is not sound, and does not seem to be mendable. Fortunately, Remark c) pp. 127–128, which contains this false statement, is completely independent from the rest of the article, so chaos does not propagate. We thank Vilmos Prokaj for this observation, and apologize to all other readers— if any.
M. Émery IRMA, CNRS et Université Unique de Strasbourg, 7 rue René Descartes, 67084 Strasbourg Cedex, France e-mail: [email protected] M. Yor LPMA, Université Pierre et Marie Curie, boîte courrier 188, 4 place Jussieu, 75252 Paris Cedex 05, France C. Donati-Martin et al. (eds.), Séminaire de Probabilités XLIV, Lecture Notes in Mathematics 2046, DOI 10.1007/978-3-642-27461-9, © Springer-Verlag Berlin Heidelberg 2012
467
Erratum to Séminaire XXXV Michel Émery and Walter Schachermayer
In Séminaire XXXV, our contribution On Vershik’s standardness criterion and Tsirelson’s notion of cosiness contains a material error: in Lemma 17 on top of page 284, we forgot to write that the random variable R is assumed to be simple. Simplicity is used in the proof, but not mentioned in the statement. This error bears no consequence on the rest of the article, for two reasons. Firstly, Lemma 17 is used only once, in the proof of Lemma 18, and the r.v. R there is simple; so the whole argument is sound. And secondly, Lemma 17 remains true if R is not simple; but this is a consequence of Lemma 18, whose proof uses the simple case (Lemma 17) as a preliminary step. We apologize to any reader whose admiration for Vershik’s truly marvellous theory might have been hindered.
M. Émery IRMA, CNRS et Université Unique de Strasbourg, 7 rue René Descartes, 67084 Strasbourg Cedex, France e-mail: [email protected] W. Schachermayer Fakultät für Mathematik, Universität Wien, Nordbergstraße 15, 1090 Wien, Österreich e-mail: [email protected] 469