Large Deviation Principle for Markov Chains in Continuous Time - Inria - Institut national de recherche en sciences et technologies du numérique Accéder directement au contenu
Rapport (Rapport De Recherche) Année : 2000

Large Deviation Principle for Markov Chains in Continuous Time

Résumé

Let $E$ be a denumerable state space, $Y_t$ be an homogeneous Markov process on $E$ with generator $R$. We introduce the \em empirical generator $G_t$ of $Y_t$, and prove strong LDP local bounds for it. This allows to prove the weak LDP in a very general setting, for irreducible non-explosive Markov processes, not necessarily ergodic. Sanov's theorem is obtained by a contraction argument from the weak LDP for $G_t$. In our opinion this is an improvement with respect to the existing literature, since LDP in the Markov case requires in general, either $E$ to be finite, or strong uniformity conditions, which important classes of chains do not verify, e.g. bounded jump networks. Moreover the empirical generator together with the representation of the rate function as an entropy allow to prove nice properties (uniqueness, continuity, convexity). It also leads to applications in simulation (importance sampling) and in the evaluation of the rate function for sample path LDP in networks. Finally it seems that some technical problems can be reduced to convex programs which can be run with fast algorithms.
Fichier principal
Vignette du fichier
RR-3877.pdf (447.35 Ko) Télécharger le fichier

Dates et versions

inria-00072776 , version 1 (24-05-2006)

Identifiants

  • HAL Id : inria-00072776 , version 1

Citer

Arnaud de La Fortelle. Large Deviation Principle for Markov Chains in Continuous Time. [Research Report] RR-3877, INRIA. 2000. ⟨inria-00072776⟩
183 Consultations
213 Téléchargements

Partager

Gmail Facebook X LinkedIn More