Large Deviation Principle for Markov Chains in Continuous Time - Archive ouverte HAL Access content directly
Reports (Research Report) Year : 2000

Large Deviation Principle for Markov Chains in Continuous Time

Abstract

Let $E$ be a denumerable state space, $Y_t$ be an homogeneous Markov process on $E$ with generator $R$. We introduce the \em empirical generator $G_t$ of $Y_t$, and prove strong LDP local bounds for it. This allows to prove the weak LDP in a very general setting, for irreducible non-explosive Markov processes, not necessarily ergodic. Sanov's theorem is obtained by a contraction argument from the weak LDP for $G_t$. In our opinion this is an improvement with respect to the existing literature, since LDP in the Markov case requires in general, either $E$ to be finite, or strong uniformity conditions, which important classes of chains do not verify, e.g. bounded jump networks. Moreover the empirical generator together with the representation of the rate function as an entropy allow to prove nice properties (uniqueness, continuity, convexity). It also leads to applications in simulation (importance sampling) and in the evaluation of the rate function for sample path LDP in networks. Finally it seems that some technical problems can be reduced to convex programs which can be run with fast algorithms.
Fichier principal
Vignette du fichier
RR-3877.pdf (447.35 Ko) Télécharger le fichier

Dates and versions

inria-00072776 , version 1 (24-05-2006)

Identifiers

  • HAL Id : inria-00072776 , version 1

Cite

Arnaud de La Fortelle. Large Deviation Principle for Markov Chains in Continuous Time. [Research Report] RR-3877, INRIA. 2000. ⟨inria-00072776⟩
187 View
213 Download

Share

Gmail Facebook X LinkedIn More