We use cookies to distinguish you from other users and to provide you with a better experience on our websites. Close this message to accept cookies or find out how to manage your cookie settings.
To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Based on the long-running Probability Theory course at the Sapienza University of Rome, this book offers a fresh and in-depth approach to probability and statistics, while remaining intuitive and accessible in style. The fundamentals of probability theory are elegantly presented, supported by numerous examples and illustrations, and modern applications are later introduced giving readers an appreciation of current research topics. The text covers distribution functions, statistical inference and data analysis, and more advanced methods including Markov chains and Poisson processes, widely used in dynamical systems and data science research. The concluding section, 'Entropy, Probability and Statistical Mechanics' unites key concepts from the text with the authors' impressive research experience, to provide a clear illustration of these powerful statistical tools in action. Ideal for students and researchers in the quantitative sciences this book provides an authoritative account of probability theory, written by leading researchers in the field.
This chapter describes how to model multiple discrete quantities as discrete random variables within the same probability space and manipulate them using their joint pmf. We explain how to estimate the joint pmf from data, and use it to model precipitation in Oregon. Then, we introduce marginal distributions, which describe the individual behavior of each variable in a model, and conditional distributions, which describe the behavior of a variable when other variables are fixed. Next, we generalize the concepts of independence and conditional independence to random variables. In addition, we discuss the problem of causal inference, which seeks to identify causal relationships between variables. We then turn our attention to a fundamental challenge: It is impossible to completely characterize the dependence between all variables in a model, unless they are very few. This phenomenon, known as the curse of dimensionality, is the reason why independence assumptions are needed to make probabilistic models tractable. We conclude the chapter by describing two popular models based on such assumptions: Naive Bayes and Markov chains.
The phenomenological theory proposed by Einstein for interpreting the phenomenon of Brownian motion is described in detail. The alternative approaches due to Langevin and Fokker–Planck are also illustrated. The theory of Markov chains is also reported as a basic mathematical approach to stochastic processes in discrete space and time; various of its applications, for example, the Monte Carlo method, are also illustrated. The theory of stochastic equations, as a representation of stochastic processes in continuous space–time, is discussed and used for obtaining a generalized, rigorous formulation of the Langevin and Fokker–Planck equations for generalized fluctuating observables. The Arrhenius formula as an example of the first exit-time problem is also derived.
This chapter provides a comprehensive overview of the foundational concepts essential for scalable Bayesian learning and Monte Carlo methods. It introduces Monte Carlo integration and its relevance to Bayesian statistics, focusing on techniques such as importance sampling and control variates. The chapter outlines key applications, including logistic regression, Bayesian matrix factorization, and Bayesian neural networks, which serve as illustrative examples throughout the book. It also offers a primer on Markov chains and stochastic differential equations, which are critical for understanding the advanced methods discussed in later chapters. Additionally, the chapter introduces kernel methods in preparation for their application in scalable Markov Chain Monte Carlo (MCMC) diagnostics.
We study a variant of the classical Markovian logistic SIS epidemic model on a complete graph, which has the additional feature that healthy individuals can become infected without contacting an infected member of the population. This additional ‘self-infection’ is used to model situations where there is an unknown source of infection or an external disease reservoir, such as an animal carrier population. In contrast to the classical logistic SIS epidemic model, the version with self-infection has a non-degenerate stationary distribution, and we derive precise asymptotics for the time to converge to stationarity (mixing time) as the population size becomes large. It turns out that the chain exhibits the cutoff phenomenon, which is a sharp transition in time from one to zero of the total variation distance to stationarity. We obtain the exact leading constant for the cutoff time and show that the window size is of constant (optimal) order. While this result is interesting in its own right, an additional contribution of this work is that the proof illustrates a recently formalised methodology of Barbour, Brightwell and Luczak (2022), ‘Long-term concentration of measure and cut-off’, Stochastic Processes and their Applications152, 378–423, which can be used to show cutoff via a combination of concentration-of-measure inequalities for the trajectory of the chain and coupling techniques.
This chapter contains topics related to matrices with special structures that arise in many applications. It discusses companion matrices that are a classic linear algebra topic. It constructs circulant matrices from a particular companion matrix and describes their signal processing applications. It discusses the closely related family of Toeplitz matrices. It describes the power iteration that is used later in the chapter for Markov chains. It discusses nonnegative matrices and their relationships to graphs, leading to the analysis of Markov chains. The chapter ends with two applications: Google’s PageRank method and spectral clustering using graph Laplacians.
It has become increasingly clear that economies can fruitfully be viewed as networks, consisting of millions of nodes (households, firms, banks, etc.) connected by business, social, and legal relationships. These relationships shape many outcomes that economists often measure. Over the past few years, research on production networks has flourished, as economists try to understand supply-side dynamics, default cascades, aggregate fluctuations, and many other phenomena. Economic Networks provides a brisk introduction to network analysis that is self-contained, rigorous, and illustrated with many figures, diagrams and listings with computer code. Network methods are put to work analyzing production networks, financial networks, and other related topics (including optimal transport, another highly active research field). Visualizations using recent data bring key ideas to life.
In this paper we extend results on reconstruction of probabilistic supports of independent and identically distributed random variables to supports of dependent stationary ${\mathbb R}^d$-valued random variables. All supports are assumed to be compact of positive reach in Euclidean space. Our main results involve the study of the convergence in the Hausdorff sense of a cloud of stationary dependent random vectors to their common support. A novel topological reconstruction result is stated, and a number of illustrative examples are presented. The example of the Möbius Markov chain on the circle is treated at the end with simulations.
In this paper, we introduce a slight variation of the dominated-coupling-from-the-past (DCFTP) algorithm of Kendall, for bounded Markov chains. It is based on the control of a (typically non-monotonic) stochastic recursion by another (typically monotonic) one. We show that this algorithm is particularly suitable for stochastic matching models with bounded patience, a class of models for which the steady-state distribution of the system is in general unknown in closed form. We first show that the Markov chain of this model can easily be controlled by an infinite-server queue. We then investigate the particular case where patience times are deterministic, and this control argument may fail. In that case we resort to an ad-hoc technique that can also be seen as a control (this time, by the arrival sequence). We then compare this algorithm to the primitive coupling-from-the-past (CFTP) algorithm and to control by an infinite-server queue, and show how our perfect simulation results can be used to estimate and compare, for instance, the loss probabilities of various systems in equilibrium.
Inaccuracy and information measures based on cumulative residual entropy are quite useful and have received considerable attention in many fields, such as statistics, probability, and reliability theory. In particular, many authors have studied cumulative residual inaccuracy between coherent systems based on system lifetimes. In a previous paper (Bueno and Balakrishnan, Prob. Eng. Inf. Sci.36, 2022), we discussed a cumulative residual inaccuracy measure for coherent systems at component level, that is, based on the common, stochastically dependent component lifetimes observed under a non-homogeneous Poisson process. In this paper, using a point process martingale approach, we extend this concept to a cumulative residual inaccuracy measure between non-explosive point processes and then specialize the results to Markov occurrence times. If the processes satisfy the proportional risk hazard process property, then the measure determines the Markov chain uniquely. Several examples are presented, including birth-and-death processes and pure birth process, and then the results are applied to coherent systems at component level subject to Markov failure and repair processes.
We propose a discrete-time discrete-space Markov chain approximation with a Brownian bridge correction for computing curvilinear boundary crossing probabilities of a general diffusion process on a finite time interval. For broad classes of curvilinear boundaries and diffusion processes, we prove the convergence of the constructed approximations in the form of products of the respective substochastic matrices to the boundary crossing probabilities for the process as the time grid used to construct the Markov chains is getting finer. Numerical results indicate that the convergence rate for the proposed approximation with the Brownian bridge correction is $O(n^{-2})$ in the case of $C^2$ boundaries and a uniform time grid with n steps.
An edge flipping is a non-reversible Markov chain on a given connected graph, as defined in Chung and Graham (2012). In the same paper, edge flipping eigenvalues and stationary distributions for some classes of graphs were identified. We further study edge flipping spectral properties to show a lower bound for the rate of convergence in the case of regular graphs. Moreover, we show by a coupling argument that a cutoff occurs at $\frac{1}{4} n \log n$ for the edge flipping on the complete graph.
Random walks on graphs are an essential primitive for many randomised algorithms and stochastic processes. It is natural to ask how much can be gained by running
$k$
multiple random walks independently and in parallel. Although the cover time of multiple walks has been investigated for many natural networks, the problem of finding a general characterisation of multiple cover times for worst-case start vertices (posed by Alon, Avin, Koucký, Kozma, Lotker and Tuttle in 2008) remains an open problem. First, we improve and tighten various bounds on the stationary cover time when
$k$
random walks start from vertices sampled from the stationary distribution. For example, we prove an unconditional lower bound of
$\Omega ((n/k) \log n)$
on the stationary cover time, holding for any
$n$
-vertex graph
$G$
and any
$1 \leq k =o(n\log n )$
. Secondly, we establish the stationary cover times of multiple walks on several fundamental networks up to constant factors. Thirdly, we present a framework characterising worst-case cover times in terms of stationary cover times and a novel, relaxed notion of mixing time for multiple walks called the partial mixing time. Roughly speaking, the partial mixing time only requires a specific portion of all random walks to be mixed. Using these new concepts, we can establish (or recover) the worst-case cover times for many networks including expanders, preferential attachment graphs, grids, binary trees and hypercubes.
Statistical models of processes where random events have an effect on partly random subsequent events are covered in this chapter. The sequence of eruptions of the geyser Old Faithful is taken as a simple example to illustrate Markov Chains. Infectious disease models are then covered and the history of various attempts at modelling them from the early twentieth century onwards is covered. Modelling religious conversion as a stochastic process is treated briefly.
We present a Markov chain on the n-dimensional hypercube
$\{0,1\}^n$
which satisfies
$t_{{\rm mix}}^{(n)}(\varepsilon) = n[1 + o(1)]$
. This Markov chain alternates between random and deterministic moves, and we prove that the chain has a cutoff with a window of size at most
$O(n^{0.5+\delta})$
, where
$\delta>0$
. The deterministic moves correspond to a linear shift register.
This book studies the large deviations for empirical measures and vector-valued additive functionals of Markov chains with general state space. Under suitable recurrence conditions, the ergodic theorem for additive functionals of a Markov chain asserts the almost sure convergence of the averages of a real or vector-valued function of the chain to the mean of the function with respect to the invariant distribution. In the case of empirical measures, the ergodic theorem states the almost sure convergence in a suitable sense to the invariant distribution. The large deviation theorems provide precise asymptotic estimates at logarithmic level of the probabilities of deviating from the preponderant behavior asserted by the ergodic theorems.
This study presents a micro-typological description of German dialects, focusing on the structure of 13,492 tokens of monosyllables, across 182 locations within Germany. Based on data from the Phonetischer Atlas der Bundesrepublik Deutschland, systematic geographical differences in both the segmental and prosodic organization of syllables are explored. The analysis reveals a North–South contrast in the organization of syllable structure. While the North tends toward more simple CVC syllables, the South tends toward the clustering of obstruents. An analysis of sonority dispersion reveals that in southern German, final demisyllables tend to follow more closely the sonority scale. Based on Markov chain models, the study reveals geographical differences in transition probabilities between the segments within monosyllables in German dialects.*
The generalized perturbative approach is an all-purpose variant of Stein’s method used to obtain rates of normal approximation. Originally developed for functions of independent random variables, this method is here extended to functions of the realization of a hidden Markov model. In this dependent setting, rates of convergence are provided in some applications, leading, in each instance, to an extra log-factor vis-à-vis the rate in the independent case.