We use cookies to distinguish you from other users and to provide you with a better experience on our websites. Close this message to accept cookies or find out how to manage your cookie settings.
To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Students are introduced the logic, foundation, and basics of statistical inference. The need for samples is first discussed and then how samples can be used to make inferences about the larger population. The normal distribution is then discussed, along with Z-scores to illustrate basic probability and the logic of statistical significance.
This chapter covers the two topics of descriptive statistics and the normal distribution. We first discuss the role of descriptive statistics and the measures of central tendency, variance, and standard deviation. We also provide examples of the kinds of graphs often used in descriptive statistics. We next discuss the normal distribution, its properties and its role in descriptive and inferential statistical analysis.
This chapter reviews some essential concepts of probability and statistics, including: line plots, histograms, scatter plots, mean, median, quantiles, variance, random variables, probability density function, expectation of a random variable, covariance and correlation, independence the normal distribution (also known as the Gaussian distribution), the chi-square distribution. The above concepts provide the foundation for the statistical methods discussed in the rest of this book.
Let
$k\geqslant 1$
be a natural number and
$\omega _k(n)$
denote the number of distinct prime factors of a natural number n with multiplicity k. We estimate the first and second moments of the functions
$\omega _k$
with
$k\geqslant 1$
. Moreover, we prove that the function
$\omega _1(n)$
has normal order
$\log \log n$
and the function
$(\omega _1(n)-\log \log n)/\sqrt {\log \log n}$
has a normal distribution. Finally, we prove that the functions
$\omega _k(n)$
with
$k\geqslant 2$
do not have normal order
$F(n)$
for any nondecreasing nonnegative function F.
No matter how much care is taken during an experiment, or how sophisticated the equipment used, values obtained through measurement are influenced by errors. Errors can be thought of as acting to conceal the true value of the quantity sought through experiment. Random errors cause values obtained through measurement to occur above and below the true value. This chapter considers statistically-based methods for dealing with variability in experimental data such as that caused by random errors. As statistics can be described as the science of assembling, organising and interpreting numerical data, it is an ideal tool for assisting in the analysis of experimental data.
Log-concave random variables and their various properties play an increasingly important role in probability, statistics, and other fields. For a distribution F, denote by 𝒟F the set of distributions G such that the convolution of F and G has a log-concave probability mass function or probability density function. In this paper, we investigate sufficient and necessary conditions under which 𝒟F ⊆ 𝒟G, where F and G belong to a parametric family of distributions. Both discrete and continuous settings are considered.
Numerous papers have investigated the distribution of birth weight. This interest arises from the association between birth weight and the future health condition of the child. Birth weight distribution commonly differs slightly from the Gaussian distribution. The distribution is typically split into two components: a predominant Gaussian distribution and an unspecified ‘residual’ distribution. In this study, we consider birth weight data from the Åland Islands (Finland) for the period 1885–1998. We compare birth weight between males and females and among singletons and twins. Our study confirms that, on average, birth weight was highest among singletons, medium among twins, and lowest among triplets. A marked difference in the mean birth weight between singleton males and females was found. For singletons, the distribution of birth weight differed significantly from the normal distribution, but for twins the normal distribution held.
We propose a novel approach to introducing hypothesis testing into the biologycurriculum. Instead of telling students the hypothesis and what kind of data to collectfollowed by a rigid recipe of testing the hypothesis with a given test statistic, we askstudents to develop a hypothesis and a mathematical model that describes the nullhypothesis. Simulation of the model under the null hypothesis allows students to comparetheir experimental data to what they would expect under the null hypothesis, thus leadingto a much more intuitive understanding of hypothesis testing. This approach has beentested both in the classroom and in faculty workshops, and we provide some suggestions forimplementations based on our experiences.
This paper presents new Gaussian approximations for the cumulative distribution function P(Aλ ≤ s) of a Poisson random variable Aλ with mean λ. Using an integral transformation, we first bring the Poisson distribution into quasi-Gaussian form, which permits evaluation in terms of the normal distribution function Φ. The quasi-Gaussian form contains an implicitly defined function y, which is closely related to the Lambert W-function. A detailed analysis of y leads to a powerful asymptotic expansion and sharp bounds on P(Aλ ≤ s). The results for P(Aλ ≤ s) differ from most classical results related to the central limit theorem in that the leading term Φ(β), with is replaced by Φ(α), where α is a simple function of s that converges to β as s tends to ∞. Changing β into α turns out to increase precision for small and moderately large values of s. The results for P(Aλ ≤ s) lead to similar results related to the Erlang B formula. The asymptotic expansion for Erlang's B is shown to give rise to accurate approximations; the obtained bounds seem to be the sharpest in the literature thus far.
The target measure μ is the distribution of a random vector in a box ℬ, a Cartesian product of bounded intervals. The Gibbs sampler is a Markov chain with invariant measure μ. A ‘coupling from the past’ construction of the Gibbs sampler is used to show ergodicity of the dynamics and to perfectly simulate μ. An algorithm to sample vectors with multinormal distribution truncated to ℬ is then implemented.
The convex hull of n independent random points in ℝd, chosen according to the normal distribution, is called a Gaussian polytope. Estimates for the variance of the number of i-faces and for the variance of the ith intrinsic volume of a Gaussian polytope in ℝd, d∈ℕ, are established by means of the Efron-Stein jackknife inequality and a new formula of Blaschke-Petkantschin type. These estimates imply laws of large numbers for the number of i-faces and for the ith intrinsic volume of a Gaussian polytope as n→∞.
Let F be a probability distribution function with density f. We assume that (a) F has finite moments of any integer positive order and (b) the classical problem of moments for F has a nonunique solution (F is M-indeterminate). Our goal is to describe a , where h is a ‘small' perturbation function. Such a class S consists of different distributions Fε (fε is the density of Fε) all sharing the same moments as those of F, thus illustrating the nonuniqueness of F, and of any Fε, in terms of the moments. Power transformations of distributions such as the normal, log-normal and exponential are considered and for them Stieltjes classes written explicitly. We define a characteristic of S called an index of dissimilarity and calculate its value in some cases. A new Stieltjes class involving a power of the normal distribution is presented. An open question about the inverse Gaussian distribution is formulated. Related topics are briefly discussed.
The differences in facial anatomical structures of the major ethnic groups,may also be reflected in nasal resistance. Active anterior rhinomanometry (AAR) is the recommended technique for the objective assessment of nasal airway resistance (NAR). This study comprised of 85 adult Malay subjects. All the subjects had to undergo a primary assessment of relevantsymptoms of nasal disease and nasal examination before undergoing AAR assessment. The mean value of total nasal airway resistance (NAR) was 0.19 Pa/cm3/s (ranged from 0.09 to 0.55Pa/cm3/s) at 75 Pa pressure point and 0.24 Pa/cm3/s ranged from 0.12 to 0.52 Pa/cm3/s) at 150 Pa pressure point. The mean unilateral NARwas 0.46 Pa/cm3/s at a reference pressure of 75 Pa and 0.51 Pa/cm3/s at a reference pressure of 150 Pa. In this study we presented normal values for NAR in healthy Malay adult subjects. AARproves to be a valuable clinical method for recording and quantitating nasal resistance.
In this paper non-normal distributions via scale mixtures are introduced into insurance applications. The symmetric distributions of interest are the Student-t and exponential power (EP) distributions. A Bayesian approach is adopted with the aid of simulation to obtain posterior summaries. We shall show that the computational burden for the Bayesian calculations is alleviated via the scale mixtures representations. Illustrative examples are given.
The class of stopping rules for a sequence of i.i.d. random variables with partially known distribution is restricted by requiring invariance with respect to certain transformations. Invariant stopping rules have an intuitive appeal when the optimal stopping problem is invariant with respect to the actual gain function. Uniformly best invariant stopping rules are derived for the gamma distribution with known shape parameter and unknown scale parameter, for the uniform distribution with both endpoints unknown, and for the normal distribution with unknown mean and variance. Some comparisons with previously published results are made.
We consider the Darvey, Ninham and Staff model for reversible chemical reactions, in the case where the ratio of the rate constants is either very large or very small. It is shown that the distribution of the number of molecules at equilibrium may sometimes be closely approximated by the Poisson distribution, and on other occasions, by a distribution with much smaller tails than the Poisson. The second type of approximating distribution is termed a Bessel distribution, and its properties are studied.
Let {W(t), 0 ≦ t < ∞} be the standard Wiener process. The probabilities of the type P[sup0≦t ≦ TW(t) − f(t) ≧ 0] have been extensively studied when f(t) is a deterministic function. This paper discusses the probabilities of the type P{sup0≦t ≦ TW(t) − [f(t) + X(t)] ≧ 0} when X(t) is a stochastic process. By taking compound Poisson processes as X(t), the paper gives procedures for finding such probabilities.
Let Yn denote the largest of n independent N(0,1) random variables. It is shown that the error in approximating the distribution of Yn by the type III extreme value distribution exp {– (–Ax + B)k}, k > 0, is uniformly of order (log n)–2 if and only if the constants A, B and k satisfy certain conditions. In particular, this holds for the penultimate form of Fisher and Tippett (1928). Furthermore, two sufficient conditions are given so that these results can be extended to a stationary Gaussian sequence.
Consider an array of binary random variables distributed over an m1(n) by m2(n) rectangular lattice and let Y1(n) denote the number of pairs of variables d, units apart and both equal to 1. We show that if the binary variables are independent and identically distributed, then under certain conditions Y(n) = (Y1(n), · ··, Yr(n)) is asymptotically multivariate normal for n large and r finite. This result is extended to versions of a model which provide clustering (repulsion) alternatives to randomness and have clustering (repulsion) parameter values nearly equal to 0. Statistical applications of these results are discussed.