1. Introduction
Urn models have been a popular research subject for many years due to their applicability in a wide range of fields, including computer science, economics, physics, business, genetics, and ecology. We refer interested readers to the entire textbook [Reference Johnson and Kotz15]; [Reference Kotz and Balakrishnan18] provides a wealth of information regarding urn models, and presents some significant developments. [Reference Johnson, Kotz and Balakrishnan16, Reference Johnson, Kemp and Kotz17] studied various urn models and their corresponding distributions with statistical inference problems [Reference Inoue and Aki13, Reference Mahmoud20].
A fundamental isomorphism between urn models with two types of balls and certain systems of ordinary differential equations was established in [Reference Flajolet, Dumas and Puyhaubert9]. This highly innovative method makes it possible to obtain probability distributions from the solutions of the differential systems. [Reference Flajolet, Gabarro and Pekari8] introduced the partial differential equation approach to urn models with two types of balls, and derived the probability generating functions as solutions to the partial differential equation.
Since the pioneering works of [Reference Flajolet, Gabarro and Pekari8, Reference Flajolet, Dumas and Puyhaubert9], the theory and applications of urn models based on analytical methods have received increased attention from probabilists, statisticians, and applied scientists [Reference Balaji and Mahmoud2, Reference Chen and Zhang6]. The urn models are recognized as exactly solvable urns, and the methodology is known as analytic combinatorics, which have been studied in various situations [Reference Flajolet and Sedgewick10]. A chapter of [Reference Mahmoud20] is devoted to analytic urns, and provides a comprehensive list of references. [Reference Hwang, Kuba and Panholzer11] considered several diminishing urns with two colors, and obtained the exact distributions by solving partial differential equations, while [Reference Morcrette and Mahmoud21] developed an analytical theory for the study of Pólya urns with random rules. Urn models under random replacement schemes have been studied by many researchers; [Reference Athreya and Karlin1, Reference Janson14, Reference Smythe22] obtained asymptotic results (see also references therein). We are concerned with exact probability distributions via the methods of analytic combinatorics.
In this study we treat urn models with $(m+1)$ types of balls under random replacement schemes, and develop the exact theory of analytic urns with their applications. The results presented here serve as a bridge between discrete-time urn processes and continuous analysis (in particular, the theory of differential equations). In Section 2, urn models containing $(m+1)$ types of balls under random replacement rules are explained. We also describe the history of urn evolution, and the enumerative approach to all histories using generating functions. In Section 3, the fundamental isomorphism between urn models with random addition rules and differential equations established in [Reference Morcrette and Mahmoud21] for urn models with two types of balls is generalized to urns with random addition rules and $(m+1)$ types of balls. In Section 4, we consider the joint distribution of the numbers of balls in the urn after n draws, and provide recurrence relations for the probability generating functions. Furthermore, we derive a partial differential equation satisfied by the generating function. Finally, in Section 5, we discuss several applications related to urn models with random addition rules for illustrative purposes. Two partition numbers used in Section 5 are examined further in the appendix.
2. Urn models
2.1. Motivating example
We provide a simple example that illustrates urn histories and their enumeration. From an urn initially containing two white balls and one blue ball, a ball is chosen at random, its color is observed, and the ball is returned to the urn with additional balls: if the color of the chosen ball is white, one white ball and two blue balls are added to the urn; if the color of the chosen ball is blue, two white balls and one blue ball are added to the urn. The trial then repeats. It is common to represent the replacement scheme with the addition matrix $A=\left(\begin{smallmatrix}1 & 2 \\[5pt] 2 & 1\end{smallmatrix}\right)$ . We denote the starting state by WWB. Note that there are initially two white balls. At the first drawing, one possible history of the urn evolution corresponds to the drawing of a white ball, in which case we return it to the urn together with one new white ball and two new blue balls. If a blue ball is chosen in the second drawing, it is returned to the urn together with two new white balls and one new blue ball. We have one possible history of urn evolution, described by the following sequences:
where the ball chosen in each step is underlined. Another urn evolution history is as follows:
We regard the two sequences as different histories, as although they share the same urn composition after two draws, they arrived at it through different stochastic paths. From this perspective, all histories are equally likely.
We will enumerate urn histories using exponential generating functions.
2.2. Urn scheme
Consider a Pólya urn model containing balls of $(m+1)$ different types of labels, which is characterized by a random replacement scheme. From an urn containing $b_i$ balls labeled i ( $i=0,1,\ldots,m$ ), a ball is drawn, its label is noted, and the ball is returned to the urn along with additional balls depending on the label of the drawn ball; if a ball labeled i ( $i=0,1,\ldots,m$ ) is drawn, $A_{ij}$ balls labeled j ( $j=0,1,\ldots,m$ ) are added, where $A_{ij}$ ( $i,j=0,1,\ldots,m$ ) are discrete random variables. This scheme is represented by the $(m+1) \times (m+1)$ replacement matrix A of discrete random entries as
The rows are indexed by the label of the drawn ball, and the columns are indexed by the label of the added ball. We assume that $A_{i0}+A_{i1}+ \cdots +A_{im}=\theta\ (\!\geq 0)$ , $A_{ii} \geq -1$ , and $A_{ij}\geq 0$ , $i \neq j$ , for $i,j=0,1,\ldots,m$ . The row sums of the replacement matrix are always constant with the value $\theta$ , which implies that a constant number of balls is added. The condition of steady linear growth of the urn size is said to be ‘balanced’, and the parameter $\theta$ is called the balance of the urn. The urn is said to be ‘tenable’ if we can perpetually sample from the urn according to the replacement scheme. Throughout this article, we limit ourselves to considering tenable balanced urns.
For $i=0,1,\ldots,m$ , let $B_n^{i}$ denote the number of balls labeled i after n draws from the urn. The total number of balls after n draws, denoted by $\tau_n=\sum_{i=0}^{m} B_n^{i}$ , is a deterministic quantity due to the balance condition. We then have $\tau_n = \tau_0 + n \theta = b_0 + b_1 + \cdots b_m + n \theta,$ where $\big(B_0^{0},B_0^{1},\ldots,B_0^{m}\big)=(b_0,b_1,\ldots,b_m)$ and $\tau_0=b_0+b_1 + \cdots + b_m$ . Assume that the joint distribution of $( A_{i0},A_{i1},\ldots,A_{im} )$ has the probability mass function $\pi_{a_{i0},a_{i1},\ldots,a_{im}}=\mathbb{P} ( A_{i0}=a_{i0},A_{i1}=a_{i1},\ldots,A_{im}=a_{im} )$ , $i=0,1,\ldots,m$ .
2.3. Enumerating histories
Let $h_n(i_0,i_1,\ldots,i_m)$ be the number of histories that corresponds to an urn with $\big(B_n^{0},B_n^{1},\ldots,B_n^{m}\big)=(i_0,i_1,\ldots,i_m)$ after n draws, and let
be the exponential generating function. The generating function of the total number of histories is expressed as $H(1,1,\ldots,1,z)$ .
Proposition 2.1. For a tenable balanced urn under scheme (2.1), the total number of possible histories after n draws is given by $\tau_n=\tau_0(\tau_0+\theta) \cdots (\tau_0+(n-1)\theta)$ , and the generating functions of the total number of histories can be expressed as
The joint probability mass function and the probability generating function of $\big(B_n^{0},\ldots,B_n^{m}\big)$ can be captured through the generating function $H(x_0,\ldots,x_m,z)$ as
respectively, where we use the notation $\big[x_1^{i_1} x_2^{i_2} \cdots x_k^{i_k}\big]g(x_1,x_2,\ldots,x_k)$ to extract the coefficient of $x_1^{i_1} x_2^{i_2} \cdots x_k^{i_k}$ in a multivariate generating function $g(x_1,x_2,\ldots,x_k)$ .
It is interesting to note that we can write the generating function $H(x_0,\ldots,x_m,z)$ through the joint probability mass function and probability generating function of $\big(B_n^{0},\ldots,B_n^{m}\big)$ :
We define the operator $\mathcal{D}$ by
which is suited for describing the urn histories.
Proposition 2.2. Under the initial urn condition $\big(B_0^{0},B_0^{1},\ldots,B_0^{m}\big)=(b_0,b_1,\ldots,b_m)$ and the replacement scheme as in (2.1), the generating function of the urn histories is expressed as
Proof. For $n \geq 0$ , we define the function $H_n ( x_0,x_1,\ldots,x_m )$ by
Then, the generating function $H(x_0,x_1,\ldots,x_m,z)$ of the urn histories can be rewritten as
The property of the operator $\mathcal{D}$ yields $H_{n+1}( x_0,x_1,\ldots,x_m) = \mathcal{D}(H_{n}( x_0,x_1,\ldots,x_m))$ . Hence, we have
3. Urn models and systems of differential equations
We study urn models with random replacement schemes by employing analytical methods. Based on an isomorphism theorem, we associated the urn model with a system of differential equations. More specifically, we obtain the following theorem.
Theorem 3.1. Under the initial urn condition $\big(B_0^{0},B_0^{1},\ldots,B_0^{m}\big)=(b_0,b_1,...,b_m)$ and the replacement scheme as in (2.1), the generating function of the urn histories is given by $H(u_0,u_1,...,u_m,z) = x_{0}^{b_0}(z)x_{1}^{b_{1}}(z) \cdots x_{m}^{b_{m}}(z)$ , where $u_0=x_0(0)$ , $u_1=x_1(0)$ , …, $u_m=x_m(0)$ such that $u_0 u_1 \cdots u_m \neq 0$ , and $(x_{0}(t),x_{1}(t),\ldots,x_{m}(t))$ is the solution to the differential system of equations
Proof. From the Taylor series expansion (around $t=0$ ) of the function $x_i^{b_i}(t+z)$ , $i=0,1,\ldots,m$ , we have
The product of the functions $x_0^{b_0}(t+z)x_1^{b_1}(t+z) \cdots x_m^{b_m}(t+z)$ is
We denote the operator ${\partial}/{\partial t}$ as $\tilde{\mathcal{D}}$ . Then, we have
By highlighting the relationship of the two expansions (2.4) and (3.2), if we let $\mathcal{D}=\tilde{\mathcal{D}}$ , or
we have $H ( x_0(t),x_1(t),\ldots,x_m(t),z) = x_0(t+z)^{b_0}x_1(t+z)^{b_1} \cdots x_m(t+z)^{b_m}$ , which is the same as letting
This is possible if
By setting $t=0$ , we get
We note that the factorial moment of $B_n^{i}$ , $i=0,1,e...,m$ , can be derived through the generating function $H(x_0,\ldots,x_m,z)$ . The next theorem provides the detail.
Theorem 3.2. The joint ( $r_0,\ldots,r_m$ ) descending factorial moment of $\big(B_n^{0},\ldots,B_n^{m}\big)$ is given by
or equivalently
where $\big(B_n^{i}\big)_{r_i} = B_n^{i} \big( B_n^{i}-1 \big) \cdots \big( B_n^{i}-r_i + 1 \big)$ , $i=0,1,\ldots,m$ .
Proof. In terms of (2.3), we have
It is easy to see that the two results are equivalent.
4. Urn models and partial differential equations
4.1. Generating functions
Consider the joint distribution of $\big(B_n^{1},B_n^{2},\ldots,B_n^{m}\big)$ . The probability generating function of $\big(B_n^{1},B_n^{2},\ldots,B_n^{m}\big)$ is denoted by $\phi_n(\boldsymbol{{t}})$ ; that is,
where $\boldsymbol{{t}} = (t_1,\ldots,t_m)$ .
We define the generating function as $\overline{H}(t_1,\ldots,t_m,z) = H(t_0,t_1,\ldots,t_m,z) \big|_{t_0=1}$ . Immediately, we can observe the following relation:
In terms of (2.2) and (2.3), the joint probability mass function and probability generating function of $\big(B_n^{1},B_n^{2},\ldots,B_n^{m}\big)$ can be expressed as
respectively.
It is worth mentioning that the generating function $\overline{H}(t_1,\ldots,t_m,z)$ can be captured through the joint probability mass function and the probability generating function of $\big(B_n^{1},\ldots,B_n^{m}\big)$ :
4.2. Recurrence relations for probability generating functions
We can evaluate the joint probability mass function of $\big(B_n^{1},\ldots,B_n^{m}\big)$ through recurrence relations that can be established easily using probability generating functions.
Theorem 4.1. Under the initial urn condition $\big(B_0^{0},B_0^{1},\ldots,B_0^{m}\big) = (b_0,b_1,\ldots,b_m)$ and the replacement scheme in (2.1), the probability generating functions $\phi_{n}(\boldsymbol{{t}})$ for $n \geq 0$ satisfy the following recurrence relations:
where $\alpha_{i}(\boldsymbol{{t}})=\mathbb{E}\left[t_1^{A_{i1}} t_2^{A_{i2}} \cdots t_m^{A_{im}} \right]$ .
Proof. (4.1) is an immediate consequence of the following conditional expectation:
The initial condition (4.2) is easily obtained by the definition of the probability generating function.
As a special case of $m=1$ , consider the Pólya urn model containing two different labels: 0 and 1. If $\alpha_0(t_1) \neq \alpha_1(t_1)$ and we can find the antiderivative
we provide the simpler recurrence. The transformation
yields
where
By differentiating (4.1) and (4.2), and making use of
we can readily establish the recurrence relations for the expected values of $B_n^{i}$ , $n \geq 0$ , $i=1,\ldots,m$ .
Corollary 4.1. For $i=1,2,\ldots,m$ , the expected values $\mathbb{E} \big[ B_n^{i} \big]$ , $n \geq 0$ satisfy the recurrence relations
In the case of $m=1$ , [Reference Hwang, Chern and Duh12] studied recurrence relations with a slight modification of (4.1) and (4.2), known as recurrences of Eulerian type, and considered the applications to the Pólya urn models.
4.3. Partial differential equations
We find the generating function $\overline{H}(\boldsymbol{{t}},z)$ by employing the partial differential equation approach. We derive the partial differential equation satisfied by the generating function $\overline{H}(\boldsymbol{{t}},z)$ .
Theorem 4.2. Under the initial urn condition $\left(B_0^{0},B_0^{1},\ldots,B_0^{m}\right) = (b_0,b_1,\ldots,b_m)$ and the replacement scheme in (2.1), the generating function $\overline{H}(\boldsymbol{{t}},z)$ satisfies the partial differential equation
where $\alpha_{i}(\boldsymbol{{t}}) = \mathbb{E}\big[t_1^{A_{i1}} t_2^{A_{i2}} \cdots t_m^{A_{im}} \big]$ .
Proof. From (4.1), we have
As a special case of $m=1$ , we examine the Pólya urn model containing the labels 0 and 1. If $\alpha_0(t_1) \neq \alpha_1(t_1)$ and we can obtain the antiderivative
we provide the simpler partial differential equation. When
(4.3) and (4.4) yield the following partial differential equation:
where
First-order partial differential equations can often be solved by the method of characteristics. The partial differential equation in (4.7) can be written as
Note that
We then see that $\eta(t_1,z)=C$ , where
Therefore, the generating function $\Psi(t_1,z)$ can be expressed as $\Psi(t_1,z) = G (\eta(t_1,z))$ , where the function G is specified by the initial condition (4.8) as $G(\eta(t_1,0)) = t_1^{b_1} I(t_1, \tau_0 )$ .
5. Applications
5.1. Coupon collector’s urn
Consider the coupon collector’s problem with a loss. Suppose that there are k distinct types of coupons bearing the numbers 1, 2, …, k, and that a coupon is placed in each package. Assume that a coupon of type i is collected with probability $1/k$ , $i=1,2,\ldots,k$ . A customer buys a product daily; they can retain the coupon with probability p, or may lose it with probability $1-p$ ( $0 < p \leq 1$ ). We are interested in the probability that they collect every type of coupon m or more times after n days. When $p=1$ , the problem corresponds to the classical coupon collector’s problem.
We can formulate this problem with the adequate urn model. Suppose that initially there are k balls all labeled (0) in the urn. When a ball labeled (0) is picked up, its label changes to (1) with probability p or remains the same with probability $1-p$ . When a ball labeled (1) is picked up, its label changes to (2) with probability p or remains (1) with probability $1-p$ , and so on. If a ball labeled ( $m-1$ ) becomes a ball labeled (m), it remains so forever. This experiment is repeated n times.
The replacement matrix is expressed as
where B(p) denotes the Bernoulli random variable taking value 1 with probability p and 0 with probability $1-p$ ( $0 < p \leq 1$ ). The system of differential equations is
with $x_0(0)=u_0$ , $x_1(0)=u_1$ , …, $x_m(0)=u_m$ . Then, the solution is
Under the initial urn condition $( B_0^{0}, B_0^{1},\ldots, B_0^{m}) = ( k,0,\ldots,0)$ , the urn history generating function is given by
The probability mass function of $B_n^{m}$ can be expressed in terms of the Stirling numbers.
For a real number a, a non-negative integer k, and a positive integer s, the numbers $S_{a,s}(n,k)$ , $n=sk,sk+1,\ldots$ , are defined by their exponential generating function
These numbers are known as the generalized non-central Stirling numbers of the second kind [Reference Koutras19].
Observing that
we have the following result under (5.3).
Proposition 5.1. The probability that a customer has all the types of coupons at least m or more times after n days is
In the case of $m=2$ , [Reference Morcrette and Mahmoud21] considered the urn for a two-type coupon collection. Note that it also mentioned a possible extension to an urn model with m colors, which stimulated our interest in analytic urns. Furthermore, it discussed the use of asymptotic tools to obtain limit laws.
5.2. Coupon collector waiting time problem
Under equivalent conditions to those presented in Section 5.1, we treat the waiting time problem in coupon collection with a loss in the special case of $m=1$ . A customer buys a product daily; they can retain the coupon with probability p, or lose it with probability $1-p$ ( $0 < p \leq 1$ ). How many days does it take to collect every type of coupon? Let that time be denoted by $T_k$ .
From the obvious identity $\mathbb{P}( T_k \leq n ) = \mathbb{P}(B_n^{1}=k)$ , $n \geq 1$ , we can express the probability mass function of the waiting time $T_k$ as
Proposition 5.2. The probability mass function of the waiting time $T_k$ is given by
By making use of (A.1) in the appendix, we can proceed to derive the probability generating function of the waiting time $T_k$ as
This representation implies that $T_k$ can be decomposed as a sum of k independent random variables.
Proposition 5.3. For $i=0,1,\ldots,k-1$ , let $Y_i$ be independent random variables with probability mass functions
Then, the waiting time $T_k$ can be decomposed as $T_k = Y_0+Y_1 + \cdots + Y_{k-1}$ .
The expected value and variance of $T_k$ are expressed, respectively, as
Remark 5.1. From (5.4) and (5.5), we have
in probability.
Consider a numerical example in the case of $n=365$ and $p=0.8$ . That is, we are interested in the number of people we should meet until we have seen at least one person with every possible birthday. In this case, the limit theorem yields approximately $({365 \log 365})/{0.8}=2691.83$ tries to get a complete set, which means that the number of trials is $7.37$ times the number of birthdays. We mention that [Reference Durrett7] considered the numerical example in the case of $n=365$ and $p=1$ .
5.3. Birthday urn
The birthday problem with recording failures can be approached with the same matrix as in (5.1). Assume that each person is equally likely to have any of the 365 days in a year as their birthday. Suppose that we randomly interview people individually. When we ask the person for their birthday, we can record it with probability p, or fail to record it with probability $1-p$ ( $0 < p \leq 1$ ). We need to determine the probability that, after interviews with n persons chosen at random, at least m persons have the same birthday on record.
The probability mass function $\mathbb{P}(B_n^{m}=0)$ can be evaluated effectively. For a real number a, a non-negative integer k, and a positive integer s, the numbers $M_{a,s}(n,k)$ , $n=0,1,\ldots$ , are defined by their exponential generating function
From (5.2) with $k=365$ , we have
From (5.6),
Furthermore, the following result is established.
Proposition 5.4. The probability that among n persons chosen at random, at least m persons have the same birthday, is given by
5.4. Waiting time in birthday problem
Consider the waiting time birthday problem with failures. Assume that each person is equally likely to have any of the 365 days in the year as their birthday. Suppose that we randomly interview people individually. When we ask each person for their birthday, we can record it with probability p or fail to record it with probability $1-p$ ( $0 < p \leq 1$ ). Let $W_m$ be the waiting time until we find m people with a common birthday on record. We investigate the distribution of the waiting time $W_m$ in terms of $B_n^{m}$ .
It is clear that $\mathbb{P}( W_m \leq n ) = \mathbb{P}(B_n^{m}=0)$ . The distribution of $W_m$ can be evaluated using (5.7).
Proposition 5.5. The probability mass function of the waiting time $W_m$ is given by
5.5. Ehrenfest urn on graphs
Consider five urns labeled 0, 1, 2, 3, and 4, initially containing k balls in total. At any given time, a ball is drawn from one urn with probability ${1}/{k}$ and moved to the adjacent urn with given probabilities (see Figure 1). For $i=0,1,2,3,4$ , let $B_n^{i}$ be the number of balls in the i-th urn.
The associated system of differential equations is
with $x_0(0)=u_0$ , $x_1(0)=u_1$ , $x_2(0)=u_2$ , $x_3(0)=u_3$ , and $x_4(0)=u_4$ .
For example, the solution $x_0(t)$ is given by
Under the initial urn condition $\left( B_0^{0}, B_0^{1}, B_0^{2}, B_0^{3}, B_0^{4} \right) = ( k,0,0,0,0 )$ , the expected values are
Therefore, we have $\mathbb{E}[B_n^{0}] \to {k}/{5}$ , $\mathbb{E}[B_n^{1}] = \mathbb{E}[B_n^{2}] \to {3k}/{10}$ , and $\mathbb{E}[B_n^{3}] = \mathbb{E}[B_n^{4}] \to {k}/{10}$ as $n \to \infty$ . When $k=20$ and $n=30$ , Figure 2 shows the graphs of the joint probability mass function $\left(B_{30}^{1}, B_{30}^{3}\right)$ .
5.6. Pólya–Ehrenfest urn
Consider two urns labeled 0 and 1. Initially, urn 0 contains $b_0$ balls, and urn 1 contains $b_1$ balls. At any given time, a ball in the urns is drawn and moved to the other urn with equal probabilities. Therefore, the total number of balls constantly increases by 1. This urn model was referred to as the Pólya–Ehrenfest model in [Reference Blom, Holst and Sandell3]. The replacement matrix is expressed as $\left(\begin{smallmatrix}-1 & \hphantom{-}2 \\[5pt] \hphantom{-}2 & -1\end{smallmatrix}\right)$ .
We consider the urn scheme alternating with equal probabilities between $\left(\begin{smallmatrix}-1 & \hphantom{-}2 \\[5pt] \hphantom{-}2 & -1\end{smallmatrix}\right)$ and $\big(\begin{smallmatrix}1 & 0 \\[5pt] 0 & 1\end{smallmatrix}\big)$ , i.e. between a Pólya–Ehrenfest urn with probability $\dfrac{1}{2}$ and a Pólya urn with probability $\dfrac{1}{2}$ .
From the partial differential equations in (4.5) and (4.6) under the initial urn condition $(B_0^{0}, B_0^{1})=(b_0,b_1)$ with $\alpha_{0}(t_1) = {t_1^2}/{2} +\dfrac{1}{2}$ and $\alpha_{1}(t_1) = {1}/{2t_1} + {t_1}/{2}$ , we have
The solution is given by
From the results presented in Corollary 4.1, the expected value of $B_{n}^{1}$ is
In the case where $b_0=1$ , $b_1=1$ , and $n=50$ , we present graphs of the distribution of the number of balls in urn 1 in Figure 3.
5.7. Multinomial urn
Assume that the random variables $(A_{i0},A_{i1},\ldots,A_{im})$ have the probability mass function
with $p_{0} + p_{1}+ \cdots +p_{m}=1$ , $i=0,1,\ldots,m$ . Then, we have $\alpha_{i}(\boldsymbol{{t}}) = ( p_{0}+p_{1} t_1 + p_{2} t_2 +\cdots + p_{m} t_m )^{\theta}$ , $i=0,1,\ldots,m$ .
From the partial differential equation in (4.5), with (4.6) under the initial urn condition $(B_n^{0},B_n^{1},\ldots,B_n^{m}) = (b_0,b_1,\ldots,b_m)$ , the solution is
Proposition 5.6. Under the initial urn condition $(B_n^{0},B_n^{1},\ldots,B_n^{m}) = (b_0,b_1,\ldots,b_m)$ , the probability generating function and probability mass function of ( $B_n^{1},B_n^{2},\ldots,B_n^{m}$ ) are given by
5.8. Uniform urn
Assume that the random variables $(A_{i1},A_{i1},\ldots,A_{im})$ have the probability mass function $\mathbb{P}(A_{i1}=a_{i1},A_{i2}=a_{i2},\ldots,A_{im}=a_{im}) = ( \ell+1 )^{-m}$ , with $A_{i0}=\theta-\sum_{j=1}^{m} A_{ij}$ , where $\theta=m \ell$ . Then, we have
From the partial differential equation in (4.5), with (4.6) under the initial urn condition $(B_n^{0},B_n^{1},\ldots,B_n^{m}) = (b_0,b_1,\ldots,b_m)$ , we obtain the solution $\overline{H}(\boldsymbol{{t}},z)$ as
Proposition 5.7. Under the initial urn condition $(B_n^{0},B_n^{1},\ldots,B_n^{m}) = (b_0,b_1,\ldots,b_m)$ , the probability generating function and probability mass function of ( $B_n^{1},B_n^{2},\ldots,B_n^{m}$ ) are given by
where $T_{\ell, n,k} = [x^k](1+t+t^2+\cdots +t^\ell)^n$ .
We note that the numbers $T_{\ell, n,k}$ are well known classically and have many combinatorial aspects [Reference Flajolet and Sedgewick10].
Appendix A. The partition numbers
In this section we examine two partition numbers and derive the recurrence relations [Reference Charalambides4, Reference Charalambides5].
For a real number a, a non-negative integer k, and a positive integer s, the generalized non-central Stirling numbers of the second kind $S_{a,s}(n,k)$ , $n=sk,sk+1,\ldots$ , are defined by their exponential generating function given by (5.3).
Proposition A.1. The numbers $S_{a,s}(n,k)$ satisfy the recurrence relations
Proof. By differentiating the generating function $f_{a,s}(t,k)$ with respect to t, we have
By equating the coefficients of ${t^n}/{n!}$ on both sides of the above expression, we derive the recurrence relations.
For $s=1$ , we define the generating function of the numbers $S_{a,1}(n,k)$ by $\varphi_{a,1}(u)= \sum_{n=k}^{\infty} S_{a,1}(n,k) u^n$ .
Proposition A.2. The generating function of the numbers $S_{a,1}(n,k)$ is given by
Proof. For $s=1$ , we have $S_{a,1}(n,k) = (k+a)S_{a,1}(n-1,k) + S_{a,1}(n-1,k-1)$ , $n \geq k$ . By multiplying the above expression by $u^n$ and summing for $n \geq k$ under the initial conditions, we get $\varphi_{a,1}(u,k) = u(k+a)\varphi_{a,1}(u,k)+u\varphi_{a,1}(u,k-1)$ and
By applying this recurrence relation repeatedly with $\varphi_{a,1}(u,0)=1$ , we obtain (A.1).
For a real number a, a non-negative integer k, and a positive integer s, the numbers $M_{a,s}(n,k)$ , $n=0,1,\ldots$ , are defined by their exponential generating function given by (5.6).
Proposition A.3. The numbers $M_{a,s}(n,k)$ satisfy the recurrence relations
Proof. As in the proof of Proposition A.1, by differentiating the generating function $g_{a,s}(t,k)$ with respect to t, we have
By equating the coefficients of ${t^n}/{n!}$ on both sides of the above expression, we derive the recurrence relations.
Acknowledgements
The author wishes to thank the editor and referees for the careful review of this paper and the helpful suggestions which led to improved results.
Funding information
There are no funding bodies to thank relating to the creation of this article.
Competing interests
There were no competing interests to declare which arose during the preparation or publication process of this article.