Publication Cover
Sequential Analysis
Design Methods and Applications
Volume 27, 2008 - Issue 1
1,518
Views
9
CrossRef citations to date
0
Altmetric
Original Articles

Bayesian Detection of Changes of a Poisson Process Monitored at Discrete Time Points Where the Arrival Rates are Unknown

Pages 68-77 | Received 05 Jul 2005, Accepted 06 Nov 2007, Published online: 04 Feb 2008

Abstract

We look at a Poisson process where the arrival rate changes at some unknown integer. At each integer, we count the number of arrivals that happened in that time interval. We assume that the arrival rates before and after the change are unknown. For a loss function consisting of the cost of late detection and a penalty for early stopping, we develop, using dynamic programming, the one- and two-step look-ahead Bayesian stopping rules. We provide some numerical results to illustrate the effectiveness of the detection procedures.

Subject Classifications:

1. INTRODUCTION

Detection of changes in the distribution of random variables has become very important in many aspects of life today. When there is an increase in the arrival rates of patients coming to a hospital, it is important to detect this change as soon as possible. This could be due to environmental factors or other issues. This is also important in industry, where quality control depends upon being able to detect changes in the process mean as soon as possible.

There are many published papers dealing with the topic of detection of changes in the distribution of data. Shewhan's (Citation1926) control charts for means and standard deviations, Page's (Citation1954) continuous inspection schemes, and Page's (Citation1962) cumulative sum schemes are all process control procedures designed to detect changes in the distributions of sequences of observed data. Chernoff and Zacks (Citation1964), Hinkley and Hinkley (Citation1970), Kander and Zacks (Citation1966), and Smith (Citation1975) investigated the Bayesian detection procedures for fixed a sample sizes. See Zacks (Citation1983) for a survey of papers on change-point problems. Shiryaev (Citation1978) studied optimal Bayesian procedures for detecting changes when the process is stopped as soon as a change is detected. Zacks and Barzily (Citation1981) extended the work of Shiryaev for the case of Bernoulli sequences. Optimal stopping rules were described based on dynamic programming procedures. In general, finding optimal stopping rules is quite difficult. Zacks and Barzily gave explicit formulas for finding one- and two-step-ahead suboptimal procedures.

Several studies were published recently on detecting changes in the intensity of a homogeneous ordinary Poisson process. Among these studies, we mention Peskir and Shiryaev (Citation2002), Herberts and Jensen (Citation2004), and Brown and Zacks (Citation2006a). These papers dealt with a Poisson process that is monitored continuously.

Brown and Zacks (Citation2006b) also studied a Poisson process that is monitored only at discrete time points. In that paper, we look at the sequence of random variables X i , where X i is the number of arrivals that occur in the time interval (i − 1, i]. Thus, we see only the number of arrivals that happen in each time interval, not exactly where the arrivals occurred within that time interval. In all of the above papers, it was assumed that the arrival rates both before and after the change are known. In the present paper, we assume that the arrival rates before and after the change are unknown and the change is positive. We assume that, the change point τ is an unknown integer. We use a Bayesian approach. We put Shiryaev's (Citation1978) geometric prior distribution on the change point τ and a uniform 0 < λ1 < λ2 < M prior distribution on the two arrival rates λ1 and λ2.

In Section 2, we discuss the posterior process of calculating the probability that a change has already occurred by time n. In Section 3, we discuss the optimal stopping rule based on dynamic programming procedures. We assume there is a cost associated with stopping early and a cost associated with late detection. We develop optimal stopping rules that will minimize the expected cost. In Section 4, we give an explicit formulation of the two-step-ahead stopping rule. We conclude with a numerical example showing how our method works.

2. THE BAYESIAN FRAMEWORK

We have a Poisson process, but we are only able to observe this process at fixed discrete time points. Let X i be the number of arrivals in the interval (i − 1, i]. We denote ℑ n as the σ-field generated by {X 1,…, X n }. At the end of some unknown interval indexed by an integer τ, the rate of arrivals increases from λ1 to λ2. We assume the arrival rates before and after the change, namely, λ1 and λ2, are unknown, as is the change point τ. Our goal is to find a detection procedure that will detect the change point τ as soon as possible. The event {τ = 0} represents the case where the change happens prior to the first observed interval. In this case all the observations are taken from a process with the arrival rate at λ2. The event {τ = r} means that the change happens at the end of the rth interval. Define as the total number of arrivals before time r and as the total number of arrivals after time r. T 0 ≡ 0 and T 0∗ = 0. Thus, when τ = r, T r is a Poisson (rλ1) and T nr is Poisson ((n − r2). If (λ1, λ2) are both known, the likelihood of τ is given by

When (λ1, λ2) are unknown, the profile likelihood function of τ is the expected value of (Equation2.1) with respect to the prior distribution of (λ1, λ2). We assume a priori that (λ1, λ2) is uniformly distributed in some interval 0 < λ1 < λ2 < M. That is, we know that when the change happens, the arrival rates have increased, and we also know that the arrival rates are bounded and uniform in a certain interval. Thus the profile likelihood function of τ, given ℱ n , is

For r = 0, let
where is the survival function of the Poisson distribution with parameter λ. For 1 ≤ r ≤ n − 1, let
Finally, let
We put Shiryaev's geometric prior distribution h on the change point τ,
Thus, we obtain that the posterior probability that there has been a change by time n is
Let  = 1 − ψ n be the probability that there has not been a change by time n.

3. OPTIMAL STOPPING RULE

In this section, we find an optimal stopping rule. Suppose the cost of stopping early is c 1 loss units and the cost per time unit of stopping late is c 2 loss units. We want to find a stopping rule that will minimize the expected cost. We assume without loss of generality that c 1 = 1 and c 2 = c is the relative cost of stopping late. Let R n be the risk (or minimal expected loss) at time n. Therefore,

The stopping rule, then, would be to stop sampling when the risk of stopping is smaller than the risk of continuing. According to (Equation3.1), this rule is equivalent to saying we stop sampling at the first n such that

In order to calculate R n , we use the principle of dynamic programming. Suppose we must stop after n∗ observations if we haven't stopped before. Let be the risk at time n when at most j more observations are allowed. When j = 0,n∗ = n. Therefore at time n we must stop and our risk is

In general, for j ≥ 1, we have a choice: we can stop and say that a change has occurred, or we can continue and take another observation. If we decide to continue, our risk is . Therefore, our risk at time n is . We can further write this as

Theorem 3.1

The sequence {(,ℑ n ), n ≥ 1} is a super martingale.

Proof

Taking the conditional expectation of ψ n+1 given the σ-field generated by the first n observations, we obtain

Therefore,
Accordingly, the sequence {,n ≥ 0} is a super martingale.

Corollary 3.1

Proof

From equation (Equation3.6),

Thus, the lim sup n→∞ E() = 0. Since  ≥ 0 with probability 1 for all n, we obtain (Equation3.7).

Thus, the one-step look-ahead procedure states that we stop sampling and declare that a change has happened at the first n such that

Define for j ≥ 1, and . Thus, the functions satisfy the recursive relationship
Hence, the risk can be written as a function of , namely,
We then make the decision to stop when , which is equivalent to stopping when . We can define the jth-step-ahead boundary to be .

Lemma 3.1

exists. That is, there exists an optimal stopping boundary.

Proof

By induction one can show that are decreasing in j. Thus, are decreasing functions of j. Also, for all j. By the monotone convergence theorem, the limit exists.

Lemma 3.2

.

Proof

Let

Therefore, if i ∊ I, then  < q∗. Since by Corollary 2.1,  → 0,
Therefore, . Thus, . By induction, we can show this for all j.

4. EXPLICIT FORMULATION OF THE TWO-STEP-AHEAD BOUNDARY

In this section, we evaluate the two-step-ahead boundary. Substituting j = 2 into equation (Equation3.10), we obtain

To evaluate the two-step-ahead risk, we find an expression for . First, we express recursively as a function of :
where, according to (Equation2.5),
and is defined as L r when the likelihood is being calculated with respect to ℑ n , the σ-field generated by the first n observations. According to (Equation3.10),
We only need to take the expectation over those X n+1 for which  > q∗.

Lemma 4.1

If X n+1 ≥ M, then  < .

Proof

Let X n+1 = m ≥ M. Let . For r = 0,

If we divide equation (Equation4.5) by given in (Equation4.3), we obtain, , where F and G are defined as

We show that . This is equivalent to showing that the first term in F is greater than the first term in G. (See Lemma A.1 Appendix for details). Therefore, . Now, let 0 < r < n:

To show that , it suffices to show that each term in the numerator times is greater than the corresponding term in the denominator of (Equation4.7) for each j, r, and n. This is equivalent to showing F j  > G j , where F j and G j axe given by

The inequality F j  > G j can be shown by induction on j, making use of Lemma A.1 in the appendix. Therefore, for all r < n. Thus,

From (Equation4.8) and (Equation4.9), we see that . Therefore,
Therefore, if the (n + 1)th observation is greater than M, the posterior probability that a change has not occurred will decrease.

We use Lemma 4.1, to evaluate the two-step-ahead stopping rule. According to equation (Equation4.1), the two-step-ahead risk function is where is defined as . We only need to take the expectation over those X n+1 for which  > q∗. Note that if  < q∗, then  < q∗ for k ≥ M. Thus for  < q∗,

Thus the two-step-ahead boundary is

Notice when  < q∗. Also when  > q∗, both the one- and two-step look ahead procedure will tell us to continue sampling. Thus the two-step-ahead procedure is equivalent to the procedure which tells us to stop when  < B n, 2. In Table , we do an example. We look at simulated values for λ1 = 1, when n ≤ 10 and λ2 = 2 when n > 10. We use a prior distribution of p = 0.01, π = 0.01. We assume we know that the upper bound on the arrival rate M = 3, and the cost c of stopping late is c = 0.06. In this example, using formula 3.9, we calculate the one-step-ahead boundary q∗ = 0.8571. We show at each n, the total number of arrivals up to time n, the posterior probability that a change has not taken place and the two-step-ahead boundary B n, 2. One can observe that the B n, 2 approaches q∗ for large n. This illustrates Lemma 3.2.

Table 1. Simulated values to show detection using the 2-step-ahead boundary

From looking at Table one observes that we would stop at time 11. Since the change took place at time 10, this shows that the detection rule is quick. We also see the convergence between the one- and two-step-ahead stopping rules.

ACKNOWLEDGMENTS

I would like to thank Dr. Shelemyahu Zacks for his helpful advice in preparing this article. I would also like to thank the Editor, Associate Editor, and referees for their helpful suggestions.

Notes

Recommended by N. Mukhopadhyay

REFERENCES

  • Brown , M. and Zacks , S. ( 2006a ). Note on Optimal Stopping for Possible Change in the Intensity of an Ordinary Polsson Process , Probability and Statistics Letters 76 : 1417 – 1425 .
  • Brown , M. and Zacks , S. ( 2006b ). On the Bayesian Detection of a Change in the Arrival Rate of a Poisson Process Monitored at Discrete Epochs , International Journal for Statistics and Systems 1 : 1 – 13 .
  • Chernoff , H. and Zacks , S. ( 1964 ). Estimating the Current Mean of a Normal Distribution Which is Subjected to Changes in Time , Annals of Mathematical Statistics 35 : 999 – 1018 .
  • Herberts , T. and Jensen , U. ( 2004 ). Optimal Detection of a Change Point in a Poisson Process for Different Observation Schemes , Scandinavian Journal of Statistics 31 : 347 – 366 .
  • Hinkley , D. and Hinkley , E. ( 1970 ). Inference about the Change-Point in a Sequence of Binomial Variables , Biometrika 57 : 477 – 488 .
  • Kander , Z. and Zacks , S. ( 1966 ). Test Procedures for Possible Changes in Parameters of Statistical Distributions Occurring at Unknown Time Points , Annals of Mathematical Statistics 37 : 1196 – 1210 .
  • Page , E. S. ( 1954 ). Continuous Inspection Schemes , Biometrika 41 : 100 – 114 .
  • Page , E. S. ( 1962 ). Cumulative Sum Schemes Using Gouging , Technometrics 4 : 97 – 109 .
  • Peskir , G. and Shiryaev , A. N. ( 2002 ). Solving the Poisson Disorder Problem , in Advances in Finance and Stochastics: Essays in Honor of Dieter Sonderman , K. Sandmann and P. J. Schonbucher , eds., pp. 295 – 312 , New York : Springer .
  • Shewhart , W. A. ( 1926 ). Quality Control Charts , Bell System Technical Journal 5 : 593 – 603 .
  • Shiryaev , A. N. ( 1978 ). Statistical Sequential Analysis: Optimal Stopping Rules , in Translations of American Mathematical Society 38 , Providence : American Mathematical Society , 149 .
  • Smith , A. F. M. ( 1975 ). A Bayesian Approach to Inference about a Change-Point in the Sequence of Random Variables , Biometrika 62 : 407 – 416 .
  • Zacks , S. ( 1983 ). Survey of Classical and Bayesian Approaches to the Change-Point Problem: Fixed Sample and Sequential Procedures for Testing and Estimation , in Recent Advances in Statistics , M. H. Rizvi , J. Rastogi , and D. Siegmund , eds., pp. 245 – 269 , New York : Academic Press .
  • Zacks , S. and Barzily , Z. ( 1981 ). Bayes Procedures for Detecting a Shift in the Probability of Success in a Series of Bernoulli Trials , Journal of Statistical Planning and Inference 5 : 107 – 119 .
  • Recommended by N. Mukhopadhyay

Appendix A

Lemma A.1

For all n, k, and m positive integers, and m ≥ M > 0,

Proof

Taking differences and utilizing the relationship between the cumulative distribution function of the Poisson distribution and the gamma distribution, we obtain

Simplifying, we obtain that equation (Equationrm A.2) is equal to

We will divide the integral into where y < x and a second integral where y > x. Thus, equation (Equationrm A.3) is equivalent to M 2k+m+1 times
Note that the second integral in (Equationrm A.4) can be rewritten as
Thus, equation (Equationrm A.2) is
g (x) = x m e Mx is an increasing function over (0, 1) when m > M. Hence,
for all 0 < y < x < 1. Therefore,
for all n, k, m, and M such that m ≥ M.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.