Research Article  Open Access
Willem Albers, "RiskAdjusted Control Charts for Health Care Monitoring", International Journal of Mathematics and Mathematical Sciences, vol. 2011, Article ID 895273, 16 pages, 2011. https://doi.org/10.1155/2011/895273
RiskAdjusted Control Charts for Health Care Monitoring
Abstract
Attribute data from highquality processes can be monitored effectively by deciding on whether or not to stop at each time where failures have occurred. The smaller the degree of change in failure rate during out of control one wants to be optimally protected against, the larger the r should be. Under homogeneity, the distribution involved is negative binomial. However, in health care monitoring, (groups of) patients will often belong to different risk categories. In the present paper, we will show how information about category membership can be used to adjust the basic negative binomial charts to the actual risk incurred. Attention is also devoted to comparing such conditional charts to their unconditional counterparts. The latter do take possible heterogeneity into account but refrain from riskadjustment. Note that in the risk adjusted case several parameters are involved, which will all be typically unknown. Hence, the potentially considerable estimation effects of the new charts will be investigated as well.
1. Introduction
We are interested in processes which exhibit only a (very) small proportion of defectives. Due to ever increasing efforts and standards, such highquality processes become more and more common in industrial setups. Moreover, for the quite different field of health care monitoring, they are in fact the rule: errors, such as malfunctioning equipment, fatally delayed help, or surgical failures, should occur only (very) rarely. Now review papers on health care monitoring (see, e.g., [1–3] and Sonesson and Bock [4]) strongly suggest to apply SPC methods, in particular control charts, and we shall follow that line here.
The common starting point for monitoring such attribute data is to watch the number of failures in a series of given sampling intervals. However, for highquality processes, this chart may not be the best choice. A first improvement is to switch to a “timebetweenevents” or “geometric” chart, which uses the number of successes between failures to judge whether the process has remained in control (). See, for example, Liu et al. [5], Yang et al. [6], Xie et al. [7], Ohta et al. [8] and Wu et al. [9], for details. When the process goes out of control (), such geometric charts quickly react to substantial increases of the failure rate from but are admittedly rather slow in detecting moderate changes. Especially in health care applications, this is undesirable, and several of the authors mentioned above have suggested a second improvement step.
Here a decision whether or not to stop is no longer made after each failure but instead only after failures have occurred. Typically, the smaller the increase of failure rate during one wants to have optimal detection power against, the larger the should be. This negative binomial chart is analyzed in some detail in Albers [10]. In particular, a simple rule of thumb is presented for selecting , and the resulting chart is both easy to understand and to apply. However, as subsequently pointed out in Albers [11], a serious complication arises if the underlying homogeneity assumption is not warranted. In industrial applications, it may often—but by no means always—be reasonable to indeed assume one and the same failure probability for each item inspected. But, in medical settings, patients tend to exhibit quite a bit of heterogeneity, and we will regularly have to take such variation between subjects into account.
In Albers [11] the basic situation is considered where in fact all we know is that such heterogeneity does occur. It can, for example, stem from the occurrence of different groups, each with its own probability of failure, but we lack further information. The only way, in which it becomes visible, is through an increase of variance, as compared to the homogeneous case. For a discussion of this overdispersion phenomenon, see, for example, Poortema [12] for a general review and Christensen et al. [13] and Fang [14] for earlier applications concerning attribute control charts. In Albers [11] it is demonstrated how the negative binomial charts can be generalized to cover the present overdispersion situation. Essentially the illfitting single parameter homogeneous model is widened there into a twoparameter model. In addition to the failure rate , a second parameter is added, in order to capture the degree of overdispersion. In view of the lack of knowledge about the precise underlying mechanism of the process, this wider family still remains an approximation of reality. But, as demonstrated in Albers [11], the results under overdispersion are far better than those provided by the homogeneous approach.
As was already pointed out in Albers [11], quite a different situation occurs when we do have information about the underlying structure. For example, suppose a number of risk categories can be distinguished, each with its own during , and for each incoming patient we register to which class he/she belongs. First of all, such detailed knowledge about the process in principle allows a more accurate analysis. But probably even more important is the fact that it opens the way to applying socalled risk adjustment methods (see, e.g., [15, 16] for an overview, and [17] for a riskadjusted version of the sets method introduced by Chen [18]). Here the baseline risk of each patient is taken into account in deciding whether the process is still . If, for example, a surgeon's performance decreases over time, an signal may nevertheless not be justified if we observe that meanwhile his/her patients are gradually shifting to higher risk categories.
Clearly this is an interesting area, giving rise to quite a few questions, both from a practical and a technical point of view. For example, in practice, one can wonder under what circumstances one should adjust the risk and when one should ignore this possibility and stick to a rigid overall chart. A more technical issue is the following. In risk adjustment several parameters are involved (cf. the above), which typically are unknown and need to be estimated. Now estimation effects for control charts tend to be conveniently ignored in practice. But if these are studied, they typically turn out to be substantial; for example, Chen et al. [19], mention a 30–90% increase in false alarm rate for a bias in the estimator for . This estimation topic was studied more systematically in Albers and Kallenberg [20, 21]. There it has been amply demonstrated that the small probabilities involved, such as , invariably produce large relative errors, and; thus, corrections are called for. In the present situation, we have not one, but several parameters, and hence the effect is likely to be even more serious. But, as Woodall [1] remarks, little work has been done on the effect of the estimation error on the performance of riskadjusted charts. Consequently, the purpose of the present paper is to remedy this by studying how the negative binomial charts from the simple homogeneous case can be adapted to the situation where risk adjustment is desirable.
As concerns the relation of the methodology proposed here to the existing methods as described in, for example, Steiner et al. [15] and Grigg and Farewell [16, 17], the following remarks are in order. The latter category are of type and as such may be slightly more efficient. In passing note that this actually is a rather subtle matter, as it also seems to depend on the type of performance criterion used (e.g., steady state or not). Nevertheless, using the negative binomial type of approach implies some aggregation of the data over time, which could indeed mean some loss of information compared to the stepwise approach. However, precisely this aggregation effect makes the resulting structure less complicated, thus, allowing a detailed analysis of estimation effects, as well as corrections of these. For the case (so far), this seems intractable. Hence, the issue in comparing the two types of approach actually is robustness. On the one hand, we have procedures which (maybe) are superior if optimal conditions hold. In the present case that means known parameters, which is almost never realistic. In addition, the impact of just plugging in estimates for these parameters is known to be huge (e.g., causing an average run length during which is systematically substantially lower than prescribed). Under such circumstances, it seems an attractive alternative to pay a small insurance premium (in the form of a small loss in detection power) in order to obtain robust procedures which allow control of the validity through adequate corrections for the charts. Incidentally, note that this robustness issue is by no means typical for the application at hand, but is of a rather general nature.
The paper is organized as follows. As far as possible, the technicalities involved are relegated to the appendix section, while the body of the paper provides the main ideas. Section 2 is devoted to introducing briefly the negative binomial chart from Albers [10], which forms our starting point. In Section 3 these charts are subsequently adapted to situations where risk adjustment is called for. The estimation aspect will be the subject of Section 4. For illustrative purposes, throughout the paper examples are presented. Moreover, at the end of the paper, we summarize the application of the proposed chart in a simple set of steps.
2. The Homogeneous Case
Here we briefly introduce the homogeneous case (see [10] for a detailed description). The process being monitored can be characterized through a sequence of independent identically distributed (i.i.d.) random variables (r.v.'s), with in the stage. During , becomes for some , and we should stop as quickly as possible. As discussed in the Section 1, the option to stop arises each time failures have been assembled, for some . Let , be the successive numbers of s involved, then these are i.i.d. as well, and moreover, distributed as a negative binomial r.v. (indices will be suppressed unless confusion might arise); that is, for , we have A stopping signal should occur the first time an , with the lower limit selected such that the false alarm rate () equals , for some small . Then the average run length () in terms of number of failures during equals for all , thus aligning the various negative binomial charts for in a proper way. Consequently, , the quantile of , which can be obtained numerically.
For the purpose of analyzing the behavior of the lower limit , a simple and transparent approximation is most useful. This can in its turn also be used for finding an approximation for the during ; that is, when has turned into . As this is a function of , and , it is in particular interesting to figure out which choice of is the best for given and . This task has been carried out in Albers [10], to which we refer for a complete description. However, to facilitate independent reading, we do present some of the details in the Appendix section. At this point we just mention that , where solves , in which the r.v. is Poisson’s with parameter . Moreover, by considering a table of numerically obtained optimal values of (i.e., resulting in the lowest ) for the various and of interest and fitting to these a simple approximation in terms of , , and , the following easy rule of thumb has been obtained: As can be seen from Table 3 in Albers [10], this simple rule works remarkably well. Quite often the solution from (2.2) is truncated at 5: most of the improvement over the geometric chart (where ) has already been achieved at . Moreover, using really large values of may be considered undesirable in practice anyhow.
As already observed in the Section 1, the underlying parameters typically need to be estimated in practice. In the simple homogeneous case, we merely have to worry about the failure rate . To this end, a Phase I is added, during which the process is observed until failures have occurred, leading to geometric 's (cf. (2.1)) and, thus, to the obvious estimator , where . Then we simply replace by in (or ), after which the actual monitoring can start using the new (or ). Consequently, application of the chart is almost as easy as that in the known parameter case. However, it does remain to investigate, and possibly to correct, the resulting estimation effects. What happens is that performance characteristics such as (or ) are no longer fixed at (or ) but instead will be random in . Noting that satisfies and , it follows that the Taylor expansion in terms of allows adequate appraisal of errors such as . This can be done in terms of bias by taking expectations or in terms of exceedance probabilities by, for example, looking at . Next, if desired, suitable corrections can be derived. Then slightly lower limits are used in which the small positive is selected such that either the bias is removed or the exceedance probability will stay below a suitably small upper bound. To illustrate the simple application of the homogeneous chart, we conclude the present section with an explicit example.
Example 2. Suppose we are monitoring some large population w.r.t. some unpleasant type of event (“a failure”). As long as matters are , its incidence rate is . Stopping during should be rare, so we decide that on average a false alarm once every 200 failures is acceptable, implying that the and . Deciding about stopping or continuing at each failure (the geometric chart, with ) is known to be not very efficient, so we increase somewhat, and, for example, take . (Background: the rule of thumb (2.2) tells us that this is optimal at the given for , that is, for an increase during to ; someone interested in optimality for lower should take an even somewhat larger ; for example, is the best for ). During , each third failure will on average arrive after 3000 observations, and, hence, the chart should signal if this arrival happens much sooner. What “much sooner” should mean under the given conditions has been derived above: the exact lower bound here equals . Using which solves produces , and hence as well, while the further approximation through (A.2) gives the quite close value 506. Monitoring is now completely straightforward: watch the sequence of realized “thirdfailuretimes” and stop as soon as a number at or below 509 (or 506) is observed. If the value of is in fact unknown, monitoring is preceded by a Phase I sample. First wait till, for example, failures have been obtained and replace by the outcome . If happens to equal 0.001, the above remains as it was; for any other value, the computation is adapted in a straightforward way.
3. The Heterogeneous Case
Our starting point is the homogeneous case described in Section 2: an underlying sequence of i.i.d. Bernoulli r.v.'s , giving rise to negative binomial r.v.'s for controlling the process. To this situation we now add heterogeneity, and subsequently we investigate how to accommodate this complication, both in the unconditional case, where the underlying information is ignored, and the conditional case, where it is used. Of course, the emphasis will be on the latter situation, where risk adjustment is applied. In fact, the unconditional case has already been dealt with in Albers [11]; we only review it here to provide the proper perspective.
In line with the common notation for bivariate situations, we will use 's for the main r.v.'s, on which the control decisions are based and 's for the r.v.'s supplying additional information. To be more precise, we will use as a risk category indicator: , , for some . If this is done individually, that is, for each separately, the unconditional situation remains simple. In fact, ignoring or not having the information from the 's means that effectively we still are in the homogeneous case. Each incoming patient simply has the overall failure probability , with during , and thus the negative binomial charts from Section 2 are still appropriate.
However, matters change if patients arrive in groups, and heterogeneity does lead to nonnegligible overdispersion. In the Appendix section, this is demonstrated in some detail for the relatively simple setting of a fixed number of groups, each of fixed size, to which routinely a standard chart would be applied. Moreover, in this context, the difference in behavior between unconditional and conditional charts is analyzed. An excursion to the situation of continuous rather than attribute data provides additional clarity, as matters for the normal case are somewhat more transparent. The conclusion is that, whenever goes , two situations should be distinguished. In the first, the p.d.f. of (and, thus, the ) remains unchanged. Then the two charts react essentially in the same way, with the conditional chart being somewhat more efficient, as it takes the additional information into account. However, once the p.d.f. of also changes during , the behavior of the charts will diverge. Which of the two provides the right answer depends on whether these changes in should be ignored or taken into account. (cf. the surgeon from Section 1 who is faced with increasingly more risky patients).
Our present setup is of course even more complicated: group sizes will typically neither be fixed nor equal. As a consequence, the p.d.f. of the waiting times till the failure is not only no longer negative binomial but in fact rather intractable in general. For this reason Albers [11] proposes by way of approximation a parametric model containing an additional parameter to cover the overdispersion aspect (c.f. the references mentioned in the Appendix section for accommodating tail length in case of the normal mixture model). In this way an adequate analysis of the unconditional chart indeed becomes possible; see Albers [11] for details.
Having set the background and explained the interpretation, we can now fill in the details for the riskadjusted chart in a straightforward way. Just as in the homogeneous case, define a new sequence on the basis of the 's. Here the are i.i.d. copies of , the number of to be observed until the failure arrives. We decide to stop as soon as a , for some suitable lower limit . In line with Section 2, “suitable” will be interpreted to mean that for some small, given . To obtain in the present case, note that while we are waiting for the realization of an , at each time , we now have at our disposal the information that , where is the number of patients from category , . Arguing along the same lines as in Section 2, we obtain (once more see the Appendix section for some of the details) that the simple approximation from the homogeneous case , with solving , is updated into (cf. (C.2)): Note that from (3.1) can obviously be written as , with weights . Hence, if we are in the homogeneous case after all; that is, , we get back exactly. Moreover, during , the will tend to be close to the underlying probabilities , and will hold approximately for arbitrary as well.
Nevertheless, we should observe that, unlike in the homogeneous case, we no longer have a single fixed . Each of the realizations from the abovementioned sequence has its own (and; thus, writing would be appropriate now). In fact, we do not even need to obtain all of these explicitly. For example, in those cases where a signal occurs, we have , which means that in this situation at time we still have . Evaluating the actual would imply continuing with the 's corresponding to . That would not be the right thing to do; fortunately, it is also superfluous, as the information suffices that the event has occurred.
Example 3 (continued). Suppose that for the situation considered in Section 2 additional information, and hence the possibility for risk adjustment, has become available. To keep matters again as simple as possible, we distinguish just two risk categories: “mild” and “severe.” Suppose of the population is severe and their risk is 11 times as high as that of the mild cases. Hence,, and thus , resulting in and . After (3.1) it was observed that the riskadjusted chart replaces the homogeneous choice by , in this way, taking the actually observed weights into account. Here this means using , which boils down to . During , will be close to , and hence will be close to 509, as in the homogeneous case. Nevertheless, monitoring becomes slightly less trivial than before: now not only the sequence of realized third failure times but also the corresponding fractions of severe cases need to be recorded. For each data pair , it is checked whether . For example, an outcome () produces a signal in the homogeneous case, (“”), but not in the riskadjusted situation (“”). In the latter case the occurrence of the third failure at a very early moment is deemed acceptable after all in view of the somewhat larger than expected presence of severe cases.
The step in (3.1) is essentially all that is needed to deal with the risk adjusted version of the chart. In the Appendix section, it is demonstrated how the approximation steps for, for example, and the carry over. Of course, if the process goes uniformly , in the sense that all are replaced by , matters are most straightforward. But even if each has its own , it remains easy to adapt the previously obtained expressions (cf. (C.3)). It is also explicitly demonstrated how the abovementioned weights influence matters. If these remain close to the , the riskadjusted chart behaves in the same manner as its unconditional counterpart. It only is slightly more efficient. However, if the ’s are quite different, the behaviour will diverge. To illustrate this, just as in the normal example from the fixed case, an explicit example is given in the Appendix section of an situation for which is completely due to the change in . The riskadjusted chart then indeed continues to consider the process as being . To illustrate matters, we conclude the present section with numerical example.
Example 3 (continued). In the setup considered before, let us next consider the situation. First let uniformly; that is, the mild category gets and the severe one . From, for example, Albers [11], we obtain that the homogeneous chart has an of 36 for this case, and, according to the above, this will continue to hold here as well. Next consider a nonuniform example: let , and let , then , and; hence, once more . If during the p.d.f. of remains the same, the ’s do not change, and will still be close to . Then the riskadjusted chart will continue to behave like the homogeneous one, that is, with an of about 36 at this . However, if instead the ’s remain the same and the ’s are used to transform into , the chart will shift towards using a lower limit close to . Moreover, as in (C.3) for this particular choice, its will remain at about the value 200. Clearly, from a risk adjustment perspective, this is precisely what should happen. The mild patients still have , and the severe ones still have . What has changed is that the latter category has shifted from to of the total. Hence, not the quality of the performance has deteriorated, but rather that of the incoming patients (cf. the surgeon example in Section 1).
4. Estimation Effects
Typically the underlying parameters of control charts are not known in practice. Here this means that we will have to estimate not just the overall , but the individual as well. As concerns , in Albers [10] a Phase I sample of “size” was used for this purpose, meaning that we observe until failures have arrived. Note that , and hence the length of this sample as well, is independent of , implying that the comparison between charts for different remains fair also w.r.t. estimation. Next, was simply estimated by . In the present context we can use this same sample, but in a more detailed way, as follows. Let be its number of patients from category (i.e., ), and in addition let , denote the corresponding 's. Then we have as a straightforward choice for estimating the :
Of course, formally there is a problem in (4.1), as each can be 0 with positive probability. Using a slightly modified definition could remedy this. However, we shall not bother to do so, as the probabilities involved are exponentially small. Moreover, if too small ’s are observed, this anyhow indicates that the design may not be right, and additional effort is required before monitoring can begin. Given , we have and . As , it follows that, ignoring terms involving , with “” denoting asymptotic normality. Hence, if the Phase I sample is chosen in this way, the estimators are indeed well behaved, in the usual sense of having a relative error which is . Only if the contribution of a certain category is really small compared to , the coefficient involved will be large.
In fact, the above is all that is needed to transform the chart into its estimated version: just replace in (3.1) the by their estimated counterparts from (4.1). To be precise, we use a lower limit defined by where once again is such that . The alternative notation then becomes , with weights . Once this lower limit has been obtained, the actual monitoring can start: each time wait till the failure, and if this occurs at or before , a signal results. Hence, straightforward application of the estimated chart remains easy.
To investigate the effects of the estimation step, we proceed as follows. As remarked in Section 2, will no longer be fixed at some given value nor will be precisely equal . These performance characteristics have now become the random variables and , respectively. They depend on the estimators and consequently fluctuate around the intended values. A rather detailed analysis of the consequences for the homogeneous case can be found in Albers [10]. To avoid repetition, we shall be much more brief here and mainly focus on the additional complication caused by the fact that the estimation step is now split into categories. See once more the Appendix section for the details. It is evaluated how large the size of the Phase I sample should be in order to ensure that the exceedance probabilities discussed at the end of Section 2 will fall below a prescribed small quantity . If the resulting is too large for use in practice, a small correction is calculated, such that for a given the desired can be met after all by using the slightly more strict rather than itself. A complication here in comparison to the homogeneous case is the presence of an additional quantity (see (D.5)), which represents the unbalance caused by (possible) differences between the ideal and the actually occurring .
To illustrate matters we again present an explicit example.
Example 4 (continued). Once again, we add a complication to our ongoing example: now the values of and are no longer known. Suppose, therefore, that we first wait till failures have occurred and then use the resulting Phase I sample to estimate these (see (4.1)). Just as in the example of Section 2, the application of the chart remains straightforward: simply plug in the to replace the unknown . However, if we also want to study the impact of the estimation step, and possibly correct for it, a bit more effort is needed (we shall refer to the Appendix section for the required formulae). The point is that, in the subsequent monitoring phase, we may still aim at an of 200, but we have to accept that the actual may differ. In particular, this can substantially be smaller than 200, leading to a lot more frequent false alarms than anticipated. To monitor not only the process but this effect as well, we can, for example, look at the probability of a deviation of more than , that is, of the falling below 160. In terms of (D.4), this is precisely with , and; thus, . As moreover , it then follows from (D.6) that . Since is close to, but smaller than, 1, a close upper bound for is . As long as is (regardless of whether the same holds for or not), will be close to 1 as well, in which case this upper bound approximately boils down to . If such a probability for a more than too low is acceptable, we can continue to the monitoring phase. Otherwise should be larger than 100; to be precise, is needed to get . Or, alternatively, for fixed at 100, this can be achieved by using with (cf. (D.7)). For, for example, , we have and and result. If in fact is , will be larger than 1, and we need to use instead. For example, consider once more the situation from the previous example, where , and subsequently . Then in addition to , we obtain , which leads through (D.5) to . Hence, to still get , should now be 131, rather than just 100. This increase is still rather mild, but it is easy to see that higher increases of can be necessary. For example, slightly generalize our example into for some small , while otherwise keeping and . In that case , producing and . This leads to , which, for example, for equals 1.9 and for already 3.8. Of course, the latter case is rather extreme, as the frequency of severe cases has increased by a factor 15.
To conclude this section, for convenience, we briefly summarize the steps involved in applying the new chart (see Algorithm 1).

Appendices
A. Approximations for the Negative Binomial Chart
In addition to a numerical solution for , it is desirable to derive a simple approximation as well, for example, to make transparent how the function behaves. Now with here binomial with parameters and and Poisson with parameter . Hence, , with solving . For , , and (which region amply suffices for our purposes), it is demonstrated in Albers [10] that this can be approximated by in which . The resulting approximation turns out to work well over the region considered.
During we have , with still such that . Again an approximation is feasible: with and as in (A.2). It is adequate for the region above and . The improvement achieved by increasing beyond 1 can be nicely judged by looking at . Due to the alignment during , these functions all start with value 1 at , after which they increase considerably. Only for really large , the decrease to the limiting value sets in.
B. Illustration for a Fixed Number of Groups of Fixed Size
Let us assume in this subsection that the patients from a given risk category arrive in groups of fixed size , for some . (Hence, stands for the special case of individual arrivals). With the risk category indicator satisfying , the number of defectives in such a group hence satisfies . Suppose we wait till such groups have arrived, and thus patients have been seen. Since and , it immediately follows that and . For controlling the process in this simple setup, we might want to apply the standard chart, based on . But, since , it is immediate that an overdispersion effect becomes apparent as soon as , that is, in the case of “real” groups. Now, and we are focusing on highquality processes, meaning that and the are (very) small. Consequently, the term , being of order , might still seem negligible. However, if is of order comparable to , as will often be the case, the first term in behaves as and the overdispersion correction as . Typically, such terms will be of the same order of magnitude, and heterogeneity does have an impact.
Hence, even in the unconditional case (where the underlying information is unavailable or simply ignored), a suitable modification should be applied to the standard chart. To be more precise, the usual upper limit it uses is , where , with denoting the standard normal p.d.f. Clearly, this should now be replaced by . Next, to make the step towards the conditional case (i.e., the risk adjusted counterpart), note that we can also write with multinomial and . Performing control conditional on the observed implies that in the riskadjusted case is compared to the upper limit . To appreciate the difference in behavior between the unconditional and the conditional version of the chart, just drop the conditioning on the and use that and . It then readily follows that the expectation of the conditional variance is, thus, increased by the nonnegligible amount . This result (as it obviously should) agrees with the form of derived above for the unconditional case.
Consequently, the picture is as follows: as long as the ’s remain unchanged once goes , both charts react similarly, with the riskadjusted version being the more precise one (as it has a smaller variance). However, if going affects not only but also (and, thus, the ), the behavior will diverge, and the choice will become dependent on the aim one has in mind.
To illustrate this conclusion more clearly, we conclude this subsection by making a twostep excursion to the corresponding normal case of monitoring the mean of a continuous process variable by means of a Shewhart chart. Hence, now is , that is, has p.d.f. , and an upper limit is used in the homogeneous case. In the presence of the group indicator , this setup transforms into being . For the unconditional case we then readily obtain that and , but we do note that now has a mixture distribution and, hence, is no longer normal. Such violations of the normality assumption in fact occur quite often in practice, and Albers et al. [22, 23] demonstrated how using an additional parameter concerning the tail length of the actual distribution can adequately remedy the ensuing model error. On the other hand, for the conditional case matters remain pretty trivial, just use for the appropriate as upper limit. Comparison of the two versions runs completely parallel to that for the attribute data above.
To obtain the intended additional clarity, a second step is needed: instead of using a category indicator, now let be normal as well. Hence, we have pairs which are bivariate normal . Unconditional monitoring then means using the upper limit , while working given the observed outcomes means applying , in view of the fact that is . Such a method of using auxiliary information has recently been discussed by Riaz [24]. The advantage of this bivariate case is that it allows a very simple comparison between the two approaches. For let now imply that the will be , then the conditional approach has an , where
Indeed, if that is, the ’s remain unchanged whenever the ’s go , then clearly , unless . Hence, in this situation, inspection using the auxiliary information from the leads to lower and, thus, to better performance, as argued in Riaz [24]. However, if is positive as well, can be smaller than and even 0 (just let , meaning that no situation is perceived. Whether this is right or wrong just depends on the perspective used. If , the behavior of is shifted “only” because of the shift in the underlying . In that sense “nothing” has happened, and not reacting may well be the appropriate response. However, if the behavior of should be judged against a fixed standard, irrespective of possible shifts in , one should clearly stick to the unconditional version by using .
C. Approximations for the Heterogeneous Case
If we denote a binomial r.v. with parameters and by , then, in analogy to (A.1), we have, with ,
For , typically will be large, and, if is (relatively) small, the will be large as well. Hence, the Poisson approximation step from (A.1) can be used here again. In fact, quite conveniently it remains possible to keep using such that ; only the relation between the lower limit and this becomes slightly more intricate than in the homogeneous case: (More formally: look for the largest value of such that the corresponding satisfies . However, as the ’s are small, this distinction is rather futile.)
Till now we excluded the case , as is not necessarily large there and the Poisson step might not be warranted. Just as in the homogeneous situation, this boundary case can be simply solved exactly. In fact, . This leads to and, thus, to, for example, , which is in line with the result from (C.2) (see [10] for details).
Hence, unlike in the unconditional case, where the Poisson approximation has to be replaced by a negative binomial one, most of the results during from the homogeneous case carry over to the riskadjusted version and the modification needed is actually already covered by (C.2). For example, all results from Albers [10] about approximating (cf. (A.2)) can be used directly. Once the process goes , in principle this carrying over continues, certainly if we translate the step from to in a uniform way into replacing each by . Then , in which , with still such that . Hence, an approximation like from (A.3) continues to hold as well.
More generally, going will mean that becomes , with the such that still . Using (C.1) again, we obtain that , with once more such that , but now Note that if the p.d.f. of the remains unchanged when the ’s go , the weights will remain close to the , and; thus, from (C.3) will in fact be close to . Hence, results from the homogeneous case (such as (A.3)) continue to hold approximately. In other words, the riskadjusted chart indeed shows behavior quite similar to that of the unconditional chart for the individual case. Moreover, it is somewhat more precise than the unconditional chart from the group arrival case, the difference being that the latter is based on a negative binomial rather than a Poisson p.d.f.
To demonstrate that matters can become quite different when the p.d.f. of is affected as well during , we argue as follows. First of all, note that there is no need that for all . To see this, without loss of generality, suppose that the ’s are ordered into increasing order of magnitude. Now assume that we have an increasing sequence of such that not only the but also the are in fact probabilities. For such indeed will hold, as follows, for example, by noting that monotone likelihood ratio implies increasing expectation [25, page 74]). Hence, this choice can serve as yet another example of the case discussed in the previous paragraph: the ’s remain unchanged, the ’s turn into , leading to an overall factor .
However, we can also choose to associate the in with rather than with . In other words, the weights are shifted from to , whereas the remain as they are. Consequently, the conditional chart uses and in (4.3) and, thus, arrives at . Hence, after riskadjustment, everything is still perfectly , and the chart sees no reason for action at all. On the other hand, the unconditional charts keeps observing that and will tend to react. As argued before, both answers are right; only the underlying questions differ.
D. Effects of and Corrections for the Estimation Step
In analogy to (C.2), we obtain for the present situation that with . From (4.3) it follows that and, thus, that , where This is precisely the same structure as we already had in (4.4) from Albers [10], the only difference being that there simply equals . Hence, the subsequent steps are quite parallel, and we shall not go into much detail. The idea again is to expand w.r.t. . Since , we arrive at where . According to Lemma 4.1 from Albers [10], this satisfies , implying that it will typically be close to 1.
From (D.3) we can readily evaluate desired quantities of interest for judging the impact of the estimation step, such as the exceedance probability: in which is some small positive constant, like 0.25. Note that, since , we can also write , so both types of performance characteristics are dealt with simultaneously through (D.4). Indeed, from (D.3), it immediately follows that . In the homogeneous case, we simply have that is , and thus the corresponding (see (4.10) in Lemma 4.3 of [10]). For the present case, we combine (4.2) and (D.2). First note that and also that (4.2) holds not only for but also for . Consequently, once more ignoring quadratic terms in the , we obtain that the present is , where
Hence, it follows that with as in (D.5).
It is immediate from (D.6) that ensuring for a certain small , like 0.10 or 0.20, requires . Moreover, by for example applying CauchySchwarz to the r.v.'s and , it is clear that , and; thus, , with equality occurring only if . As we observed in Section 3, the latter situation will be approximately true during and, moreover, during , when only is affected and is not. However, as soon as the behaviors of the riskadjusted and the unconditional charts start to differ, this will mean that the ’s are no longer close to the , and then the exceedance probability will start to be larger than its homogeneous counterpart. The verbal explanation is quite straightforward: as remarked after (4.2), the relative precision of the increases in . As long as the ’s are such that the weights ’s are close to the supposed , this effect is adequately balanced in (D.2). However, once rare categories become more prominent, this balance is disturbed.
Through (D.6) we can check the behavior of the estimated chart and prescribe the minimal to ensure a desired bound on . However, if this cannot be realized and we are stuck with some given lower value, this bound can also be achieved by using a somewhat more strict lower limit , with as in (4.3) and small. Then from (D.1) becomes , and; thus, is replaced by in what follows, leading through (D.6) to A requirement like can now be enforced by choosing (cf. (4.11) from Albers [10]). Indeed, the term of order shows that is small, and the correction becomes superfluous (i.e., ) as soon as reaches the aforementioned value . Incidentally, note that, unlike the other quantities involved, is not yet known during Phase I. Hence, the bound on is valid as long as the actually observed values of after Phase I are at most as large as the one used beforehand in obtaining and .
References
 W. H. Woodall, “The use of control charts in healthcare and publichealth surveillance,” Journal of Quality Technology, vol. 38, no. 2, pp. 89–104, 2006. View at: Google Scholar
 J. Thor, J. Lundberg, J. Ask et al., “Application of statistical process control in healthcare improvement: systematic review,” Quality and Safety in Health Care, vol. 16, no. 5, pp. 387–399, 2007. View at: Publisher Site  Google Scholar
 S. H. Shaha, “Acuity systems and control charting,” Quality management in health care, vol. 3, no. 3, pp. 22–30, 1995. View at: Google Scholar
 C. Sonesson and D. Bock, “A review and discussion of prospective statistical surveillance in public health,” Journal of the Royal Statistical Society. Series A, vol. 166, no. 1, pp. 5–21, 2003. View at: Publisher Site  Google Scholar  MathSciNet
 J. Y. Liu, M. Xie, T. N. Goh, and P. Ranjan, “Timebetweenevents charts for online process monitoring,” in Proceedings of the IEEE International Engineering Management Conference (IEMC '04), pp. 1061–1065, October 2004. View at: Google Scholar
 Z. Yang, M. Xie, V. Kuralmani, and K. L. Tsui, “On the performance of geometric charts with estimated control limits,” Journal of Quality Technology, vol. 34, no. 4, pp. 448–458, 2002. View at: Google Scholar
 M. Xie, T. N. Goh, and X. S. Lu, “A comparative study of CCC and CUSUM charts,” Quality and Reliability Engineering International, vol. 14, no. 5, pp. 339–345, 1998. View at: Google Scholar
 H. Ohta, E. Kusukawa, and A. Rahim, “A CCCr chart for highyield processes,” Quality and Reliability Engineering International, vol. 17, no. 6, pp. 439–446, 2001. View at: Publisher Site  Google Scholar
 Z. Wu, X. Zhang, and S. H. Yeo, “Design of the sumofconformingrunlength control charts,” European Journal of Operational Research, vol. 132, no. 1, pp. 187–196, 2001. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 W. Albers, “The optimal choice of negative binomial charts for monitoring highquality processes,” Journal of Statistical Planning and Inference, vol. 140, no. 1, pp. 214–225, 2010. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 W. Albers, “Control charts for health care monitoring under overdispersion,” Metrika, vol. 74, no. 1, pp. 67–83, 2011. View at: Publisher Site  Google Scholar
 K. Poortema, “On modelling overdispersion of counts,” Statistica Neerlandica, vol. 53, no. 1, pp. 5–20, 1999. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 A. Christensen, H. Melgaard, J. Iwersen, and P. Thyregod, “Environmental monitoring based on a hierarchical Poissongamma model,” Journal of Quality Technology, vol. 35, no. 3, pp. 275–285, 2003. View at: Google Scholar
 Y. Fang, “ccharts, Xcharts, and the Katz family of distributions,” Journal of Quality Technology, vol. 35, no. 1, pp. 104–114, 2003. View at: Google Scholar
 S. H. Steiner, R. J. Cook, V. T. Farewell, and T. Treasure, “Monitoring surgical performance using riskadjusted cumulative sum charts,” Biostatistics, vol. 1, no. 4, pp. 441–452, 2000. View at: Google Scholar
 O. Grigg and V. Farewell, “An overview of riskadjusted charts,” Journal of the Royal Statistical Society. Series A, vol. 167, no. 3, pp. 523–539, 2004. View at: Publisher Site  Google Scholar  MathSciNet
 O. A. Grigg and V. T. Farewell, “A riskadjusted sets method for monitoring adverse medical outcomes,” Statistics in Medicine, vol. 23, no. 10, pp. 1593–1602, 2004. View at: Publisher Site  Google Scholar
 R. Chen, “A surveillance system for congenital malformations,” Journal of the American Statistical Association, vol. 73, pp. 323–327, 1978. View at: Google Scholar
 R. Chen, R. R. Connelly, and N. Mantel, “The efficiency of the sets and the cuscore techniques under biased baseline rates,” Statistics in Medicine, vol. 16, no. 12, pp. 1401–1411, 1997. View at: Publisher Site  Google Scholar
 W. Albers and W. C. M. Kallenberg, “Estimation in Shewhart control charts: effects and corrections,” Metrika, vol. 59, no. 3, pp. 207–234, 2004. View at: Publisher Site  Google Scholar  MathSciNet
 W. Albers and W. C. M. Kallenberg, “Are estimated control charts in control?” Statistics, vol. 38, no. 1, pp. 67–79, 2004. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 W. Albers, W. C. M. Kallenberg, and S. Nurdiati, “Parametric control charts,” Journal of Statistical Planning and Inference, vol. 124, no. 1, pp. 159–184, 2004. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 W. Albers, W. C. M. Kallenberg, and S. Nurdiati, “Exceedance probabilities for parametric control charts,” Statistics, vol. 39, no. 5, pp. 429–443, 2005. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 M. Riaz, “Monitoring process mean level using auxiliary information,” Statistica Neerlandica, vol. 62, no. 4, pp. 458–481, 2008. View at: Publisher Site  Google Scholar
 E. L. Lehmann, Testing Statistical Hypotheses, Wiley, New York, NY, USA, 1959.
Copyright
Copyright © 2011 Willem Albers. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.