Abstract

This paper considers a more general shock model with insurance and financial risk background, in which the system is subject to two types of shocks called primary shocks and secondary shocks. Each primary shock causes a series of secondary shocks according to some cluster pattern. In reliability applications, a primary shock can represent an issue of insurance policies of an insurer company, and the secondary shocks then denote the relevant insurance claims generated by the policy. We focus on the local cumulative shock process where only a certain number of the most recent primary and secondary shocks are accumulated. This process is a very new topic in the available literature which is more flexible and realistic in modeling some more complex reliability situations such as bankrupt behavior of an insurance company. Based on the theory of infinite divisibility and stable distributions, we establish a central limit theorem for the local cumulative shock process and obtain the conditions for the process to converge to an infinitely divisible distribution or to an -stable law. Also, by choosing the proper scale parameters, the process converges to a normal distribution.

1. Introduction

A shock model in reliability is an operating system subject to successive shocks of random magnitudes and random arrival times. In applications, such a system may appear in engineering, economics, and natural objects. The shocks can be overload and abrupt changes of temperature or voltage for a mechanical device and electronic equipment and be natural disasters for an ecological system or financial crises for an economic system. These shocks have impacts on the system and cause the final breakdown of the system.

In the literatures, the standard mathematical setup for the reaching law of shocks is a stochastic point process with the shock instants , and the shock magnitudes come from a family of nonnegative random variables . The main object in focus is the lifetime (or the failure time) of the system. Let denote the lifetime and be the prefixed threshold of the system, then the two classical cases, cumulative shock model and extreme shock model, are defined, respectively, as for arbitrary . Hence, shock effects are accumulated in the cumulative shock case, and the system fails just when the cumulative magnitudes exceeds the threshold level. While in the extreme shock case, shock effects are memoryless and the system breaks down as soon as the magnitude of an individual shock is larger than the threshold. With relations (1) and (2), the so-called cumulative shock process and maximum shock process can be defined by respectively. It is clear that the lifetime properties of the system are determined entirely by the processes in the cumulative case and in the extreme case.

Due to the important theory value and the broad application areas, shock models remain an academic focus in reliability researches during the last three decades. The main literatures on the two types of shock models include Agrafiotis and Tsoukalas [1], Bai et al. [2], Gut [3, 4], Gut and Hüsler [5], Igaki et al. [6], Skoulakis [7], Finkelstein and Marais [8], Mercier and Pham [9], Omey and Vesilo [10], Sumita and Zuo [11], Wang et al. [12], and others. In these works, various reliability backgrounds are provided, the distributed characteristics of the system lifetime are discussed, and the asymptotic properties of the cumulative and maximum shock processes are investigated. In general, the study develops along the two directions. One is gradually profound exploration of the reliability properties of models. We can find that, with an evolvement of research objects from the early simple systems to some complex systems presently, the key problems also turn to distribution properties and limiting behaviors of the system lifetime and estimations for the failure probability from the lifetime distribution classes. The other is realistic extensions of models. In the past several years, the model structure and the relevant failure mechanism are rechanged or readjusted based on classical models to meet the various features of the real reliability systems, which is regarded as a main trend in current research.

In this paper, we setup a new shock model (called local cumulative shock model) based on a cluster point process and discuss the limit properties of the relevant shock process (called local cumulative shock process). Under the cluster structure, a primary shock, whenever it occurs, can trigger a series of secondary shocks, and the system fails when the superposed effect of primary and secondary shocks just over a certain local period exceeds the threshold. This model is an extension of the classical extreme shock model. With a more practical structure and failure mechanism than classical models, it is suitable to describe some complex reliability systems relating to earthquake disaster, network failure, and insurance risk.

The rest of this paper is organized as follows. In Section 2, we present a local cumulative shock model with a cluster structure and give the basic assumptions. The fundamental properties and weak limit theorems (based on the infinite divisibility and Lévy-Khintchine representation) of the local cumulative shock process are discussed in Sections 3 and 4, respectively. Finally, Section 5 concludes the paper.

2. Local Cumulative Shock Model with Cluster Structure

At the beginning, we list the main notations which are used in this paper. primary shock process with shock instants , which is assumed to be a nonhomogeneous Poisson process. intensity function of , and then cumulative intensity function is . magnitude of the th primary shock. arrival process of secondary shocks triggered by the th primary shock. For , ’s are assumed to be i.i.d. stochastic point processes. magnitude of the th secondary shock caused by the th primary shock. cumulative shock process which is a superposition of a primary process and a group of secondary processes. local cumulative shock process defined as . moment generating function of . characteristic function of . characteristic triplet of , where is a Lévy measure. regularized process of with centering function and regularizing function . characteristic function of . characteristic triplet of with Lévy measure . i.i.d. random variables defined on with common distribution function , where . indicator function of a random event .

A generalized cumulative shock model and its lifetime properties are already discussed in our latest work [13], where the system considered is subject to two types of shocks, called primary shocks and secondary shocks, respectively, and each primary shock causes a series of secondary shocks according to a “cluster” mechanism. Then, the shock process has a cluster structure and is a superposition of a primary (shock) process and a group of adjunct (shock) processes, and the system fails once the totally superposed effect of the primary and secondary shocks exceeds the threshold level. Let be the primary process with shock points and let be the adjunct process caused by the th primary shock. The relevant cumulative shock process can be defined, through the totally superposed shock effect by time , aswhere and represent the occurrence instant and magnitude of the th primary shock, ; and for each , is a point sequence satisfying ; then, and represent the occurrence instant and magnitude of the th secondary shock caused by the th primary shock; ; is the indicator function of event .

In particular, the coefficient “” in Model (5) can be set differently to describe the different superposition patterns of primary shocks and secondary shocks for different applications. For example, for a seismic hazard, the total damage is the accumulated effect of both the main-quake (the primary shock) and all associated after-quakes (the secondary shocks). In this case we set . This is a very natural situation and the details for earthquake cluster background can be found in Ogata [14], Daley and Vere-Jones [15], and relevant references therein. Also, an upgrade of a computer software does not affect the computer system, but it may induce some software consistency issues which affect the system operation for a period of time. This is just the situation of , where only secondary shocks damage the system and all primary shocks are ineffective. The relevant cases are discussed by Hohn et al. [16] and Fäy et al. [17]. Moreover, we consider the repair on some system as the primary shock, each repair is imperfect and causes a cluster of redundant faults which can be considered as the secondary shocks. Thus, the two types of shocks have the opposite effects to the system and . Another interesting interpretation of this case is an insurance risk issue, whenever the insurer company issues an insurance policy and charges a corresponding premium, it has to burden a series of potential claim risks induced by this policy. Where a policy premium means a primary shock and the insurance claims play the secondary shocks. For the corresponding details of insurance and finance applications, please see mainly Rolski et al. [18], Denuit et al. [19], and Lindskog and McNeil [20].

With the above cluster structure, the new cumulative shock model associating with (5) is more appropriate for modeling some real and complex reliability situations. However, is the system’s failure behaviour dependent purely on the total accumulation of shocks over , the entire history of the system operation? An observation on the fate of Lehman Brothers Holdings implies a negative answer. Indeed, it is difficult to convince that a century-old insurance or financial firm will be immortal as its long successful experience. In fact, for many realistic reliability systems such as an earthquake-prone region and an insurance or financial company, an unpredictable misfortune may be initiated by some momentous events only recently rather than very long ago. This suggests that we concentrate on the impact events over some latest period and to regard a new reliability issue.

By this background, we consider a further case based on Model (5). For some , letwhere denotes the arrival counts of a counting process in the interval . Thus, measures the cumulative effect of both primary and secondary shocks over a recent period of length and induces a stochastic process , according to which we can define a new lifetime of the system asand the relevant failure event can be expressed asIt establishes a novel failure mechanism: the system fails if and only if the superposed effect of the primary and secondary shocks accumulated in the recent duration , rather than the entire history as in Model (5), exceeds the threshold level.

Associating with the failure mechanism (8) is a new shock model; we call the local cumulative shock model with cluster structure, whose lifetime property is completely determined by , the local cumulative shock process. Note that when , the failure occurs only when the magnitude of a single shock exceeds , resulting in the classical extreme shock model defined by (2) and (4). Also, as for each given , and of (5) and the new model becomes the generalized cumulative shock model of cluster structure.

To illustrate the differences between the two shock processes and , we give a group of MATLAB numerical simulations below. Where the primary process follows a homogeneous Poisson process and the adjunct processes are an independent and identically distributed (i.i.d.) family of homogeneous Poisson processes. When the secondary shocks have an i.i.d. light-tailed distribution (an exponential distribution is used here), the sample paths of the cumulative shock process and the corresponding local cumulative shock process with the case of are shown in Figure 1, and the sample paths with the case of in Figure 2, respectively. Meanwhile, the sample paths of and with the secondary shocks of an i.i.d. heavy-tailed distribution (heavy-tailed Weibull distribution here) display as Figure 3 () and Figure 4 (), respectively.

These illustrations show some distinct characteristics between the local cumulative shock process and the cumulative shock process . For example, (i) taking different values ( or ) of the parameter (which means different superposition patterns of primary shocks and secondary shocks) changes the trend of , but it seems to have no obvious influence on ; (ii) when the secondary shocks follow a heavy-tailed distribution, the sample paths of have more drastic upward jumps than those in the case of has a light-tailed distribution. The similar differences between the two cases can also be found from the sample paths of ; (iii) it can be observed that fluctuates more intensely than in the same time span, whether the secondary shocks have light-tailed or heavy-tailed distributions, also whether the coefficient or . In summary, these characteristics show stronger randomness (or lesser predictability) for the local cumulative shock process than for the cumulative shock process and provide a more reasonable explanation for the bankruptcy behavior in insurance and financial industry such as Lehman Brothers Holdings: the fateful risk is more like short-term determined and not necessarily long-history dependent.

In this paper, we consider the local cumulative shock model with cluster shock structure and mainly focus on the weak convergence of , the local cumulative shock process defined by (6). is a very novel stochastic process, it defines a new reliability mechanism for some applications and has not been found in the current research literature. Our discussion is based on the following basic assumptions.()The primary process is a nonhomogeneous Poisson process with intensity function and cumulative intensity function satisfying as .()The adjunct processes , , are an i.i.d. family of counting processes of stationary increments and with arrival instants of . For any , we have .()The primary shock magnitudes ’s are nonnegative and i.i.d. random variables with finite variance .()The secondary shock magnitudes ’s are nonnegative and i.i.d. random variables, for all and .()The families , , , and are mutually independent.

3. Fundamental Properties

For describing the properties of , we use the notationsThen, the local cumulative shock process has a new representationBy assumptions (A2)–(A5), is a family of i.i.d. random vectors and independent of the primary process , and the components , , and in are mutually independent. We present several basic properties of based on the moment generating function.

Proposition 1. Under assumptions (A1)–(A5), given , the moment generating function of can be expressed aswhere is a random variable defined on with the distribution function .

Proof. By the definition of the moment generating function, we havewhere are i.i.d. random variables on with the common distribution function . The last step based on the fact thatare i.i.d. random variables.

From (11), we obtainNote that (11) and (14) contain the moment generating function and the moments of . Thus, we need to present the properties of .

Proposition 2. Under assumptions (A1)–(A5), given , the moment generating function of iswhere .

Proof. Letting , thenNote that has stationary increments (by assumption (A2)); we haveThe result has been proved.

For a given , we can compute first two moments (if exists) of and of and get the following conclusion.

Proposition 3. Under assumptions (A1)–(A5), for a given , the moment generating function of iswhere is given by (15), and the mean and variance are

Also define the distribution function of asWe have another result.

Proposition 4. Under assumptions (A1)–(A5), for a given , the logarithmically characteristic function of iswhere is a measure on .

Proof. Denote the characteristic function of by . It follows from (11) thatwhere is the characteristic function of . Conditioning on , we haveThe last equality is due to the definition of (22) and the fact thatwhere “” means limit equivalence.

It can be shown that the measure defined by (23) satisfiesand also guaranteesThus, is the Lévy measure and is the characteristic triplet of for given , respectively.

4. Central Limit Theorems

In this section, we discuss the limit theorems of , that is, the weak convergence of the regularized processas , where and are the appropriately selected centering and regularizing functions. To this end, we need to examine the convergence property of the corresponding characteristic function. The discussion of this section is based on the standard theories of infinitely divisible distributions and stable distributions (please see Cont and Tankov [21], Sato [22], Lin et al. [23], and Embrechts et al. [24] for the details).

4.1. Lévy-Khintchine Representation and Infinite Divisibility

By (21) and the Lévy-Khintchine representation of an infinitely divisible distribution, we know that is infinitely divisible for given , and its characteristic function is fully determined by the Lévy triplet defined in (22) and (23). Such a one-to-one correspondence also holds for their convergence results as . Thus, we use the notationto represent this relation and only discuss the weak convergence of triplet . Note that the infinite divisibility of is due to the nature of the Poisson primary shock process.

Then, by (29) and Proposition 4, the regularized process is also infinitely divisible. For appropriately selected and , we have the following theorem.

Theorem 5. Under assumptions (A1)–(A5), for a given , is infinitely divisible andwhereThus, the Lévy-Khintchin representation of is given by

Proof. The characteristic function of , denoted by , can be written asThen,where . The rest of proof is the same as the last part proof of Proposition 4.

Because the weak limit of the sequence of infinitely divisible random variables is also infinitely divisible, we can obtain the following weak convergence results of the regularized process based on Theorem 5 and Proposition 4.

Corollary 6. Under assumptions (A1)–(A5), there exists an infinitely divisible random variable with and constants and , such that the sufficient and necessary conditions for as arewhere means for all bounded and continuous functions .

4.2. Convergence to -Stable Distribution

Now, we present the conditions for to converge weakly to an -stable distribution. Let denote the common distribution function of secondary shocks . We consider the case that has a regular-tailed distribution with index and then have the following theorem.

Theorem 7. Under assumptions (A1)–(A5), if , where and is some slowly varying function, then and can be chosen asrespectively, such that as ,where and is the -stable distribution with index .

Proof. The selected implies that as . From Theorem 5, we know that ; according to Corollary 6 and Subsection 2.4 of Klüppelberg et al. [25], to prove this theorem, we need to show, as , (i)for each , , and for , satisfies the following homogeneous conditions: (ii);(iii);note that (i) holds if we show, for , where the convergence is in the sense of Corollary 6. From (32) and (23), we havewhere , . Due to the existence of the integral above, the finite variance of , and the regular-tailed distribution property of , we have . It follows from Lebesgue’s dominated convergence theorem thatFrom (39), we have . Using the properties of regular-tailed distributions and slowly varying functions, we obtainLet ; then, is a Lévy measure; it implies (43) and also satisfies the homogeneous conditions (42). Thus, (i) has been proved.
For showing (ii), we note from (46) that, as ,and using (34),Since , the integral above converges to zero, thus .
Finally, we show (iii). From (33) and (40), when ,Since all the integrands are bounded and continuous over their integration regions, the above integrals exist. Let , thus we have .
When ,Similarly, the above integral exists, and we also have by choosing .

We remark that the another representation of in Theorem 7 iswhere . Note thatwhich implieswhere is a slowly varying function.

The theorem below gives the conditions for the regularized process to converge to normal distribution. This result is natural.

Theorem 8. Under assumptions (A1)–(A5), if exists and is finite, we can choosesuch that, as , where represents standard normal distribution.

5. Summary and Discussion

The local cumulative shock model defined by (6), (7), and (8) is a generalization of the classical extreme shock models. There are two types of shocks in this model called primary shocks and secondary shocks, respectively. The primary shocks are generated by a nonhomogeneous Poisson process and each of them causes a series of secondary shocks. Therefore, the shock process has a cluster structure and can be considered as the superposition of two streams of shock processes. In this model, the local cumulative shock process plays the core role in determining the lifetime behavior of the system.

This paper discusses the limit properties of . Based on the theories of infinite divisibility and stable distributions, we establish the central limit theorems of , and find the conditions under which the regularized process of converges to infinitely divisible, -stable or normal distributions. Our work provides a foundation for the next work of studying lifetime property and failure behavior of the local cumulative shock model.

is a very novel stochastic process and has not been found in the available literature. The new failure mechanism defined by is more reasonable than the classical models to explain some reliability phenomenons such as the bankruptcy behavior in insurance and financial industry, where the fateful risk is not necessarily long-history dependent but more like short-term determined. For such a situation, the classical models may not provide realistic risk measures. There are two main factors that make our model more reasonable in describing the risk behavior for the operations of an insurance or a financial firm. The first one is that the cluster structure is based on the event sequence of purchase and claims of each insurance policy. The second is that the local cumulative shock is more realistic and flexible in modeling the bankrupt condition for modern insurance services.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgment

J. M. Bai’s work is supported by the National Natural Science Foundation of China Grant nos. 71171103 and 71072070. C. Yuan’s work is supported by Humanities & Social Sciences Project of the Ministry of Education of China Grant no. 13YJC630215 and partly by the National Natural Science Foundation of Guangdong Province Grant no. S2012040007369. The authors would like to thank the editor and the anonymous referees for their valuable comments and suggestions, and also they are grateful to Dr. Zhe George Zhang in the Department of Decision Sciences, Western Washington University, and Professor Zehui Li of the School of Mathematics and Statistics, Lanzhou University, for their useful suggestions and modifications.