Journal of Applied Mathematics

Journal of Applied Mathematics / 2013 / Article

Research Article | Open Access

Volume 2013 |Article ID 902972 | 10 pages | https://doi.org/10.1155/2013/902972

A New Improved Parsimonious Multivariate Markov Chain Model

Academic Editor: Marco H. Terra
Received12 Nov 2012
Revised17 Dec 2012
Accepted02 Jan 2013
Published04 Feb 2013

Abstract

We present a new improved parsimonious multivariate Markov chain model. Moreover, we find a new convergence condition with a new variability to improve the prediction accuracy and minimize the scale of the convergence condition. Numerical experiments illustrate that the new improved parsimonious multivariate Markov chain model with the new convergence condition of the new variability performs better than the improved parsimonious multivariate Markov chain model in prediction.

1. Introduction

The probability distribution of Markov chains plays an important role in a wide range of applications such as telecommunication systems, manufacturing systems, and inventory systems, see, for instance, [1] and the references therein. In recent years, the prediction of multivariate Markov chain models [2] has become more and more useful in many real-world applications: sales demand predictions [2, 3], DNA sequencing [4], and credit and financial data modeling [5]. The major merit of the multivariate Markov chain model is to detect the relations among the sequences and to predict more precisely.

Different models for multiple categorical data sequences are introduced in the following part. A multivariate Markov chain model has been presented in [2]. Ching et al. constructed a new matrix by means of the transition probability matrices among different sequences. To improve prediction accuracy, Ching et al. incorporated positive and negative parts in an improved parsimonious multivariate Markov chain model [5]. Miao and Hambly presented recursive formulas for the default probability distribution which is feasible for computation in this simple version [6]. A more advanced model, namely, higher-order multivariate Markov chain model has been exhibited in [7]. To reduce the number of parameters of the model, a parsimonious higher-order multivariate Markov chain model has been proposed in [8], where the number of parameters is . Certainly, there are many other papers contributing to the multivariate Markov chain models, for example, [1, 9, 10].

With the development of science technologies with their applications, the number of data in sequences become larger, and the results need to be more precise. It is inevitable that a large categorical data sequence group will cause high computational costs, especially using the convergence condition as in [5]. In this paper, a new improved parsimonious multivariate Markov chain model and a new convergence condition with a new variability are presented to enhance the precision of the prediction and save the computational costs.

The rest of the paper is organized as follows. In Section 2, we briefly review several multivariate Markov chain models. In Section 3, a new improved parsimonious multivariate Markov chain model and a new convergence condition with a new variability are presented. Section 4 gives the estimation methods for the parameters of the new improved parsimonious multivariate Markov chain model with different convergence conditions. Numerical experiments with three examples are presented to demonstrate the effectiveness of our proposed model with the new convergence condition of the new variability in Section 5. Finally, concluding remarks are given in Section 6.

2. Review of the Multivariate Markov Chain Models

In this section, we briefly introduce several multivariate Markov chain models, for example, the Markov chain model [3], the multivariate Markov chain model [2], and the improved parsimonious multivariate Markov chain model [5].

2.1. The Markov Chain Model

First, we introduce some definitions of the Markov chain [2, 11]. Let the state set of the categorical data sequences be . The discrete-time Markov chain with states satisfies the following relations: where , . The conditional probability is called the one-step transition probability of the Markov chain. The transition probability is and the transition probability matrix is . The Markov chain model can be represented as follows: where is the initial probability distribution, and is the state probability distribution at time .

2.2. The Multivariate Markov Chain Model

Suppose the number of categorical data sequences . The multivariate Markov chain model [2] is represented as follows: where Here, is the initial probability distribution of the th sequence, is the state probability distribution of the th sequence at time , and is the state probability distribution of the th sequence at time . Here, is the one-step transition probability from the state in the th sequence at time to the state in the th sequence at time . In matrix form, we have Here, and can be obtained from the categorical data sequences and the corresponding linear programming, for details, refer to [2].

2.3. The Improved Parsimonious Multivariate Markov Chain Model

With the same notations as introduced in Section 2.2, we introduce the improved parsimonious multivariate Markov chain model originating from multivariate Markov chain model. Consider where the factor is a constant for normalization. Here, is the vector of all ones and . In matrix form, the improved parsimonious multivariate Markov chain model can be represented as where and are the positive and negative parts of the transition probability matrices in (9). The above model has two directions to approach the steady solution .

Let , . If for and there exist satisfying , then we note that .

Lemma 1 (see [12]). Let be a nonnegative and irreducible matrix, a complex matrix, and an eigenvalue of . If , then .

3. A New Improved Parsimonious Multivariate Markov Chain Model

In this section, we propose a new improved parsimonious multivariate Markov chain model and a new convergence condition with a new variability.

In the new model, the state probability distribution of the th sequence at time depends on the state probability distribution of all the sequences at time . Let the number of categorical data sequences be , the number of states in every sequences, , then the new improved parsimonious multivariate Markov chain model can be represented as follows: where is the initial probability distributions of the th sequence and Here, is the state probability distribution of the th sequence at time , is the one-step transition probability matrix from the state in the th sequence at time to the state in the th sequence at time . Here, is the state probability distribution of the th sequence at time . Let then the new improved parsimonious multivariate Markov chain model in matrix form is which also can be represented as where and are, respectively, the positive and negative parts of the transition probability matrices in (14) where . Each column sum of is equal to one.

From (14), after times iterations, it has

If , the iteration of the new improved parsimonious multivariate Markov chain model is convergent. For finding a more simple and efficient convergence condition from the point of view of properties of special matrices, we get the following theorem.

Theorem 2. In the new improved parsimonious multivariate Markov chain model, if where and , then the iteration of the new improved parsimonious multivariate Markov chain model is convergent.

Proof. Because , we obtain For , it has Then, we obtain From Lemma 1, it has In the new model, . It suffices to prove that . Then, we obtain The new improved parsimonious multivariate Markov chain model is convergent.

4. Estimation of the Parameters of the New Improved Parsimonious Multivariate Markov Chain Model

In this section, we estimate the parameters of the new improved parsimonious multivariate Markov chain model in the new convergence condition with the new variability which has been proved in Theorem 2. The transition probability matrices are estimated at first. If the data sequences are given and the state set is , is the frequency from the state in the th sequence at time to the state in the th sequence at time with , then the transition frequency matrix can be constructed as follows: Here, can be obtained by normalizing the frequency transition probability matrix as where

Subsequently, the way of estimating the parameter is introduced. is a joint state probability distribution of the new improved parsimonious multivariate Markov chain model at and can be represented as which satisfies where , have been denoted in Section 3 satisfying and . Based on the idea of the convergence condition in [5], the iteration matrix of the new improved parsimonious multivariate Markov chain model satisfies By imposing an upper bound , the convergence condition of the new improved parsimonious multivariate Markov chain model is Then the new improved parsimonious multivariate Markov chain model in this convergence condition can be represented as a set of linear programming problems subject to where Here, covers all of the rows of each component taking one of the two possible values, and , particularly, in th column taking and . Then has rows. This convergence condition is viable only when the sequence group is not large.

To speed up the convergence of the new improved parsimonious multivariate Markov chain model and minimize the scale of the convergence condition, we apply a new convergence condition with and for the new model. Then, the form of the new improved parsimonious multivariate Markov chain model can be represented as subject to: where is the th entry of the vector. Certainly, the optimization problem can also be represented as a linear programming problem: subject to

5. Numerical Experiments

In this section, numerical experiments with three examples of different improved parsimonious multivariate Markov chain models with different convergence conditions are reported.

Noting that the new improved parsimonious multivariate Markov chain model with the original convergence condition [5] is “IPM1,” (especially, the new improved parsimonious multivariate Markov chain model is an improved multivariate Markov chain model when we choose ), the new improved parsimonious multivariate Markov chain model with the new convergence condition of the new variability is “IPM2,” the convergence factor of the original convergence condition is “,” and the variabilities of the new convergence condition are “” and “.” The stopping criterion can be found in Matlab order of . We add a notation “” at the back of data in Tables 1, 2, 3, and 4 when the stopping criterion is satisied but the accuracy is not reached.



IPM1 when 31.1845* 31.0831* 29.8349* 29.8349* 31.0906* 31.0853*
IPM1 when 27.0472* 26.9434* 26.8745* 26.8745* 26.9486* 26.8462*
IPM1 when 22.8309* 22.7298* 22.7296* 22.7302* 22.7321* 22.7310*
IPM1 when 18.6215* 18.5266* 18.5256* 18.5256* 18.5282* 18.5211*
IPM1 when 14.3298* 14.2594* 14.2614* 14.2614* 14.2586* 14.2705*
IPM1 when 10.4531* 10.2961* 10.3231* 10.2962* 10.2795* 10.2971*
IPM1 when 9.9017 9.8994 9.8872 9.8879 9.8896 9.8552
IPM1 when 10.1173 10.1440 10.1255 10.1271 10.1283 10.0536
IPM1 when 10.3910 10.4131 10.3945 10.3886 10.3803 10.2726
IPM1 when 10.4059 10.4268 10.4059 10.4059 10.4059 10.2590



IPM2 when 8.9588 9.1114 9.0964 9.0894 9.0860 9.0778
IPM2 when 8.8126 8.8788 9.0326 9.1028 9.0904 9.0832
IPM2 when 8.9746 8.8142 8.8544 8.9447 9.0892 9.0974
IPM2 when 8.6957 8.9355 8.8167 8.8449 8.9054 9.0014
IPM2 when 8.9006 8.6790 8.7050 8.8189 8.8410 8.8849
IPM2 when 9.4994 8.7730 8.6838 8.7168 8.8209 8.8393
IPM2 when 9.1024* 9.1909 8.7041 8.6911 8.7275 8.8228
IPM2 when 9.3544* 9.5649* 9.0116 8.6994 8.6993 8.7371
IPM2 when 12.3474* 9.0286* 9.3882 8.8994 8.7029 8.7076
IPM2 when 13.5788* 9.4195* 9.4562* 9.1952 8.8252 8.7046



IPM1 when 1624.2* 1628.5* 1631.7* 1634.1* 1636.1* 1637.7*
IPM1 when 1354.0* 1355.8* 1359.3* 1361.1* 1363.8* 1365.1*
IPM1 when 1083.7* 1086.4* 1088.4* 1089.9* 1091.1* 1092.1*
IPM1 when 809.2156* 811.0515* 812.4306* 813.5030* 814.3598* 815.0606*
IPM1 when 543.7497* 545.0841* 546.0720* 546.8347* 547.4426* 547.9362*
IPM1 when 366.9729 366.9684 366.9580 366.9577 366.9533 366.9589
IPM1 when 367.3149 367.2101 367.1579 367.1367 367.1448 367.1498
IPM1 when 367.5954 367.5916 367.5874 367.5769 367.5771 367.5766
IPM1 when 368.2437 368.2422 368.2464 368.2530 368.2550 368.2639
IPM1 when 369.1255 369.1145 369.1161 369.1119 369.1088 369.1062



IPM2 when 348.9387 348.8616 348.6692 348.4583 346.8533 347.9374
IPM2 when 341.3333 348.4608 348.5667 348.6407 348.6676 348.6457
IPM2 when 333.1122 334.4701 339.0412 343.5102 348.3476 348.3801
IPM2 when 334.2837 331.1873 333.7044 334.5513 337.5047 340.6662
IPM2 when 333.1028* 332.8355 330.7546 332.4308 333.8498 334.6122
IPM2 when 327.5181* 333.2228* 335.7616 331.0009 331.4727 332.7485
IPM2 when 481.5991* 327.9535* 333.2418* 335.7036 334.4605 331.1581
IPM2 when 550.2992* 457.7242* 329.7488* 333.3142* 335.7388 333.6250
IPM2 when 618.8261* 521.8871* 452.3191* 331.4777* 333.4008* 335.0657
IPM2 when 527.6297* 579.7398* 501.1963* 442.6577* 331.6313* 333.5236*

5.1. Example 1

There are three categorical data sequences [13]: IPM1 with [5] can be represented as IPM2 with is In IPM2 with , it has

5.2. Example 2

Let the three categorical data sequences be

Suppose that is the prediction probability at time and is the fact value at time where , , is the fact state at time in the th categorical data sequence. “nA” is the number of the categorical data in one sequence and “pe” is the prediction error of the models which can be estimated by the equation:

In Table 1, the prediction errors of the new improved parsimonious multivariate Markov chain model when is better than the prediction errors of other values of . Table 1 illustrates the efficiency of the new improved parsimonious multivariate Markov chain model when in the original convergence condition [5].

In Table 2, numerical experiments on the prediction errors of the new improved parsimonious multivariate Markov chain model of the new convergence condition with the new variability “” are reported. The best performance of the prediction errors of the new model with the new convergence condition of the new variability is 8.6790 when we choose and . In different cases of , the best prediction errors of the new model with the new convergence condition of the new variability are in the diagonal line of the result matrix between and .

For comparing the performances of the new improved parsimonious multivariate Markov chain model with the new convergence condition more clearly, we present Figure 1.

To compare the performances of the new improve parsimonious multivariate Markov chain model in different convergence conditions, we show Figure 2 with the data of Tables 1 and 2. Obviously, the results of the new model in the new convergence condition with the new variability are much better than those of the new model in the original convergence condition.

5.3. An Application to Sales Demand Predictions

In this part, the sales demand sequences are presented to show the effectiveness of the new improved parsimonious multivariate Markov chain model of the new convergence condition with the new variability. Since the requirement of the market fluctuates heavily, the production planning and the inventory control directly affect the estate cost. Thus, studying the interplay between the storage space requirement and the overall growing sales demand is a pressing issue for the company. Suppose that the products are classified into six possible states , for example, no sales volume, very low sales volume, low sales volume, standard sales volume, fast sales volume, and very fast sales volume. The data of customer’s sales demand of five important products can be found in [3].

By computing the proportion of the occurrence of each state in the sequences, the initial probability distributions of the five categorical data sequences are The transition probability matrix can be obtained after normalizing the transition frequency matrix. By solving the linear programming problem corresponding to the new improved parsimonious multivariate Markov chain model with the new convergence condition where and , is obtained and the model is presented as follows:

Here, the prediction error equation are represented as

In Table 3, we present the prediction errors of the new improved parsimonious multivariate Markov chain model with the original convergence condition [5]. From the data of Table 3, we get Figure 3. In Figure 3, it is obvious that the new improved parsimonious multivariate Markov chain model of the original convergence condition performs the best when we choose . The smaller we choose, the better error prediction we get.

In Table 4, the best performance of the prediction errors of the new improved parsimonious multivariate model with the new convergence condition in sales demand prediction is 331.1873 when and . In different cases of , the best prediction errors are in the diagonal line of the result matrix between and .

For comparing the performances of the new improved parsimonious multivariate model with the new convergence condition of the new variability more clearly, we present Figure 4 where the data are extracted from Table 4. As the value of increases, the parameter of the best performances of the prediction results increases. The best performances of the IPM2 are almost the same in different cases of .

With the data of Tables 3 and 4, we get Figure 5. It illustrates the benefits of the new parsimonious multivariate Markov chain model of the new convergence condition with the new variability in prediction accuracy.

6. Conclusion

In this paper, we present a new improved parsimonious multivariate Markov chain model and a new convergence condition with a new variability, which can enhance the the prediction accuracy of the models and save the computational estate. Numerical experiments with three examples illustrate that the new improved parsimonious multivariate Markov chain model of the new convergence condition with the new variability performs better than the improved parsimonious multivariate Markov chain model with the original convergence condition. Certainly, our new model can also be applied into credit risk and other research areas.

Acknowledgments

This research is supported by the Chinese Universities Specialized Research Fund for the Doctoral Program (20110185110020) and Sichuan Province Sci. & Tech. Research Project (12ZC1802).

References

  1. W.-K. Ching and M. K. Ng, Markov Chains: Models, Algorithms and Applications, International Series on Operations Research and Management Science, Springer, New York, NY, USA, 2006. View at: MathSciNet
  2. W.-K. Ching, E. S. Fung, and M. K. Ng, “A multivariate Markov chain model for categorical data sequences and its applications in demand predictions,” IMA Journal of Management Mathematics, vol. 13, no. 3, pp. 187–199, 2002. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  3. W. K. Ching, E. S. Fung, and M. K. Ng, “Higher-order Markov chain models for categorical data sequences,” Naval Research Logistics, vol. 51, no. 4, pp. 557–574, 2004. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  4. W. K. Ching, M. M. Ng, E. S. Fung, and T. Akutsu, “On construction of stochastic genetic networks based on gene expression sequences,” International Journal of Neural Systems, vol. 15, no. 4, pp. 297–310, 2005. View at: Publisher Site | Google Scholar
  5. W. Ching, T. Siu, and L. Li, “An improved parsimonious multivariate Markov chain model for credit risk,” Journal of Credit Risk, vol. 5, pp. 1–25, 2009. View at: Google Scholar
  6. D. W. C. Miao and B. M. Hambly, “Recursive formulas for the default probability distribution of a heterogeneous group of defauleable entities,” 2012. View at: Google Scholar
  7. W.-K. Ching, M. K. Ng, and E. S. Fung, “Higher-order multivariate Markov chains and their applications,” Linear Algebra and Its Applications, vol. 428, no. 2-3, pp. 492–507, 2008. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  8. C. Wang, T. Z. Huang, and C. Wen, “A simplified higher-order multivariate Markov chains model,” submitted. View at: Publisher Site | Google Scholar
  9. C. Wang and T. Z. Huang, “Improved multivariate Markov chain model with the new convergence condition,” submitted. View at: Publisher Site | Google Scholar
  10. T.-K. Siu, W.-K. Ching, E. S. Fung, and M. K. Ng, “On a multivariate Markov chain model for credit risk measurement,” Quantitative Finance, vol. 5, no. 6, pp. 543–556, 2005. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  11. R. A. Horn and C. R. Johnson, Matrix Analysis, Cambridge University Press, Cambridge, UK, 1985. View at: MathSciNet
  12. Z.-Y. You and C.-L. Wang, “A concept of nonlinear block diagonal dominance,” Journal of Computational and Applied Mathematics, vol. 83, no. 1, pp. 1–10, 1997. View at: Google Scholar | Zentralblatt MATH | MathSciNet
  13. C. Wang, T. Z. Huang, and W. K. Ching, “On simplified parsimonious models for higher-order multivariate Markov chains,” submitted. View at: Google Scholar

Copyright © 2013 Chao Wang and Ting-Zhu Huang. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

912 Views | 587 Downloads | 2 Citations
 PDF  Download Citation  Citation
 Download other formatsMore
 Order printed copiesOrder

We are committed to sharing findings related to COVID-19 as quickly and safely as possible. Any author submitting a COVID-19 paper should notify us at help@hindawi.com to ensure their research is fast-tracked and made available on a preprint server as soon as possible. We will be providing unlimited waivers of publication charges for accepted articles related to COVID-19.