References
- Abbasnejad, M., N. R. Arghami, and M. Tavakoli. 2012. A goodness of fit test for exponentiality based on Lin Wong information. Journal of the Iranian Statistical Society 11 (2):191–202.
- Alizadeh Noughabi, H. 2010. A new estimator of entropy and its application in testing normality. Journal of Statistical Computation and Simulation 80 (10):1151–62.
- Bitaraf, M., M. Rezaei, and F. Yousefzadeh. 2017. Test for normality based on two new estimators of entropy. Journal of Statistical Computation and Simulation 87 (2):280–94.
- Cha, S. H. 2007. Comprehensive survey on distance/similarity measures between probability density functions. International Journal of Mathematical Models and Methods in Applied Sciences 1 (4):300–7.
- Correa, J. C. 1995. A new estimator of entropy. Communications in Statistics-Theory and Methods 24 (10):2439–49.
- Deza, M. M., and E. Deza. 2009. Encyclopedia of distance. Heidelberg, Germany: Springer Berlin Heidelberg.
- Ebrahimi, N., K. Pflughoeft, and E. S. Soofi. 1994. Two measures of sample entropy. Statistics & Probability Letters 20 (3):225–34.
- Esteban, M., M. Castellanos, D. Morales, and I. Vajda. 2001. Monte carlo comparison of four normality tests using different entropy estimates. Communication in Statistics-Simulation and Computation 30 (4):761–85.
- Joarder, A., H. Krishna, and D. Kundu. 2011. Inferences on weibull parameters with conventional Type-I censoring. Computational Statistics and Data Analysis 55 (1):1–11.
- Koziol, J. A., and D. P. Byar. 1975. Percentage points of the asymptotic distributions of one and two sample KS statistics for truncated or censored data. Technometrics 17 (4):507–10.
- Kullback, S., and R. A. Leibler. 1951. On information and sufficiency. The Annals of Mathematical Statistics 22 (1):79–86.
- Lin, J., and S. K. M. Wong. 1990. A new directed divergence measure and its characterization. International Journal of General System 17 (1):73–81.
- Lin, J. 1991. Divergence measures based on the shannon entropy. IEEE Transactions on Information Theory 37 (1):145–51.
- Nelson, W. 1982. Applied life data analysis. NewYork: JohnWiley and Sons.
- Pakyari, R., and N. Balakrishnan. 2013. Testing exponentiality based on Type-I censored. Journal of Statistical Computation and Simulation 83 (12):2369–78.
- Pakyari, R., and K. R. Nia. 2017. Testing goodness-of-fit for some lifetime distributions with conventional Type-I censoring. Communications in Statistics-Simulation and Computation 46 (4):2998–3009.
- Park, S., and M. Shin. 2014. Kullback–leibler information of a censored variable and its applications. Statistics 48 (4):756–65.
- Parzen, E. 1979. Nonparametric statistical data modeling. Journal of the American Statistical Association 74 (365):105–21.
- Persson, T., and H. Rootzen. 1977. Simple and highly efficient estimators for a Type-I censored normal sample. Biometrika 64 (1):123–8.
- Pettitt, A. N., and M. A. Stephens. 1976. Modified cramer-von mises statistics for censored data. Biometrika 63 (2):291–8.
- Shannon, C. E. 1948. A mathematical theory of communication. ACM Sigmobile Mobile Computing and Communications Review 5 (1):3–55.
- Shioya, H., and T. Da-Te. 1995. A generalization of lin divergence and the derivation of a new information divergence. Electronics and Communications in Japan (Part III: Fundamental Electronic Science) 78 (7):34–40.
- Silverman, B. 1986. Density estimation. London, UK: Chapman & Hall/CRC.
- Topsoa, F. 1999. Some inqualities for information divergence and related measures of discrimination. Research Report Collection 2 (1):73–83.
- Vasicek, O. 1976. A test for normality based on sample entropy. Journal of Research Statistical Society 38 (Serial. B):54–9.
- Wasserman, L. 2010. All of nonparametric statistics. New York: Springer Texts in Statistics. (No. 519. 52 W3).
- Wieczorkowski, R., and P. Grzegorzewski. 1999. Entropy estimators - Improvements and comparisons. Communication Statistics Computing and Simulation 28 (2):541–67.
- Zamanzade, E., and N. R. Arghami. 2011. Goodness of fit test based on correcting moments of modified entropy estimator. Journal of Statistical Computation and Simulation 81 (12):2077–93.
- Zamanzade, E., and N. R. Arghami. 2012. Testing normality based on new entropy estimators. Journal of Statistical Computation and Simulation 82 (11):1701–13.