Item request has been placed! ×
Item request cannot be made. ×
loading  Processing Request

Minimum K-phi-divergence estimator

Item request has been placed! ×
Item request cannot be made. ×
loading   Processing Request
  • نوع التسجيلة:
    Electronic Resource
  • الدخول الالكتروني :
    https://hdl.handle.net/20.500.14352/50248
    http://www.sciencedirect.com/science/article/pii/S0893965904900766
    http://www.sciencedirect.com
    DGI BFM-2000-0800
  • معلومة اضافية
    • Publisher Information:
      Pergamon-Elsevier Science 2023-06-20T09:43:23Z 2023-06-20T09:43:23Z 2004-04
    • Added Details:
      Pérez, T.
      Pardo Llorente, Julio Ángel
    • نبذة مختصرة :
      In the present work, the problem of estimating parameters of statistical models for categorical data is analyzed. The minimum K-phi-divergence estimator is obtained minimizing the K-phi-divergence measure between the theoretical and the empirical probability vectors. Its asymptotic properties are obtained. Rom a simulation study, the conclusion is that our estimator emerges as an attractive alternative to the classical maximum likelihood estimator.
      Depto. de Estadística e Investigación Operativa
      Fac. de Ciencias Matemáticas
      TRUE
      pub
    • الموضوع:
    • Note:
      application/pdf
      0893-9659
      English
    • Other Numbers:
      ESRCM oai:docta.ucm.es:20.500.14352/50248
      S. Kullback and A. Leibler, On information and sufficiency, Annals of Mathematical Statistics 22, 76-86, (1951). N. Cressie and T.P~.C. Read, Multinomial goodness-of-fit tests, Journal of the Royal Statistical Society, Series B 46, 440-464, (1984). D. Morales, L. Pardo and I. Vajda, Asymptotic divergence of estimates of discrete distributions, Journal of Statistical Planning and Inference 48, 347-369, (1995). S.M. Ali and S.D. Silvey, A general class of coefficients of divergence of one distribution from another, Journal of the Royal Statistical Society, Series B 26, 131-142, (1966). I. Csisz£r, Eine Informationstheoretische Ungleichung und ihre Anwendung auf den Beweis der Ergodizit~it on Markhoffschen Ketten., Publications of the Mathematical Institute of Hungarian Academy of Sciences, Series A 8, 85-108, (1963). M.C. Pardo, Asymptotic behaviour of an estimator based on Rao's divergence, Kybernetika 33 (5), 489-504, (1997). M.C. Pardo, A comparison of some estimators of the mixture proportion of mixed normal distributions, Journal of Computational and Applied Mathematics 84, 207-217, (1997). J. Burbea and C.R. Rao, On the convexity of some divergence measures based on entropy functions, IEEE Transactions on Information Theory 28, 489-495, (1982). N.W. Birch, A new proof of the Pearson-Fisher theorem, Annals of Mathematical Statistics 35, 817-824, (1964).
      0893-9659
      10.1016/s0893-9659(04)00040-0
      1413950402
    • Contributing Source:
      REPOSITORIO E-PRINTS UNIVERSIDAD COMPLU
      From OAIster®, provided by the OCLC Cooperative.
    • الرقم المعرف:
      edsoai.on1413950402
HoldingsOnline