Minimum message length analysis of multiple short time series

https://doi.org/10.1016/j.spl.2015.09.021Get rights and content

Abstract

This paper applies the Bayesian minimum message length principle to the multiple short time series problem, yielding satisfactory estimates for all model parameters as well as a test for autocorrelation. Connections with the method of conditional likelihood are also discussed.

Introduction

Consider data Y=(y1,,ym)Rm×n comprised of m sequences yi=(yi,1,,yi,n)Rn generated by the following stationary first order Gaussian autoregressive model: yi,j=μi+εi,j,εi,j=ρεi,j1+vi,j, where (i=1,,m;j=1,,n), μ=(μ1,,μm)Rm are the sequence means, ρ(1,1) is a common autoregressive parameter and vi,j denotes the innovations which are independently and identically distributed as N(0,τ). The starting point of this paper is to make inferences about the parameters θ=(μ,ρ,τ)Rm+2 given data sampled from the model  (1)–(2). The sequences are considered exchangeable, in the sense that inferences made about the model parameters should be invariant under the interchange of any pair of sequences in the matrix Y. This model appears frequently in epidemiological and medical studies in which several measurements have been made over time on a large number of people. In this case, the autocorrelation parameter ρ is of particular interest, as it represents how well the physical quantity “tracks” over time.

Making inferences about ρ in this setting is complicated by the fact that the number of parameters grows with the number of sequences m and a straightforward application of the maximum likelihood principle leads to inconsistent estimates of both τ and ρ. A likelihood-based solution to the problem of estimating ρ in the model  (1)–(2) using the method of approximate conditional likelihood was presented in  Cruddas et al. (1989) and shown to yield significant improvements over the standard maximum likelihood estimates. Two frequentist test procedures for the presence of autocorrelation are also discussed in  Cox and Solomon (1988).

A solution within the Bayesian framework of inference would be of great value. Unfortunately, with the choice of sensible priors that reflect the invariance properties required of the problem, the usual method of analysing the posterior distribution formed from the product of the prior distribution and likelihood is unsatisfactory. The posterior distribution does not concentrate probability mass around the true parameter values even as the number of sequences m, and parameter estimation based on this posterior is subsequently inconsistent. This paper demonstrates that estimation based on the alternative information-theoretic Bayesian principle of minimum message length (Wallace, 2005) leads to satisfactory estimates of all parameters θ as well as providing a simple basis for testing for autocorrelation.

This paper has three aims: (1) to produce satisfactory point estimates for all parameters of the first order Gaussian autoregressive model, (2) to produce a suitable test for autocorrelation, and (3) demonstrate the resolution of a difficult estimation problem using the minimum message length principle.

Section snippets

Minimum message length

The minimum message length (MML) principle (Wallace, 2005, Wallace and Boulton, 1968, Wallace and Freeman, 1987) is a Bayesian principle for inductive inference based on information theory. The essential idea behind the minimum message length principle is that compressing data is equivalent to learning structure in the data. The key measure of the quality of fit of a model to data is the length of the data after it has been compressed by the model under consideration. As the compressed data

Wallace–Freeman estimates

Inference using the Wallace–Freeman estimator requires specifying a likelihood function, the corresponding Fisher information matrix and prior densities over all parameters. In the multiple short time series setting specified by (1)–(2), the negative log-likelihood function for the parameters θ=(μ,ρ,τ)Rm+2 is logp(Y|θ)=mn2log2πτm2log(1ρ2)+12τi=1mTi(μi,ρ), where Ti(μi,ρ)=j=1n(yijμi)2+ρ2j=2n1(yijμi)22ρj=2n(yijμi)(yij1μi). The determinant of the Fisher information matrix for θ is |J

Estimation of ρ

The Wallace–Freeman estimates ρˆ of ρ were compared against the approximate conditional likelihood estimates ρˆλ derived in  Cruddas et al. (1989) and the regular maximum likelihood estimates. Due to the translation invariance of the estimates for μ given by (12), and the fact that the estimates for ρ and τ are based on the residuals (yijμˆi) for both the Wallace–Freeman procedure and approximate conditional likelihood, the particular choice of μi will have no effect on the behaviour of the

Comparison with approximate conditional likelihood

The approximate conditional likelihood procedure can be shown to be equivalent to a restricted Wallace–Freeman estimator. In particular, the approximate conditional likelihood estimate for a parameter of interest ψ orthogonal to nuisance parameters λ can be written as ψˆ=argminψ{I87(y,ψ,λˆML(ψ))} with prior densities of (ψ,λ)πλ(λ)1,πψ(ψ)Jψ(ψ)(1/2), where λˆML(ψ) is the maximum likelihood estimate of the orthogonalised parameters given ψ. Although this estimate is permissible in the sense that

References (15)

  • J.O. Berger et al.

    Noninformative priors and Bayesian testing for the AR(1) model

    Econ. Theory

    (1994)
  • J.H. Conway et al.

    Sphere Packing, Lattices and Groups

    (1998)
  • D.R. Cox et al.

    On testing for serial correlation in large numbers of small samples

    Biometrika

    (1988)
  • A.M. Cruddas et al.

    A time series illustration of approximate conditional likelihood

    Biometrika

    (1989)
  • D.L. Dowe et al.

    Resolving the Neyman–Scott problem by minimum message length

  • P. Druilhet et al.

    Invariant HPD credible sets and MAP estimators

    Bayesian Anal.

    (2007)
  • D.F. Schmidt

    A new message length formula for parameter estimation and model selection

There are more references available in the full text version of this article.
View full text