Recursive Maximum Likelihood Algorithm for Dependent Observations

Schwartz B, Gannot S, Habets E, Noam Y (2019)


Publication Type: Journal article

Publication year: 2019

Journal

Book Volume: 67

Pages Range: 1366-1381

Article Number: 8598896

Journal Issue: 5

DOI: 10.1109/TSP.2018.2889945

Abstract

A recursive maximum-likelihood algorithm (RML) is proposed that can be used when both the observations and the hidden data have continuous values and are statistically dependent between different time samples. The algorithm recursively approximates the probability density functions of the observed and hidden data by analytically computing the integrals with respect to the state variables, where the parameters are updated using gradient steps. A full convergence proof is given, based on the ordinary differential equation approach, which shows that the algorithm converges to a local minimum of the Kullback-Leibler divergence between the true and the estimated parametric probability density functions-a result that is useful even for a miss-specified parametric model. Compared to other RML algorithms proposed in the literature, this contribution extends the state-space model and provides a theoretical analysis in a nontrivial statistical model that was not analyzed so far. We further extend the RML analysis to constrained parameter estimation problems. Two examples, including nonlinear state-space models, are given to highlight this contribution.

Authors with CRIS profile

Involved external institutions

How to cite

APA:

Schwartz, B., Gannot, S., Habets, E., & Noam, Y. (2019). Recursive Maximum Likelihood Algorithm for Dependent Observations. IEEE Transactions on Signal Processing, 67(5), 1366-1381. https://dx.doi.org/10.1109/TSP.2018.2889945

MLA:

Schwartz, Boaz, et al. "Recursive Maximum Likelihood Algorithm for Dependent Observations." IEEE Transactions on Signal Processing 67.5 (2019): 1366-1381.

BibTeX: Download