Authors
Hyoungjin Lim, Gwonsoo Che, Wonyeol Lee, Hongseok Yang
Publication date
2020/2
Conference
AAAI 2020
Pages
4828-4835
Description
We present an algorithm for marginalising changepoints in time-series models that assume a fixed number of unknown changepoints. Our algorithm is differentiable with respect to its inputs, which are the values of latent random variables other than changepoints. Also, it runs in time O (mn) where n is the number of time steps and m the number of changepoints, an improvement over a naive marginalisation method with O (n m) time complexity. We derive the algorithm by identifying quantities related to this marginalisation problem, showing that these quantities satisfy recursive relationships, and transforming the relationships to an algorithm via dynamic programming. Since our algorithm is differentiable, it can be applied to convert a model non-differentiable due to changepoints to a differentiable one, so that the resulting models can be analysed using gradient-based inference or learning techniques. We empirically show the effectiveness of our algorithm in this application by tackling the posterior inference problem on synthetic and real-world data.
Total citations
Scholar articles
H Lim, G Che, W Lee, H Yang - Proceedings of the AAAI Conference on Artificial …, 2020