DenseHMM: Learning Hidden Markov Models by Learning Dense Representations

December 17, 2020 ยท Entered Twilight ยท ๐Ÿ› International Conference on Pattern Recognition Applications and Methods

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 5.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: LICENSE, README.md, code_dense_hmm, dense_hmm.yml

Authors Joachim Sicking, Maximilian Pintz, Maram Akila, Tim Wirtz arXiv ID 2012.09783 Category cs.LG: Machine Learning Cross-listed stat.ML Citations 3 Venue International Conference on Pattern Recognition Applications and Methods Repository https://github.com/fraunhofer-iais/dense-hmm โญ 2 Last Checked 1 month ago
Abstract
We propose DenseHMM - a modification of Hidden Markov Models (HMMs) that allows to learn dense representations of both the hidden states and the observables. Compared to the standard HMM, transition probabilities are not atomic but composed of these representations via kernelization. Our approach enables constraint-free and gradient-based optimization. We propose two optimization schemes that make use of this: a modification of the Baum-Welch algorithm and a direct co-occurrence optimization. The latter one is highly scalable and comes empirically without loss of performance compared to standard HMMs. We show that the non-linearity of the kernelization is crucial for the expressiveness of the representations. The properties of the DenseHMM like learned co-occurrences and log-likelihoods are studied empirically on synthetic and biomedical datasets.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning