Hidden 1-Counter Markov Models and How to Learn Them
Loading...

Date
2022
Journal Title
Journal ISSN
Volume Title
Publisher
International Joint Conferences on Artificial Intelligence
Open Access Color
OpenAIRE Downloads
OpenAIRE Views
Abstract
We introduce hidden 1-counter Markov models (H1MMs) as an attractive sweet spot between standard hidden Markov models (HMMs) and probabilistic context-free grammars (PCFGs). Both HMMs and PCFGs have a variety of applications, e.g., speech recognition, anomaly detection, and bioinformatics. PCFGs are more expressive than HMMs, e.g., they are more suited for studying protein folding or natural language processing. However, they suffer from slow parameter fitting, which is cubic in the observation sequence length. The same process for HMMs is just linear using the well-known forward-backward algorithm. We argue that by adding to each state of an HMM an integer counter, e.g., representing the number of clients waiting in a queue, brings its expressivity closer to PCFGs. At the same time, we show that parameter fitting for such a model is computationally inexpensive: it is bi-linear in the length of the observation sequence and the maximal counter value, which grows slower than the observation length. The resulting model of H1MMs allows us to combine the best of both worlds: more expressivity with faster parameter fitting. © 2022 International Joint Conferences on Artificial Intelligence. All rights reserved.
Description
Artificial Intelligence Journal; Didi Chuxing; et al.; FinVolution Group; International Joint Conferences on Artifical Intelligence (IJCAI); Shanghai Artificial Intelligence Industry Association
Keywords
Fields of Science
Citation
WoS Q
N/A
Scopus Q
Q2
Source
IJCAI International Joint Conference on Artificial Intelligence -- 31st International Joint Conference on Artificial Intelligence, IJCAI 2022 -- 2022-07-23 through 2022-07-29 -- Vienna -- 182301
Volume
Issue
Start Page
4857
End Page
4863
