A stochastic memoizer for sequence data

Frank Wood*, Cédric Archambeau, Jan Gasthaus, Lancelot James, Yee Whye Teh

*Corresponding author for this work

Research output: Chapter in Book/Conference Proceeding/ReportConference Paper published in a bookpeer-review

Abstract

We propose an unbounded-depth, hierarchical, Bayesian nonparametric model for discrete sequence data. This model can be estimated from a single training sequence, yet shares statistical strength between subsequent symbol predictive distributions in such a way that predictive performance generalizes well. The model builds on a specific parameterization of an unbounded-depth hierarchical Pitman-Yor process. We introduce analytic marginalization steps (using coagulation operators) to reduce this model to one that can be represented in time and space linear in the length of the training sequence. We show how to perform inference in such a model without truncation approximation and introduce fragmentation operators necessary to do predictive inference. We demonstrate the sequence memoizer by using it as a language model, achieving state-of-the-art results.

Original languageEnglish
Title of host publicationProceedings of the 26th Annual International Conference on Machine Learning, ICML'09
DOIs
Publication statusPublished - 2009
Event26th Annual International Conference on Machine Learning, ICML'09 - Montreal, QC, Canada
Duration: 14 Jun 200918 Jun 2009

Publication series

NameACM International Conference Proceeding Series
Volume382

Conference

Conference26th Annual International Conference on Machine Learning, ICML'09
Country/TerritoryCanada
CityMontreal, QC
Period14/06/0918/06/09

Fingerprint

Dive into the research topics of 'A stochastic memoizer for sequence data'. Together they form a unique fingerprint.

Cite this