Simplifying mixture models through function approximation

  • Kai Zhang*
  • , James T. Kwok
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

15 Scopus citations

Abstract

Finite mixture model is a powerful tool in many statistical learning problems. In this paper, we propose a general, structure-preserving approach to reduce its model complexity, which can bring significant computational benefits in many applications. The basic idea is to group the original mixture components into compact clusters, and then minimize an upper bound on the approximation error between the original and simplified models. By adopting the L2norm as the distance measure between mixture models, we can derive closed-form solutions that aremore robust and reliable than using the KL-based distancemeasure. Moreover, the complexity of our algorithm is only linear in the sample size and dimensionality. Experiments on density estimation and clustering-based image segmentation demonstrate its outstanding performance in terms of both speed and accuracy.

Original languageEnglish
Title of host publicationAdvances in Neural Information Processing Systems 19 - Proceedings of the 2006 Conference
PublisherNeural information processing systems foundation
Pages1577-1584
Number of pages8
ISBN (Print)9780262195683
StatePublished - 1 Dec 2007
Externally publishedYes
Event20th Annual Conference on Neural Information Processing Systems, NIPS 2006 - Vancouver, BC, Canada
Duration: 4 Dec 20067 Dec 2006

Publication series

NameAdvances in Neural Information Processing Systems
ISSN (Print)1049-5258

Conference

Conference20th Annual Conference on Neural Information Processing Systems, NIPS 2006
Country/TerritoryCanada
CityVancouver, BC
Period4/12/067/12/06

Fingerprint

Dive into the research topics of 'Simplifying mixture models through function approximation'. Together they form a unique fingerprint.

Cite this