Header logo is ei

Factorial Coding of Natural Images: How Effective are Linear Models in Removing Higher-Order Dependencies?




The performance of unsupervised learning models for natural images is evaluated quantitatively by means of information theory. We estimate the gain in statistical independence (the multi-information reduction) achieved with independent component analysis (ICA), principal component analysis (PCA), zero-phase whitening, and predictive coding. Predictive coding is translated into the transform coding framework, where it can be characterized by the constraint of a triangular filter matrix. A randomly sampled whitening basis and the Haar wavelet are included into the comparison as well. The comparison of all these methods is carried out for different patch sizes, ranging from 2x2 to 16x16 pixels. In spite of large differences in the shape of the basis functions, we find only small differences in the multi-information between all decorrelation transforms (5% or less) for all patch sizes. Among the second-order methods, PCA is optimal for small patch sizes and predictive coding performs best for large patch sizes. The extra gain achieved with ICA is always less than 2%. In conclusion, the `edge filters‘ found with ICA lead only to a surprisingly small improvement in terms of its actual objective.

Author(s): Bethge, M.
Volume: 9
Pages: 90
Year: 2006
Month: March
Day: 0

Department(s): Empirical Inference
Bibtex Type: Poster (poster)

Digital: 0
Event Name: 9th T{\"u}bingen Perception Conference (TWK 2006)
Event Place: T{\"u}bingen, Germany
Language: en
Organization: Max-Planck-Gesellschaft
School: Biologische Kybernetik

Links: Web


  title = {Factorial Coding of Natural Images: How Effective are Linear Models in Removing Higher-Order Dependencies?},
  author = {Bethge, M.},
  volume = {9},
  pages = {90},
  organization = {Max-Planck-Gesellschaft},
  school = {Biologische Kybernetik},
  month = mar,
  year = {2006},
  month_numeric = {3}