The following expands the Discussions of the two technical reports.

Generative vs discriminative: image-specific explaining-away


Sparse coding: attributed elements and further coding of attributes


AND-OR graph: selective composition for AND-nodes and arg-max inference for OR-nodes


Max pooling: max for arg-max and arg-max for sketch


Cortex-like architecture: sparse connection and arg-max synchronization?


Layered belief network: learning is trimming (more generally, learning is wiring)?


Avoiding explaining-away in inference

The following are some strategies: We prefer the last strategy, which is to hardwire the explaining-away in learning.

Past issues

Two issues with generative models that had puzzled us for a long time: These two obstacles have been circumvented in our current work as we explained above.

Key principles


Key references

Form of representation: Scheme of learning: Architecture of inference:

Past papers

This work is a continuation of our long term search for generative models and model-based algorithms, as well as our attempt to understand these models within a common information-theoretical framework. The active basis model can be considered a revision of our previous model on textons. It can also be viewed as an inhomogeneous version of the Markov random field model that we previously developed for textures. More important, the active basis model is a simplest instance of the and-or graph in the compositional framework that we have been studying. The and-or grammar naturally suggests to further compose multiple active bases to represent more articulate shapes. The architecture of the sum-max maps is a natural computational tool for parsing the observed image according to the and-or grammar.

Back to active basis homepage