MoTIF: An Efficient Algorithm for Learning Translation Invariant Dictionaries
The performances of approximation using redundant expansions rely on having dictionaries adapted to the signals. In natural high-dimensional data, the statistical dependencies are, most of the time, not obvious. Learning fundamental patterns is an alternative to analytical design of bases and is nowadays a popular problem in the field of approximation theory. In many situations, the basis elements are shift invariant, thus the learning should try to find the best matching filters. We present a new algorithm for learning iteratively generating functions that can be translated at all positions in the signal to generate a highly redundant dictionary.
Jost2006_1450.pdf
openaccess
165.37 KB
Adobe PDF
f9540017418babcc3087737c89ec0fb6