Turner, RE and Sahani, M (2007) A Maximum-Likelihood Interpretation for Slow Feature Analysis. Neural Computation, 19. pp. 1022-1038. ISSN 0899-7667Full text not available from this repository.
The brain extracts useful features from a maelstrom of sensory information, and a fundamental goal of theoretical neuroscience is to work out how it does so. One proposed feature extraction strategy is motivated by the observation that the meaning of sensory data, such as the identity of a moving visual object, is often more persistent than the activation of any single sensory receptor. This notion is embodied in the slow feature analysis (SFA) algorithm, which uses “slowness” as an heuristic by which to extract semantic information from multi-dimensional time-series. Here, we develop a probabilistic interpretation of this algorithm showing that inference and learning in the limiting case of a suitable probabilistic model yield exactly the results of SFA. Similar equivalences have proved useful in interpreting and extending comparable algorithms such as independent component analysis. For SFA, we use the equivalent probabilistic model as a conceptual spring-board, with which to motivate several novel extensions to the algorithm.
|Divisions:||Div F > Computational and Biological Learning|
|Depositing User:||Unnamed user with email email@example.com|
|Date Deposited:||15 Dec 2015 13:19|
|Last Modified:||10 Feb 2016 23:37|