Cogprints

Learning as Extraction of Low-Dimensional Representations

Edelman, Shimon and Intrator, Nathan (1997) Learning as Extraction of Low-Dimensional Representations. [Preprint]

Full text available as:

[img]Postscript
486Kb

Abstract

Psychophysical findings accumulated over the past several decades indicate that perceptual tasks such as similarity judgment tend to be performed on a low-dimensional representation of the sensory data. Low dimensionality is especially important for learning, as the number of examples required for attaining a given level of performance grows exponentially with the dimensionality of the underlying representation space. In this chapter, we argue that, whereas many perceptual problems are tractable precisely because their intrinsic dimensionality is low, the raw dimensionality of the sensory data is normally high, and must be reduced by a nontrivial computational process, which, in itself, may involve learning. Following a survey of computational techniques for dimensionality reduction, we show that it is possible to learn a low-dimensional representation that captures the intrinsic low-dimensional nature of certain classes of visual objects, thereby facilitating further learning of tasks involving those objects.

Item Type:Preprint
Subjects:Psychology > Cognitive Psychology
ID Code:562
Deposited By:Edelman, Shimon
Deposited On:17 Oct 1997
Last Modified:11 Mar 2011 08:54

Metadata

Repository Staff Only: item control page