creators_name: Rachkovskij, Dmitri A. creators_name: Kussul, Ernst M. type: preprint datestamp: 1999-04-30 lastmod: 2011-03-11 08:54:02 metadata_visibility: show title: Binding and Normalization of Binary Sparse Distributed Representations by Context-Dependent Thinning ispublished: unpub subjects: comp-sci-art-intel subjects: comp-sci-neural-nets full_text_status: public keywords: distributed representation, sparse coding, binary coding, binding, variable binding, thinning, representation of structure, structured representation, recursive representation, nested representation, compositional representation, connectionist symbol processing. abstract: Distributed representations were often criticized as inappropriate for encoding of data with a complex structure. However Plate's Holographic Reduced Representations and Kanerva's Binary Spatter Codes are recent schemes that allow on-the-fly encoding of nested compositional structures by real-valued or dense binary vectors of fixed dimensionality. In this paper we consider procedures of the Context-Dependent Thinning which were developed for representation of complex hierarchical items in the architecture of Associative-Projective Neural Networks. These procedures provide binding of items represented by sparse binary codevectors (with low probability of 1s). Such an encoding is biologically plausible and allows to reach high information capacity of distributed associative memory where the codevectors may be stored. In distinction to known binding procedures, Context-Dependent Thinning allows to support the same low density (or sparseness) of the bound codevector for varied number of constituent codevectors. Besides, a bound codevector is not only similar to another one with similar constituent codevectors (as in other schemes), but it is also similar to the constituent codevectors themselves. This allows to estimate a structure similarity just by the overlap of codevectors, without the retrieval of the constituent codevectors. This also allows an easy retrieval of the constituent codevectors. Examples of algorithmic and neural network implementations of the thinning procedures are considered. We also present representation examples of various types of nested structured data (propositions using role-filler and predicate-arguments representation, trees, directed acyclic graphs) using sparse codevectors of fixed dimension. Such representations may provide a fruitful alternative to the symbolic representations of traditional AI, as well as to the localist and microfeature-based connectionist representations. date: 1999-03 date_type: published refereed: FALSE citation: Rachkovskij, Dmitri A. and Kussul, Ernst M. (1999) Binding and Normalization of Binary Sparse Distributed Representations by Context-Dependent Thinning. [Preprint] (Unpublished) document_url: http://cogprints.org/537/2/Cdt.ps