The identification of context-sensitive features: A formal definition of context for concept learning

Turney, Peter (1996) The identification of context-sensitive features: A formal definition of context for concept learning. [Conference Paper]

Full text available as:



A large body of research in machine learning is concerned with supervised learning from examples. The examples are typically represented as vectors in a multi- dimensional feature space (also known as attribute-value descriptions). A teacher partitions a set of training examples into a finite number of classes. The task of the learning algorithm is to induce a concept from the training examples. In this paper, we formally distinguish three types of features: primary, contextual, and irrelevant features. We also formally define what it means for one feature to be context-sensitive to another feature. Context-sensitive features complicate the task of the learner and potentially impair the learner's performance. Our formal definitions make it possible for a learner to automatically identify context-sensitive features. After context-sensitive features have been identified, there are several strategies that the learner can employ for managing the features; however, a discussion of these strategies is outside of the scope of this paper. The formal definitions presented here correct a flaw in previously proposed definitions. We discuss the relationship between our work and a formal definition of relevance.

Item Type:Conference Paper
Subjects:Computer Science > Artificial Intelligence
Computer Science > Machine Learning
Computer Science > Statistical Models
ID Code:1866
Deposited By:Turney, Peter
Deposited On:11 Nov 2001
Last Modified:11 Mar 2011 08:54

References in Article

Select the SEEK icon to attempt to find the referenced article. If it does not appear to be in cogprints you will be forwarded to the paracite service. Poorly formated references will probably not work.

Bergadano, F., Matwin, S., Michalski, R.S., and Zhang, J.

(1992). Learning two-tiered descriptions of flexible

concepts: The POSEIDON system. Machine Learning, 8,


John, G.H., Kohavi, R., and Pfleger, K. (1994). Irrelevant

features and the subset selection problem, Machine

Learning: Proceedings of the Eleventh International Con-ference,

pp. 121-129, California: Morgan Kaufmann.

Katz, A.J., Gately, M.T., and Collins, D.R. (1990). Robust

classifiers without robust features, Neural Computation, 2,


Michalski, R.S. (1987). How to learn imprecise concepts:

A method employing a two-tiered knowledge representa-tion

for learning. Proceedings of the Fourth International

Workshop on Machine Learning, pp. 50-58, California:

Morgan Kaufmann.

Pratt, L.Y., Mostow, J, and Kamm, C.A. (1991). Direct

transfer of learned information among neural networks.

Proceedings of the 9th National Conference on Artificial

Intelligence (AAAI-91), pp. 584-580, Anaheim, California.

Turney, P.D. (1993a). Exploiting context when learning to

classify. In Proceedings of the European Conference on

Machine Learning, ECML-93, pp. 402-407. Vienna,

Austria: Springer-Verlag.

Turney, P.D. (1993b). Robust classification with context-sensitive

features. In Industrial and Engineering Applica-tions

of Artificial Intelligence and Expert Systems, IEA/

AIE-93, pp. 268-276. Edinburgh, Scotland: Gordon and


Watrous, R.L. (1991). Context-modulated vowel discrimi-nation

using connectionist networks. Computer Speech

and Language, 5, 341-362.

Widmer, G. and Kubat, M. (1992). Learning flexible

concepts from streams of examples: FLORA2. In Pro-ceedings

of the 10th European Conference on Artificial

Intelligence (ECAI-92), Vienna. Chichester: Wiley and


Widmer, G. and Kubat, M. (1993). Effective learning in

dynamic environments by explicit context tracking. In

Proceedings of the European Conference on Machine

Learning (ECML-93), 227-243, Vienna, Austria. Berlin:

Springer Verlag.


Repository Staff Only: item control page