Information Gain Impurity Function
(Redirected from Information Gain Criteria)
Jump to navigation
Jump to search
An Information Gain Impurity Function is an impurity function based on Kullback-Leibler Divergence.
- Context
- It can be used by a CART algorithm.
- See: information entropy, Gini Index/CART, Chi-Square/CHAID.
References
2011
- http://en.wikipedia.org/wiki/Information_gain_in_decision_trees
- In information theory and machine learning, 'information gain is an alternative synonym for Kullback–Leibler divergence. In particular, the information gain about a random variable X obtained from an observation that a random variable A takes the value A=a is the Kullback-Leibler divergence DKL( p(x|a) || p(x|I) ) of the prior distribution p(x|I) for x from the posterior distribution p(x|a) for x given a. The expected value of the information gain is the mutual information I(X;A) of X and A — i.e. the reduction in the entropy of X achieved by learning the state of the random variable A. In machine learning this concept can be used to define a preferred sequence of attributes to investigate to most rapidly narrow down the state of X. Such a sequence (which depends on the outcome of the investigation of previous attributes at each stage) is called a decision tree. Usually an attribute with high information gain should be preferred to other attributes.
- In general terms, the expected information gain is the change in information entropy from a prior state to a state that takes some information as given: [math]\displaystyle{ IG(Ex,a) = H(Ex) - H(Ex|a) }[/math]
- http://en.wikipedia.org/wiki/Decision_tree_learning#Information_gain
- Used by the ID3, C4.5 and C5.0 tree generation algorithms. Information gain is based on the concept of entropy used in information theory. [math]\displaystyle{ I_{E}(f) = - \sum^{m}_{i=1} f_i \log^{}_2 f_i }[/math]
2004
- (Raileanu & Stoffel, 2004) ⇒ Laura Elena Raileanu, and Kilian Stoffel. (2004). “Theoretical Comparison between the Gini Index and Information Gain Criteria.” In: Annals of Mathematics and Artificial Intelligence, 41(1). doi:10.1023/B:AMAI.0000018580.96245.c6
1993
- (Quinlan, 1993a) ⇒ J. Ross Quinlan. (1993). “C4.5: Programs for machine learning." Morgan Kaufmann. ISBN:1558602380