The Entropy of Relations and a New Approach for Decision Tree Learning

The formula for scaling how much information in relations on the finite universe is proposed, which is called the entropy of relation R and denoted by H (R). Based on the concept of H (R), the entropy of predicates and the information of propositions are measured. We can use these measures to evaluate predicates and choose the most appropriate predicate for some given cartesian set. At last, H (R) is used to induce decision tree. The experiment show that the new induction algorithm denoted by IDIR do better than ID3 on the aspects of nodes and test time.