Calculate the information gain when splitting on A and B. Which attribute would the decision tree induction algorithm choose?
Consider the following data set for a binary class problem.
The contingency tables after splitting on attributes A and B are:
The overall entropy before splitting is:
Eorig = ?0.4 log 0.4 ? 0.6 log 0.6=0.9710
The information gain after splitting on A is:
Computer Science & Information Technology
You might also like to view...
When charting data, typically you should not include __________.
a. column headings b. row headings c. totals d. nonadjacent columns
Computer Science & Information Technology
A runtime error is usually the result of
a. a logical error b. a syntax error c. a compiler error d. bad data
Computer Science & Information Technology