Calculate the information gain when splitting on A and B. Which attribute would the decision tree induction algorithm choose?

Consider the following data set for a binary class problem.

The contingency tables after splitting on attributes A and B are:





The overall entropy before splitting is:


Eorig = ?0.4 log 0.4 ? 0.6 log 0.6=0.9710


The information gain after splitting on A is:





Computer Science & Information Technology

You might also like to view...

When charting data, typically you should not include __________.

a. column headings b. row headings c. totals d. nonadjacent columns

Computer Science & Information Technology

A runtime error is usually the result of

a. a logical error b. a syntax error c. a compiler error d. bad data

Computer Science & Information Technology