* Entropy is an indicator of how messy your data is. It decreases as you reach closer to the leaf node.
* The Information Gain is based on the decrease in entropy after a dataset is split on an attribute. It keeps on increasing as you reach closer to the leaf node.