Basic Algorithm - 2024.2 English - XD160

Vitis Libraries

Document ID
XD160
Release Date
2024-11-29
Version
2024.2 English
Starting from the root node, compute all information gains of all features’ splits to select the best partition as the judging conditions. The following above rules are used to grow tree nodes, until you reach stopping rules:
  1. The node depth is equal to the maxDepth training parameter.
  2. No split candidate leads to an information gain greater than minInfoGain.
  3. No split candidate produces child nodes which each have at least minInstancesPerNode training instances.

Node impurity and -information gain:

The node impurity is a measure of the homogeneity of the labels at the node. The information gain is the difference between the parent node impurity and the weighted sum of the two child node impurities. Use the gini impurity and variance impurity for classification and regression scenario, respectively. Meanwhile, the information gain is used to find the best feature split in your implementation.

gini impurity
variance impurity
info gain

Caution

The current implementation provides one impurity measure (gini) for classification and one impurity (variance) for regression. Entropy (only for regression) is to be extended.