Advance Learning Algorithms

Week 3 - Advice on applying ML

Error metrics for skewed datasets

Week 4 - Decision Tree / Tree Ensembles

Decision tree

Steps:

Decisions

  1. How to choose what feature to split on at each node?
    • Maximize Purity (or minimize impurity): How even the split (more = good)?
  2. When do you stop splitting?
    • When a node is 100% one class
    • When splitting a node will result in the tree exceeding a maximum depth (defined parameter)
    • When improvements in purity score are below a threshold
    • When number of examples in a node is below a threshold

Regression Decision Tree

Tree Ensemble (Multiple Decision Trees)

When to use Decision Tree and Tree Ensembles

Neural Networks