Machine Learning.

random forest
Random Forest Variable Importance: Calculation, Accuracy, & Examples
30 January 2025 | by Jake Hoare

Learn how variable importance is calculated in random forests using both accuracy-based and Gini-based measures.

Continue reading

Decision Trees Are Usually Better Than Logistic Regression
25 October 2018 | by Tim Bock

Logistic regression is a standard approach to building a predictive model. However, decision trees are an alternative which are clearer and often superior.

Continue reading

Feature Engineering for Categorical Variables
24 October 2018 | by Tim Bock

There are two types of predictors in predictive models: numeric and categorical. There are several methods of transforming categorical variables.

Continue reading

Feature Engineering for Numeric Variables
24 October 2018 | by Tim Bock

When building a predictive model, it is often practical to improve predictive performance by modifying the numeric variables. This is called transformation.

Continue reading

view of a forest
How Random Forests Fit to Data
06 August 2018 | by Jake Hoare

A random forest is a collection of decision trees, which is used to learn patterns in data and make predictions based on those patterns.

Continue reading

tree on a hill
How Decision Trees Choose the Best Split (with Examples)
02 August 2018 | by Jake Hoare

Decision trees work by repeatedly splitting the data to lead to the option which causes the greatest improvement. We explain how these splits are chosen.

Continue reading

Machine learning pruning
Pruning Decision Trees: A Guide to Pre-Pruning and Post-Pruning
04 July 2017 | by Jake Hoare

Machine learning is an issue of trade-offs. Here we look at pruning and other ways of managing these trade-offs in the context of decision trees. Read more.

Continue reading