A Classification And Regression Tree (CART), is a predictive model, which explains how an outcome variable's values can be predicted based on other values. A CART output is a decision tree where each branch is a split in a predictor variable and each end node contains a prediction for the outcome variable.
This article describes how to create a Classification And Regression Tree (CART). The below example shows an interactive tree created using the Sankey output option using 'Preferred Cola' as the Outcome variable and 'Age', 'Gender', and 'Exercise Frequency' as the Predictor variables.
Requirements
 An Outcome variable to be predicted (eg, Prefer Cola).
 Predictors variables will be considered as predictors of the outcome variable (eg, Age, Gender, Excercise frequency). Note: Predictors that are considered to be uninformative will be automatically excluded from the model.
Please note these steps require a Displayr license.
Method
 Click Anything > Advanced Analysis > Machine Learning > Classification and Regression Trees (CART).
 In the object inspector go to the Inputs tab.
 On the Data menu select the variable to be predicted by the predictor variables.
 Select the predictor variable(s) from the Predictor(s) list.
 OPTIONAL: Select the algorithm you wish to use in the Data > Algorithm. By default, the CART algorithm is used.
 OPTIONAL: Select the desired Output type:

 Sankey: An interactive tree (as shown above). This is the default.
 Tree: A greyscale tree plot.
 Text: A text representation of the tree.
 PredictionAccuracy Table: Produces a table relating the observed and predicted outcome.
 Cross Validation: A plot of the crossvalidation accuracy versus the size of the tree in terms of the number of leaves.

 OPTIONAL: Select the desired Missing Data treatment. (See Missing Data Options).
 OPTIONAL: Select Variable names to display variable names in the output instead of labels.
 OPTIONAL: Select the type of postpruning applied to the tree from the Pruning menu:
 Minimum error: Prune back leaf nodes to create the tree with the smallest cross validation error.
 Smallest tree: Prune to create the smallest tree with cross validation error at most 1 standard error greater than the minimum error.
 None: Retain the tree as it has been built. Note that choosing this option without Early stopping (see below) is prone to overfitting.
 OPTIONAL: Tick the Early stopping box if you wish to stop splitting nodes when the fit stops improving.

OPTIONAL: To shorten category labels from categorical predictor variables select the desired option from the Predictor category labels menu:
 Full labels: The complete labels.
 Abbreviated labels: Labels that have been shortened by taking the first few letters from each word.
 Letters: Letters from the alphabet where "a" corresponds to the first category, "b" to the second category, and so on.
 OPTIONAL: To shorten category labels of the outcome variable select the desired option from the Outcome category labels menu. The choices are the same as above (see Step 11).
 OPTIONAL: Select Allow longrunning calculations to allow categorical variables with more than 30 categories to be included amongst Predictor variables. Note, that predictors with m categories require evaluation of 2^(m  1) split points. Enabling this option may cause calculations to run for a long time.
See Also
How to Run Machine Learning Diagnostics  PredictionAccuracy Table
How to Run Machine Learning Diagnostics  Table of Discriminant Function Coefficients extension
How to Create an Ensemble of Machine Learning Models
How to Run a Gradient Boosting Machine Learning Model
How to Compare Machine Learning Models
How to Run Machine Learning Linear Discriminant Analysis
How to Save Machine Learning Discrimination Variables
How to Save Machine Learning Predicted Values Variables
How to Save Machine Learning Probability of Each Response Variable
How to Run Support Vector Machine