HES 505 Fall 2025: Session 21
Grading complete by end of week
NO CLASS WEDS 5 Nov
Final project descriptions!!
Differentiate supervised from unsupervised classification
Recognize the linkage between statistical learning models and interpolation
Define the key elements of tree-based classifiers
Articulate the differences in statistical learning for spatial data
Cognitive Description: collection ordering and classification of data
Cause and Effect: design-based or model-based testing of the factors that give rise to geographic distributions
Systems Analysis: describes the entire complex set of interactions that structure an activity
Deterministic Interpolation: best explanation for variation is distance
Probabilistic Interpolation: best explanation is covariance (plus spatial trends)
\[ \begin{equation} z(\mathbf{x}) = \mu(\mathbf{x}) + \epsilon(\mathbf{x}) \end{equation} \]
Often we actually care about the “drivers” of \(\mu(x)\)
Inference not prediction
Data volume
Sampling designs
Nonlinearity
Autocorrelation
“Machine Learning”
Broad class of methods that rely on efficient, algorithmic learning
Probabilities derived from “large” data (rather than theoretical distributions)
Relax traditional statistical assumptions
Goal is inference/prediction not grouping
Training data: observations with known values used to generate model parameters
Validation data: observations with known values used to tune hyperparameters prior to full model implementation
Testing data: observations held out of initial model-fitting to determine accuracy
Use decision rules to segment the predictor space
Series of consecutive decision rules form a ‘tree’
Terminal nodes (leaves) are the outcome; internal nodes (branches) the splits
Divide the predictor space (\(R\)) into \(J\) non-overlapping regions
Every observation in \(R_j\) gets the same prediction
Recursive binary splitting
Maximize the information gained at each split
Pruning and over-fitting
Benefits
Easy to explain
Links to human decision-making
Graphical displays
Easy handling of qualitative predictors
Costs
Lower predictive accuracy than other methods
Not necessarily robust
Grow 100(000s) of trees using bootstrapping
Random sample of predictors considered at each split
Avoids correlation amongst multiple predictions
Average of trees improves overall outcome (usually)
Lots of extensions
How to build ensembles?
The number of and tuning of hyperparameters
Boosting
No direct way to incorporate spatial autocorrelation
Training Set qualities vs. Inferential conditions
Cross-validation considerations