Main Content

This example shows how to detect outliers using quantile random forest. Quantile random forest can detect outliers with respect to the conditional distribution of $$Y$$ given $$X$$. However, this method cannot detect outliers in the predictor data. For outlier detection in the predictor data using a bag of decision trees, see the `OutlierMeasure`

property of a `TreeBagger`

model.

An *outlier* is an observation that is located far enough from most of the other observations in a data set and can be considered anomalous. Causes of outlying observations include inherent variability or measurement error. Outliers significant affect estimates and inference, so it is important to detect them and decide whether to remove them or consider a robust analysis.

Statistics and Machine Learning Toolbox™ provides several functions to detect outliers, including:

`zscore`

— Compute*z*scores of observations.`trimmean`

— Estimate mean of data, excluding outliers.`boxplot`

— Draw box plot of data.`probplot`

— Draw probability plot.`robustcov`

— Estimate robust covariance of multivariate data.`fitcsvm`

— Fit a one-class support vector machine (SVM) to determine which observations are located far from the decision boundary.`dbscan`

— Partition observations into clusters and identify outliers using the density-based spatial clustering of application with noise (DBSCAN) algorithm.

Also, MATLAB® provides the `isoutlier`

function, which finds outliers in data.

To demonstrate outlier detection, this example:

Generates data from a nonlinear model with heteroscedasticity and simulates a few outliers.

Grows a quantile random forest of regression trees.

Estimates conditional quartiles ($${Q}_{1}$$, $${Q}_{2}$$, and $${Q}_{3}$$) and the interquartile range ($$IQR$$) within the ranges of the predictor variables.

Compares the observations to the

*fences*, which are the quantities $${F}_{1}={Q}_{1}-1.5IQR$$ and $${F}_{2}={Q}_{3}+1.5IQR$$. Any observation that is less than $${F}_{1}$$ or greater than $${F}_{2}$$ is an outlier.

Generate 500 observations from the model

$${y}_{t}=10+3t+t\mathrm{sin}(2t)+{\epsilon}_{t}.$$

$$t$$ is uniformly distributed between 0 and $$4\pi $$, and $${\epsilon}_{t}\sim N(0,t+0.01)$$. Store the data in a table.

n = 500; rng('default'); % For reproducibility t = randsample(linspace(0,4*pi,1e6),n,true)'; epsilon = randn(n,1).*sqrt((t+0.01)); y = 10 + 3*t + t.*sin(2*t) + epsilon; Tbl = table(t,y);

Move five observations in a random vertical direction by 90% of the value of the response.

numOut = 5; [~,idx] = datasample(Tbl,numOut); Tbl.y(idx) = Tbl.y(idx) + randsample([-1 1],numOut,true)'.*(0.9*Tbl.y(idx));

Draw a scatter plot of the data and identify the outliers.

figure; plot(Tbl.t,Tbl.y,'.'); hold on plot(Tbl.t(idx),Tbl.y(idx),'*'); axis tight; ylabel('y'); xlabel('t'); title('Scatter Plot of Data'); legend('Data','Simulated outliers','Location','NorthWest');

Grow a bag of 200 regression trees using `TreeBagger`

.

Mdl = TreeBagger(200,Tbl,'y','Method','regression');

`Mdl`

is a `TreeBagger`

ensemble.

Using quantile regression, estimate the conditional quartiles of 50 equally spaced values within the range of `t`

.

```
tau = [0.25 0.5 0.75];
predT = linspace(0,4*pi,50)';
quartiles = quantilePredict(Mdl,predT,'Quantile',tau);
```

`quartiles`

is a 500-by-3 matrix of conditional quartiles. Rows correspond to the observations in `t`

, and columns correspond to the probabilities in `tau`

.

On the scatter plot of the data, plot the conditional mean and median responses.

meanY = predict(Mdl,predT); plot(predT,[quartiles(:,2) meanY],'LineWidth',2); legend('Data','Simulated outliers','Median response','Mean response',... 'Location','NorthWest'); hold off;

Although the conditional mean and median curves are close, the simulated outliers can affect the mean curve.

Compute the conditional $$IQR$$, $${F}_{1}$$, and $${F}_{2}$$.

iqr = quartiles(:,3) - quartiles(:,1); k = 1.5; f1 = quartiles(:,1) - k*iqr; f2 = quartiles(:,3) + k*iqr;

`k = 1.5`

means that all observations less than `f1`

or greater than `f2`

are considered outliers, but this threshold does not disambiguate from extreme outliers. A `k`

of `3`

identifies extreme outliers.

Plot the observations and the fences.

figure; plot(Tbl.t,Tbl.y,'.'); hold on plot(Tbl.t(idx),Tbl.y(idx),'*'); plot(predT,[f1 f2]); legend('Data','Simulated outliers','F_1','F_2','Location','NorthWest'); axis tight title('Outlier Detection Using Quantile Regression') hold off

All simulated outliers fall outside $$[{F}_{1},{F}_{2}]$$, and some observations are outside this interval as well.