Esta página aún no se ha traducido para esta versión. Puede ver la versión más reciente de esta página en inglés.
This example shows how to detect outliers using quantile random forest. Quantile random forest can detect outliers with respect to the conditional distribution of given . However, this method cannot detect outliers in the predictor data. For outlier detection in the predictor data using a bag of decision trees, see the
OutlierMeasure property of a
An outlier is an observation that is located far enough from most of the other observations in a data set and can be considered anomalous. Causes of outlying observations include inherent variability or measurement error. Outliers significant affect estimates and inference, so it is important to detect them and decide whether to remove them or consider a robust analysis.
The Statistics and Machine Learning Toolbox™ software provides several functionalities to detect or remove outliers including:
zscore — Compute z scores of observations.
trimmean — Estimate mean of data, excluding outliers.
boxplot — Draw box plot of data.
probplot — Draw probability plot.
robustcov — Estimate robust covariance of multivariate data.
fitcsvm — Fit a one-class support vector machine (SVM) to determine which observations are located far from the decision boundary.
To demonstrate outlier detection, this example:
Generates data from a nonlinear model with heteroscedasticity and simulates a few outliers.
Grows a quantile random forest of regression trees.
Estimates conditional quartiles (, , and ) and the interquartile range () within the ranges of the predictor variables.
Compares the observations to the fences, which are the quantities and . Any observation that is less than or greater than is an outlier.
Generate 500 observations from the model
is uniformly distributed between 0 and , and . Store the data in a table.
n = 500; rng('default'); % For reproducibility t = randsample(linspace(0,4*pi,1e6),n,true)'; epsilon = randn(n,1).*sqrt((t+0.01)); y = 10 + 3*t + t.*sin(2*t) + epsilon; Tbl = table(t,y);
Move five observations in a random vertical direction by 90% of the value of the response.
numOut = 5; [~,idx] = datasample(Tbl,numOut); Tbl.y(idx) = Tbl.y(idx) + randsample([-1 1],numOut,true)'.*(0.9*Tbl.y(idx));
Draw a scatter plot of the data and identify the outliers.
figure; plot(Tbl.t,Tbl.y,'.'); hold on plot(Tbl.t(idx),Tbl.y(idx),'*'); axis tight; ylabel('y'); xlabel('t'); title('Scatter Plot of Data'); legend('Data','Simulated outliers','Location','NorthWest');
Grow a bag of 200 regression trees using
Mdl = TreeBagger(200,Tbl,'y','Method','regression');
Mdl is a
Using quantile regression, estimate the conditional quartiles of 50 equally spaced values within the range of
tau = [0.25 0.5 0.75]; predT = linspace(0,4*pi,50)'; quartiles = quantilePredict(Mdl,predT,'Quantile',tau);
quartiles is a 500-by-3 matrix of conditional quartiles. Rows correspond to the observations in
t, and columns correspond to the probabilities in
On the scatter plot of the data, plot the conditional mean and median responses.
meanY = predict(Mdl,predT); plot(predT,[quartiles(:,2) meanY],'LineWidth',2); legend('Data','Simulated outliers','Median response','Mean response',... 'Location','NorthWest'); hold off;
Although the conditional mean and median curves are close, the simulated outliers can affect the mean curve.
Compute the conditional , , and .
iqr = quartiles(:,3) - quartiles(:,1); k = 1.5; f1 = quartiles(:,1) - k*iqr; f2 = quartiles(:,3) + k*iqr;
k = 1.5 means that all observations less than
f1 or greater than
f2 are considered outliers, but this threshold does not disambiguate from extreme outliers. A
3 identifies extreme outliers.
Plot the observations and the fences.
figure; plot(Tbl.t,Tbl.y,'.'); hold on plot(Tbl.t(idx),Tbl.y(idx),'*'); plot(predT,[f1 f2]); legend('Data','Simulated outliers','F_1','F_2','Location','NorthWest'); axis tight title('Outlier Detection Using Quantile Regression') hold off
All simulated outliers fall outside , and some observations are outside this interval as well.