After you choose a particular type of model to train, for example a decision tree or a support vector machine (SVM), you can tune your model by selecting different advanced options. For example, you can change the maximum number of splits for a decision tree or the box constraint of an SVM. Some of these options are internal parameters of the model, or hyperparameters, that can strongly affect its performance. Instead of manually selecting these options, you can use hyperparameter optimization within the Classification Learner app to automate the selection of hyperparameter values. For a given model type, the app tries different combinations of hyperparameter values by using an optimization scheme that seeks to minimize the model classification error, and returns a model with the optimized hyperparameters. You can use the resulting model as you would any other trained model.
Because hyperparameter optimization can lead to an overfitted model, the recommended approach is to create a separate test set before importing your data into the Classification Learner app. After you train your optimizable model, you can export it from the app and see how it performs on your test set. For an example, see Train Classifier Using Hyperparameter Optimization in Classification Learner App.
To perform hyperparameter optimization in Classification Learner, follow these steps:
Choose a model type and decide which hyperparameters to optimize. See Select Hyperparameters to Optimize.
Hyperparameter optimization is not supported for logistic regression models.
(Optional) Specify how the optimization is performed. For more information, see Optimization Options.
Train your model. Use the Minimum Classification Error Plot to track the optimization results.
Inspect your trained model. See Optimization Results.
In the Classification Learner app, in the Model Type section of the Classification Learner tab, click the arrow to open the gallery. The gallery includes optimizable models that you can train using hyperparameter optimization.
After you select an optimizable model, you can choose which of its hyperparameters you want to optimize. In the Model Type section, select Advanced > Advanced. The app opens a dialog box in which you can select Optimize check boxes for the hyperparameters that you want to optimize. Under Values, specify the fixed values for the hyperparameters that you do not want to optimize or that are not optimizable.
This table describes the hyperparameters that you can optimize for each type of model and the search range of each hyperparameter. It also includes the additional hyperparameters for which you can specify fixed values.
|Model||Optimizable Hyperparameters||Additional Hyperparameters||Notes|
For more information, see Advanced Tree Options.
For more information, see Advanced Discriminant Options.
|Optimizable Naive Bayes||
For more information, see Advanced Naive Bayes Options.
For more information, see Advanced SVM Options.
For more information, see Advanced KNN Options.
For more information, see Advanced Ensemble Options.
By default, the Classification Learner app performs hyperparameter tuning by using Bayesian optimization. The goal of Bayesian optimization, and optimization in general, is to find a point that minimizes an objective function. In the context of hyperparameter tuning in the app, a point is a set of hyperparameter values, and the objective function is the loss function, or the classification error. For more information on the basics of Bayesian optimization, see Bayesian Optimization Workflow.
You can specify how the hyperparameter tuning is performed. For example, you can change the optimization method to grid search or limit the training time. On the Classification Learner tab, in the Model Type section, select Advanced > Optimizer Options. The app opens a dialog box in which you can select optimization options.
This table describes the available optimization options and their default values.
The optimizer values are:
When the app performs Bayesian optimization for hyperparameter tuning, it uses the acquisition function to determine the next set of hyperparameter values to try.
The acquisition function values are:
For details on how these acquisition functions work in the context of Bayesian optimization, see Acquisition Function Types.
Each iteration corresponds to a combination of
hyperparameter values that the app tries. When you use
Bayesian optimization or random search, specify a positive
integer that sets the number of iterations. The default
When you use grid search, the app ignores the Iterations value and evaluates the loss at every point in the entire grid. You can set a training time limit to stop the optimization process prematurely.
|Training time limit||To set a training time limit, select this option and set the Maximum training time in seconds option. By default, the app does not have a training time limit.|
|Maximum training time in seconds||Set the training time limit in seconds as a positive real
number. The default value is |
|Number of grid divisions||When you use grid search, set a positive integer as the
number of values the app tries for each numeric hyperparameter.
The app ignores this value for categorical hyperparameters. The
default value is |
After specifying which model hyperparameters to optimize and setting any additional optimization options (optional), train your optimizable model. On the Classification Learner tab, in the Training section, click Train. The app creates a Minimum Classification Error Plot that it updates as the optimization runs.
When you train an optimizable model, the app disables the Use Parallel button. After the training is complete, the app makes the button available again when you select a nonoptimizable model. The button is off by default.
The minimum classification error plot displays the following information:
Estimated minimum classification error at each iteration – Each light blue point corresponds to the minimum classification error estimated by the optimization process when considering the hyperparameter values at that iteration.
If you use grid search or random search to perform hyperparameter optimization, the app does not display these points.
Observed minimum classification error up to each iteration – Each dark blue point corresponds to the minimum classification error computed so far by the optimization process. For example, at the third iteration, the dark blue point corresponds to the minimum of the classification error observed in the first, second, and third iterations.
Bestpoint hyperparameters – The red square indicates the iteration that corresponds to the optimized hyperparameters. You can find the values of the optimized hyperparameters listed in the upper right of the plot under Optimization Results.
The optimized hyperparameters do not always provide the minimum
classification error. When the app performs hyperparameter tuning by
using Bayesian optimization (see Optimization Options for
a brief introduction), it chooses the set of hyperparameter values that
minimizes an upper confidence interval of the classification error
objective model, rather than the set that minimizes the classification
error. For more information, see the
name-value pair argument of
Minimum error hyperparameters – The yellow point indicates the iteration that corresponds to the hyperparameters that yield the minimum classification error.
If you use grid search to perform hyperparameter optimization, the Bestpoint hyperparameters and the Minimum error hyperparameters are the same.
Missing points in the plot correspond to
classification error values.
When the app finishes tuning model hyperparameters, it returns a model trained with the optimized hyperparameter values. The model metrics, displayed plots, and exported model correspond to this trained model with fixed hyperparameter values.
To inspect the optimization results of a trained optimizable model, select the model in the History list and look at the Current Model pane.
The Current Model pane includes these sections:
Results – Shows the performance of the optimizable model
Model Type – Displays the type of optimizable model and lists any fixed hyperparameter values
Optimized Hyperparameters – Lists the values of the optimized hyperparameters
Hyperparameter Search Range – Displays the search ranges for the optimized hyperparameters
Optimizer Options – Shows the selected optimizer options
When you perform hyperparameter tuning using Bayesian optimization and you export
the resulting trained optimizable model to the workspace as a structure, the
structure includes a
BayesianOptimization object in the
HyperParameterOptimizationResult field. The object contains
the results of the optimization performed in the app.
When you generate MATLAB® code from a trained optimizable model, the generated code uses the fixed and optimized hyperparameter values of the model to train on new data. The generated code does not include the optimization process. For information on how to perform Bayesian optimization when you use a fit function, see Bayesian Optimization Using a Fit Function.