Training a classifier
You can train a classifier by providing it with training data that it uses to determine how documents should be classified.
About this task
After you create and save a classifier, the classifier training page Overview tab is displayed. This shows the status of the latest model, if any have been previously created.
Procedure
Results
Models list - Your new model is
added to the Models table on the Overview tab. You can browse model details from the
list.
- You can Deploy/Undeploy, Cancel training, or Delete a model from here
- For a normal model, the Model details dialog is open for the model detail
- For federated models, a tab for the federated model is opened
Model details dialog - Once model training is completed (step 1 in the
procedure), detailed model information is available in the Model details dialog.
- Details tab
- Model evaluation scores: F1, precision and recall of the model
- Number of labels: number of labels that the model is expected to output (except for federated models)
- Labels tab
- Label evaluation scores: F1, precision and recall scores for each label and confusion
matrix, showing how many of the validation data the trained classifier classifies as:
- True positive - the data has the label and the classifier correctly predicts it
- False positive - the data doesn't have the label but the classifier wrongly predicts it
- False negative - validation data has the label but the classifier failed to the label)
Note: You can switch to a Cards-style view from the icon at the right-top of the table.
- Label evaluation scores: F1, precision and recall scores for each label and confusion
matrix, showing how many of the validation data the trained classifier classifies as:
- Training tab
- Training resource configuration: configuration used to train this particular model (except for federated models)
- Training loss: A graph showing how the loss value moved during the training
- Runtime tab
- Deployed instance: Information about "labeler" enrichment when this model is deployed (except for individual models)
- Parameters (except for federated models)
- Probability threshold (read-only)Note: Change in this value should affect model evaluation scores, but the scores available in the UI show the initial scores (the values calculated with the initial threshold value).
- Probability threshold (read-only)
Federated model tab - The Federated model tab displays a model information card
with overview information about your federated model. Select the Show model
details menu to open the Model information dialog for the federated model
- Individual models table - Provides a list of individual models in the federated model
- Name, F1 score, Recall, and Precision columns: Show the individual model information for those metrics
- Probability threshold column: Shows the current value for the model. When the value has a
(Pending)
suffix, the value is used to update the model the next time you deploy the federated model. You can set the(Pending)
value by selecting the row and choosing the Edit probability threshold button shown on top of the table - Model training configuration column: Shows how the model will be retrained the next time
you train a federated model, based on the current federated model. The options are:
-
- Update training set (default): Keeps the current test and validation set, and updates the training set with newly-added documents
- Reinitialize training resource: Splits training data into training, validation, and test sets by specified ratios
- Keep the current model: Keeps using the current model
You can update the strategy by selecting the row and choosing the Edit model training configuration button shown on top of the table
-