Finally, the data reaches the output layer, which consolidates the extracted features and produces a final prediction. This prediction is compared to the annotated training dataset to calculate errors and adjust the network’s weights for improved accuracy.
For example, to train a model to recognize images of cats, engineers might use supervised learning, labeling thousands of images with tags such as “cat” or “not cat” so the model can learn key features such as fur texture, whiskers and ear shape.
Alternatively, in unsupervised learning, the model works with unlabeled data to discover patterns independently. The model identifies relationships without predefined categories by clustering images based on shared characteristics (for example, similar shapes or textures).
This approach is helpful for tasks such as fraud detection, quality control and pattern analysis when labeled data is unavailable. In unsupervised learning, the model would independently cluster images based on shared patterns, grouping all cat images without explicitly knowing they are cats.
A third approach, self-supervised learning, combines aspects of unsupervised learning by starting with unlabeled data but generating pseudo labels from the data’s inherent structure, enabling models to learn meaningful representations without traditional labels, making them powerful for tasks with limited labeled datasets.
With self-supervised learning, the model could analyze parts of an image, such as reconstructing a partially obscured cat face, to identify patterns and features. Ultimately, the trained model—whether using machine learning or deep learning—could accurately identify and classify new, unseen images of cats, distinguishing them from other animals or objects.