Yolo Vision Shenzhen
Shenzhen
Rejoindre maintenant
Glossaire

Apprentissage par transfert

Libérez la puissance de l'apprentissage par transfert pour gagner du temps, améliorer les performances de l'IA et aborder de nouvelles tâches avec des données limitées à l'aide de modèles pré-entraînés.

Transfer learning is a powerful technique in machine learning (ML) where a model developed for a specific task is reused as the starting point for a model on a second, related task. Instead of training a neural network from scratch—which requires massive datasets and significant computational power—developers leverage the knowledge an AI has already gained. This approach mimics how humans learn; for example, knowing how to play the piano makes it much easier to learn the organ because the foundational understanding of music theory and finger dexterity transfers over. In the context of deep learning, this means a model can achieve high accuracy on a new problem with significantly less data and time.

Fonctionnement de l'apprentissage par transfert

The effectiveness of transfer learning lies in the hierarchical nature of feature extraction. Deep learning models, particularly those used in computer vision, learn to recognize patterns in layers. The initial layers of the backbone detect simple, universal features like edges, curves, and textures. These low-level features are applicable to almost any visual task.

The process typically involves two main phases:

  1. Pre-training: A model is trained on a large-scale benchmark dataset, such as ImageNet, to learn general visual representations. This results in a set of model weights that already understand visual structure.
  2. Adaptation: The pre-trained model is then adapted to a specific niche task. This is often done by "freezing" the early layers (keeping their weights fixed) and retraining only the final layers, or the detection head, on a smaller, custom dataset.

Applications concrètes

Transfer learning has democratized AI by allowing specialized solutions to be built without Big Tech resources.

  • AI in Healthcare: It is difficult to gather millions of annotated medical images for every specific disease. However, researchers can take a model pre-trained on everyday objects and apply it to medical image analysis. The model transfers its ability to detect shapes and anomalies to identify tumors in X-rays or MRI scans with high precision.
  • AI in Manufacturing: In industrial settings, visual inspection systems must adapt quickly to new product lines. A generalized defect detection model can be rapidly updated to spot flaws in a specific new component, such as a microchip, utilizing smart manufacturing workflows to minimize downtime.

Relation avec d'autres concepts

It is helpful to distinguish transfer learning from closely related terms:

  • vs. Fine-Tuning: Fine-tuning is a specific method of implementing transfer learning. While transfer learning is the overarching concept of reusing knowledge, fine-tuning refers to the mechanical process of unfreezing parts of the model and training them on new data with a lower learning rate.
  • vs. Apprentissage sans données: L'apprentissage par transfert nécessite une phase d'entraînement avec certaines données étiquetées pour la nouvelle tâche. En revanche, l'apprentissage sans données tente de classify que le modèle n'a jamais vus auparavant, en s'appuyant souvent sur des descriptions sémantiques plutôt que sur des exemples visuels.

Exemple pratique

Le texte suivant Python snippet demonstrates transfer learning using the ultralytics library. We load the YOLO26 model, which comes with pre-trained weights derived from the COCO dataset. When we initiate training on a new dataset, the model automatically transfers its pre-learned features to the new task.

from ultralytics import YOLO

# Load a pre-trained YOLO26 model (transferring weights from COCO)
model = YOLO("yolo26n.pt")

# Train the model on a new, smaller dataset to adapt its knowledge
# This leverages the pre-learned backbone for faster convergence
results = model.train(data="coco8.yaml", epochs=5)

For managing datasets and executing these training runs in the cloud, tools like the Ultralytics Platform streamline the process, allowing teams to collaborate on annotating data and deploying transfer-learned models efficiently.

For a deeper dive into the academic theory, the Stanford CS231n notes offer an excellent overview, while the PyTorch Transfer Learning Tutorial provides extensive technical details for implementation.

Rejoindre la communauté Ultralytics

Rejoignez le futur de l'IA. Connectez-vous, collaborez et évoluez avec des innovateurs mondiaux.

Rejoindre maintenant