Yolo Tầm nhìn Thâm Quyến
Thâm Quyến
Tham gia ngay
Bảng chú giải thuật ngữ

Học chuyển giao

Mở khóa sức mạnh của học chuyển giao để tiết kiệm thời gian, tăng hiệu suất AI và giải quyết các nhiệm vụ mới với dữ liệu hạn chế bằng cách sử dụng các mô hình được huấn luyện trước.

Transfer learning is a powerful technique in machine learning (ML) where a model developed for a specific task is reused as the starting point for a model on a second, related task. Instead of training a neural network from scratch—which requires massive datasets and significant computational power—developers leverage the knowledge an AI has already gained. This approach mimics how humans learn; for example, knowing how to play the piano makes it much easier to learn the organ because the foundational understanding of music theory and finger dexterity transfers over. In the context of deep learning, this means a model can achieve high accuracy on a new problem with significantly less data and time.

Cách hoạt động của Transfer Learning

The effectiveness of transfer learning lies in the hierarchical nature of feature extraction. Deep learning models, particularly those used in computer vision, learn to recognize patterns in layers. The initial layers of the backbone detect simple, universal features like edges, curves, and textures. These low-level features are applicable to almost any visual task.

The process typically involves two main phases:

  1. Pre-training: A model is trained on a large-scale benchmark dataset, such as ImageNet, to learn general visual representations. This results in a set of model weights that already understand visual structure.
  2. Adaptation: The pre-trained model is then adapted to a specific niche task. This is often done by "freezing" the early layers (keeping their weights fixed) and retraining only the final layers, or the detection head, on a smaller, custom dataset.

Các Ứng dụng Thực tế

Transfer learning has democratized AI by allowing specialized solutions to be built without Big Tech resources.

  • AI in Healthcare: It is difficult to gather millions of annotated medical images for every specific disease. However, researchers can take a model pre-trained on everyday objects and apply it to medical image analysis. The model transfers its ability to detect shapes and anomalies to identify tumors in X-rays or MRI scans with high precision.
  • AI in Manufacturing: In industrial settings, visual inspection systems must adapt quickly to new product lines. A generalized defect detection model can be rapidly updated to spot flaws in a specific new component, such as a microchip, utilizing smart manufacturing workflows to minimize downtime.

Mối quan hệ với các khái niệm khác

It is helpful to distinguish transfer learning from closely related terms:

  • vs. Fine-Tuning: Fine-tuning is a specific method of implementing transfer learning. While transfer learning is the overarching concept of reusing knowledge, fine-tuning refers to the mechanical process of unfreezing parts of the model and training them on new data with a lower learning rate.
  • So với Học không cần huấn luyện (Zero-Shot Learning ): Học chuyển giao yêu cầu một giai đoạn huấn luyện với một số dữ liệu được gắn nhãn cho nhiệm vụ mới. Ngược lại, học không cần huấn luyện cố gắng... classify các đối tượng mà mô hình chưa từng thấy trước đây, thường dựa vào mô tả ngữ nghĩa hơn là ví dụ trực quan.

Ví dụ thực tế

Sau đây là Python snippet demonstrates transfer learning using the ultralytics library. We load the YOLO26 model, which comes with pre-trained weights derived from the COCO dataset. When we initiate training on a new dataset, the model automatically transfers its pre-learned features to the new task.

from ultralytics import YOLO

# Load a pre-trained YOLO26 model (transferring weights from COCO)
model = YOLO("yolo26n.pt")

# Train the model on a new, smaller dataset to adapt its knowledge
# This leverages the pre-learned backbone for faster convergence
results = model.train(data="coco8.yaml", epochs=5)

For managing datasets and executing these training runs in the cloud, tools like the Ultralytics Platform streamline the process, allowing teams to collaborate on annotating data and deploying transfer-learned models efficiently.

For a deeper dive into the academic theory, the Stanford CS231n notes offer an excellent overview, while the PyTorch Transfer Learning Tutorial provides extensive technical details for implementation.

Tham gia Ultralytics cộng đồng

Tham gia vào tương lai của AI. Kết nối, hợp tác và phát triển cùng với những nhà đổi mới toàn cầu

Tham gia ngay