Yolo 비전 선전
선전
지금 참여하기
용어집

모델 앙상블

모델 앙상블로 모델 정확도와 견고성을 향상시키십시오. 배깅, 부스팅, 스태킹 및 실제 응용 프로그램과 같은 기술을 살펴보십시오.

A Model Ensemble is a strategic approach in machine learning where predictions from multiple individual models are combined to produce a final output that is often more accurate and robust than any single model could achieve alone. Much like a committee of experts deliberating to reach a better decision than a single individual, a model ensemble leverages the strengths of diverse architectures to mitigate errors. This technique is widely used to improve performance in complex tasks, reduce the risk of overfitting, and handle the inherent bias-variance tradeoff found in statistical modeling.

The Mechanics of Ensembling

The core principle behind a model ensemble is "diversity." By training multiple models—often referred to as "base learners" or "weak learners"—on different subsets of training data or using different algorithms, the ensemble ensures that the errors made by one model are likely corrected by others. In the context of deep learning, this often involves running multiple neural networks in parallel during inference.

Common methods for combining these predictions include:

  • Voting: Used in image classification, where the class selected by the majority of models becomes the final prediction.
  • Averaging: Often used in regression tasks, where numerical outputs are averaged to smooth out noise.
  • Weighted Fusion: In object detection, techniques like Weighted Box Fusion (WBF) merge bounding boxes from different detectors based on confidence scores.

실제 애플리케이션

Model ensembles are essential in high-stakes environments where maximizing accuracy is paramount and computational resources allow for running multiple models.

  1. Medical Diagnostics: In medical image analysis, missing a diagnosis can have severe consequences. Radiologists often use ensembles that combine a standard Convolutional Neural Network (CNN) with a Vision Transformer (ViT). The CNN excels at local texture analysis, while the ViT captures global context, allowing the system to detect tumors with higher sensitivity than either architecture alone.
  2. Autonomous Driving: The perception systems in autonomous vehicles must be fail-safe. Engineers frequently deploy an ensemble of detection models—for instance, fusing the real-time speed of YOLO26 with the transformer-based accuracy of RT-DETR. This ensures that pedestrians or obstacles are detected even if one model struggles with specific lighting conditions, such as glare or shadows.

Python으로 앙상블 구현하기

While complex ensemble strategies can be built using libraries like Scikit-learn, you can create a basic inference ensemble for computer vision by simply loading multiple models and processing the same input. The following example demonstrates how to load two distinct Ultralytics YOLO models to generate predictions on the same image.

from ultralytics import YOLO

# Load two different model variants to create a diverse ensemble
model_a = YOLO("yolo26n.pt")  # Nano model (Speed focused)
model_b = YOLO("yolo26s.pt")  # Small model (Higher accuracy)

# Perform inference on an image with both models
# In production, results are typically merged programmatically
results_a = model_a("https://ultralytics.com/images/bus.jpg")
results_b = model_b("https://ultralytics.com/images/bus.jpg")

print(f"Model A detected {len(results_a[0].boxes)} objects.")
print(f"Model B detected {len(results_b[0].boxes)} objects.")

모델 앙상블 대 전문가 혼합

It is helpful to differentiate a standard model ensemble from a Mixture of Experts (MoE), a term often seen in modern Large Language Model (LLM) research.

  • Model Ensemble: Typically queries every model in the collection for every input and aggregates the results. This maximizes metrics like mean average precision (mAP) but significantly increases inference latency and computational cost. It is a brute-force approach to quality.
  • Mixture of Experts: Uses a "gating network" to route data to only a few specific "expert" sub-models best suited for the current input. This allows for massive scalability in foundation models without the computational penalty of running every parameter for every token.

Advantages and Considerations

The primary advantage of using a model ensemble is the boost in performance. Ensembles frequently dominate leaderboards in data science challenges like Kaggle competitions because they can model complex patterns that single models miss. However, this comes with a cost: deploying ensembles requires more memory and compute power.

For teams looking to manage these resource demands efficiently, the Ultralytics Platform offers tools to train, track, and benchmark different model architectures. By comparing performance metrics easily, developers can decide if the accuracy gain from an ensemble justifies the extra infrastructure required for deployment in edge AI scenarios.

Ultralytics 커뮤니티 가입

AI의 미래에 동참하세요. 글로벌 혁신가들과 연결하고, 협력하고, 성장하세요.

지금 참여하기