Yolo 비전 선전
선전
지금 참여하기
용어집

병합 현실

Explore how merged reality fuses physical and digital worlds using computer vision. Learn how to [deploy YOLO26](https://docs.ultralytics.com/models/yolo26/) for real-time spatial interaction.

Merged Reality (MR), also widely known as Mixed Reality, describes the convergence of the physical world with computer-generated digital content. Unlike strictly virtual or augmented environments, merged reality creates a seamless space where physical and digital objects co-exist and interact in real time. This technology relies heavily on advanced computer vision and spatial computing to map the real-world environment accurately, allowing digital artifacts to be anchored to physical surfaces and respond to physical changes. By leveraging sensors, cameras, and deep learning algorithms, MR systems can understand depth, geometry, and lighting, creating immersive experiences that feel tangible and grounded in the user's actual surroundings.

AI 및 머신 러닝에서의 관련성

The evolution of merged reality is intrinsically linked to advancements in artificial intelligence. To successfully merge digital and physical worlds, a system must possess a sophisticated understanding of the environment. This is where visual perception tasks become critical. Techniques such as object detection allow the system to recognize furniture or people, while SLAM (Simultaneous Localization and Mapping) enables the device to track its own position relative to those objects.

Modern MR applications utilize deep learning models to process complex sensory data instantly. For example, pose estimation is used to track hand movements for gesture control, eliminating the need for physical controllers. Furthermore, semantic segmentation helps the system distinguish between a floor, a wall, and a table, ensuring that a digital character walks on the floor rather than floating through a table.

실제 애플리케이션

Merged reality is transforming industries by enhancing productivity and training through immersive simulations.

  • Industrial Maintenance and Training: In manufacturing, technicians wear MR headsets that overlay digital schematics onto physical machinery. If a worker looks at a specific engine part, the system uses real-time inference to identify the component and display repair instructions or torque specifications directly on the part. This hands-free guidance reduces errors and speeds up complex tasks.
  • Medical Surgery and Planning: Surgeons utilize MR to overlay 3D medical imaging scans (like MRI or CT data) directly onto a patient's body during operations. This allows for precise visualization of internal anatomy without making large incisions. By integrating segmentation models, the system can highlight specific organs or tumors in real-time, aiding in navigation and improving surgical outcomes.

주요 용어 차별화

It is important to distinguish Merged Reality from related concepts in the "XR" (Extended Reality) spectrum:

  • Augmented Reality (AR): AR typically overlays digital information onto a camera feed (like a smartphone filter) without deep spatial interaction. MR goes further by ensuring digital objects interact physically with the real world (e.g., a digital ball bouncing off a real table).
  • Virtual Reality (VR): VR creates a completely synthetic environment, blocking out the physical world entirely. MR keeps the user present in their physical environment while enhancing it.

Applying Computer Vision for MR

To build a basic component of an MR system, such as detecting surfaces or objects to anchor digital content, developers often use high-speed detection models. The Ultralytics YOLO26 model is particularly well-suited for this due to its low latency and high accuracy, which are essential for maintaining the illusion of reality.

The following example demonstrates how to perform instance segmentation on a video stream. In an MR context, this pixel-level mask could define the "walkable" area for a digital character.

from ultralytics import YOLO

# Load the YOLO26 segmentation model
model = YOLO("yolo26n-seg.pt")

# Predict on a video source to identify physical objects and their boundaries
# This data helps anchor digital assets to real-world objects
results = model.predict(source="room_scan.mp4", show=True, stream=True)

for result in results:
    # Process masks to determine occlusion or physics interactions
    if result.masks:
        print(f"Detected {len(result.masks)} physical segments for interaction.")

Future of Merged Reality

As hardware becomes lighter and edge computing capabilities improve, MR is expected to become ubiquitous. The integration of generative AI will likely allow MR environments to populate themselves dynamically, creating digital twins of real-world spaces automatically. With tools like the Ultralytics Platform, developers can easily train custom models to recognize specific objects within these merged environments, pushing the boundaries of how we interact with information in three-dimensional space.

Ultralytics 커뮤니티 가입

AI의 미래에 동참하세요. 글로벌 혁신가들과 연결하고, 협력하고, 성장하세요.

지금 참여하기