Otonom Araçlar
Otonom araçların, güvenlik, verimlilik ve yenilik ile ulaşımda devrim yaratmak için yapay zekayı, bilgisayarlı görü'yü ve sensörleri nasıl kullandığını keşfedin.
Autonomous Vehicles (AVs), frequently referred to as self-driving cars, are intelligent transportation systems capable
of sensing their environment and operating without human involvement. These systems represent the pinnacle of
AI in automotive innovation, combining
sophisticated hardware with advanced software algorithms to interpret complex surroundings. The primary objective of
AV technology is to enhance road safety by minimizing accidents caused by human error, while also optimizing traffic
efficiency and providing mobility to those unable to drive. At their core, these vehicles rely on
artificial intelligence (AI) to perceive
stimuli, process information, and make split-second driving decisions.
Perception And Sensor Technologies
For an autonomous vehicle to navigate safely, it must possess a comprehensive understanding of its surroundings. This
is achieved through a perception layer that aggregates data from a suite of sensors.
-
Computer Vision (CV): Cameras serve as the primary visual sensors, mimicking human sight. Algorithms process video feeds to recognize
lane markings, traffic lights, and signs.
-
LiDAR Technology: Light Detection and Ranging (LiDAR) uses laser pulses to create precise, high-resolution 3D maps of the
environment, essential for depth perception.
-
Object Detection: Deep learning models identify and localize dynamic obstacles. High-speed models like
YOLO26 are crucial here for detecting pedestrians and
other vehicles with low latency.
-
Sensor Fusion: No single sensor is perfect in all conditions (e.g., cameras in fog). Fusion
algorithms combine data from cameras, radar, and LiDAR to form a robust environmental model.
-
Semantic Segmentation: This technique classifies every pixel in an image, helping the vehicle distinguish between the drivable road
surface, sidewalks, and vegetation.
Levels Of Autonomy
The capabilities of autonomous systems are categorized by the
SAE J3016 levels of driving automation, which define the
extent of computer control versus human intervention.
-
Advanced Driver Assistance Systems (ADAS): Covering Levels 1 and 2, these systems assist with steering or acceleration (e.g., adaptive cruise control) but
require the driver to remain engaged.
-
Conditional Automation: At Level 3, the vehicle can handle most driving tasks in specific conditions, such as highway traffic jams, but
the human must be ready to take over when alerted.
-
High And Full Automation: Levels 4 and 5 represent vehicles that can operate without human input. Level 4 is limited to geo-fenced areas,
while Level 5 aims for full autonomy on any road, often requiring powerful
Edge AI hardware.
Gerçek Dünya Yapay Zeka Uygulamaları
Autonomous vehicle technology is currently being deployed across various sectors, relying on heavy
machine learning (ML) computation to handle
real-world complexity.
-
Robotaxis: Companies like Waymo utilize fleets of fully
autonomous vehicles to transport passengers in urban environments. These vehicles use
predictive modeling to anticipate the
behavior of pedestrians and other drivers in complex cityscapes.
-
Autonomous Trucking: Long-haul logistics benefit from automation on predictable highway routes.
Innovators like Aurora develop self-driving trucks that leverage long-range
perception to improve fuel efficiency and safety.
-
Last-Mile Delivery: Small, autonomous robots use
object tracking to navigate sidewalks and deliver packages,
reducing the cost and carbon footprint of logistics.
İlgili Kavramları Ayırt Etme
It is important to differentiate Autonomous Vehicles from related terms in the robotics and automotive fields.
-
Vs. Robotics: While AVs are technically mobile robots, the field of robotics is broader, encompassing stationary industrial
arms and humanoid assistants. AVs are specifically specialized for transportation logic.
-
Vs.
Connected Vehicles (V2X): Connected vehicles communicate with each other (V2V) and infrastructure (V2I) to share data like speed and
location. A vehicle can be connected without being autonomous, though connectivity often enhances the safety of AVs.
-
Vs.
Teleoperation: Teleoperation involves a human remotely driving a vehicle. In contrast, true AVs rely on onboard
neural networks to make decisions locally.
YOLO26 ile Algıyı Uygulama
A critical component of any autonomous system is the ability to track objects over time. The following example
demonstrates how to use the Ultralytics Platformu compatible
ultralytics library to perform object tracking on a video, simulating a vehicle's perception system.
from ultralytics import YOLO
# Load the YOLO26 model, optimized for speed and accuracy
model = YOLO("yolo26n.pt")
# Track vehicles and pedestrians in a video stream
# This simulates the continuous perception required by an AV
results = model.track(
source="https://www.ultralytics.com/blog/ultralytics-yolov8-for-speed-estimation-in-computer-vision-projects",
show=True,
)
# Process results (e.g., counting objects or estimating speed)
for r in results:
print(r.boxes.xywh) # Print bounding box coordinates