Glossary

Cloud Computing

Discover the power of cloud computing for AI/ML! Scale efficiently, train Ultralytics YOLO models faster, and deploy seamlessly with cost-effectiveness.

Cloud computing is the on-demand delivery of computing services—including servers, storage, databases, networking, software, analytics, and intelligence—over the Internet ("the cloud"). Instead of owning and maintaining their own computing infrastructure, organizations can access these services from a cloud provider like Amazon Web Services (AWS), Google Cloud, or Microsoft Azure. This model allows for faster innovation, flexible resources, and economies of scale, making it an essential foundation for modern Artificial Intelligence (AI) and Machine Learning (ML). The core idea, as defined by the National Institute of Standards and Technology (NIST), is to provide ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources.

How Cloud Computing Works

Cloud providers maintain a global network of data centers with massive amounts of hardware. They offer services through different models, the most common being:

  • Infrastructure as a Service (IaaS): Provides fundamental computing resources like virtual machines, storage, and networking. This gives users maximum control and is ideal for custom deep learning environments.
  • Platform as a Service (PaaS): Offers a platform allowing customers to develop, run, and manage applications without the complexity of building and maintaining the infrastructure. This includes managed databases and Kubernetes services.
  • Software as a Service (SaaS): Delivers software applications over the internet on a subscription basis. Ultralytics HUB is an example of a SaaS platform that provides tools for training and managing computer vision models.

This structure enables key benefits like cost savings, global scalability, high performance, and enhanced data security, which are managed in partnership with organizations like the Cloud Security Alliance (CSA).

Importance in AI and Machine Learning

The cloud is the primary engine for AI development today. Training advanced models, like Ultralytics YOLO, requires immense computational power and data, which is often impractical to host locally.

Key uses include:

  • Training Powerful Models: The cloud provides access to high-performance hardware like GPUs and TPUs necessary for distributed training on large datasets. Platforms like Ultralytics HUB Cloud Training leverage this to accelerate model development.
  • Managing Large Datasets: AI models are trained on vast amounts of training data. Cloud storage solutions provide scalable and accessible repositories for these datasets, from ImageNet to custom collections for specific tasks like object detection.
  • Scalable Model Deployment: Once a model is trained, it can be deployed to the cloud for real-time inference. The cloud's elastic nature allows applications to automatically scale to handle fluctuating demand, a core principle of MLOps. You can learn more about different model deployment options in our documentation.

Real-World Applications

  1. AI in Automotive: Companies developing autonomous vehicles collect petabytes of driving data. They use cloud-based GPU clusters to train and validate perception models that can identify pedestrians, vehicles, and road signs, a process detailed in our AI in Automotive solutions page.
  2. AI in Healthcare: A research hospital might use a secure, HIPAA-compliant cloud environment to train a diagnostic model for medical image analysis. By pooling anonymized data, they can build a robust model using a framework like PyTorch to detect anomalies in X-rays or MRIs, leading to faster and more accurate diagnoses for improved AI in healthcare.

Cloud Computing Vs. Related Concepts

  • Serverless Computing: Serverless computing is an execution model within cloud computing, not an alternative to it. While broader cloud computing may involve managing virtual servers (IaaS), serverless abstracts away all server management. You simply provide code (as functions), and the cloud provider automatically provisions resources to run it, scaling from zero to massive volumes as needed.
  • Edge Computing: Edge computing involves processing data locally on devices at the "edge" of the network, close to the data source. This is the opposite of the centralized model of cloud computing. However, they are often used together in a hybrid approach. For instance, an Edge AI device like an NVIDIA Jetson might perform initial object detection and then send only relevant metadata to the cloud for long-term storage, aggregation, or more intensive analysis. This approach combines the low latency of the edge with the massive power of the cloud. You can find more insights on our blog about deploying applications on edge devices.

Join the Ultralytics community

Join the future of AI. Connect, collaborate, and grow with global innovators

Join now
Link copied to clipboard