Cloud Run is a managed compute platform that lets you run containers directly on top of Google's scalable infrastructure.
You can deploy code written in any programming language on Cloud Run if you can build a container image from it. In fact, building container images is optional. If you're using Go, Node.js, Python, Java, .NET Core, Ruby, or a supported framework you can use the source-based deployment option that builds the container for you, using the best practices for the language you're using.
Google has built Cloud Run to work well together with other services on Google Cloud, so you can build full-featured applications.
In short, Cloud Run lets developers spend their time writing their code, and very little time operating, configuring, and scaling their Cloud Run service. You don't have to create a cluster or manage infrastructure to be productive with Cloud Run.
Services, jobs, and worker pools: three ways to run your code
On Cloud Run, your code can run as a service, job, or worker pool. All of these resource types run in the same environment and can use the same integrations with other services on Google Cloud.
The following table provides a high-level look at the options provided by each Cloud Run resource type.
Resource | Description |
---|---|
Service | Responds to HTTP requests sent to a unique and stable endpoint, using stateless, ephemeral instances that autoscale based on a variety of key metrics, also responds to events and functions. |
Job | Handles non-request-based parallelizable tasks that are executed manually, or on a schedule, and run to completion. |
Worker pool | Handles non-request-based workloads such as pull-based workloads, for example, Kafka consumers, Pub/Sub pull queues, or RabbitMQ consumers. |
Cloud Run services
A Cloud Run service provides you with the infrastructure required to run a reliable HTTPS endpoint. Your responsibility is to make sure your code listens on a TCP port and handles HTTP requests.
The following diagram shows a Cloud Run service running several container instances to handle web requests and events from the client using an HTTPS endpoint.
A standard service includes the following features:
- Unique HTTPS endpoint for every service
- Every Cloud Run service has an HTTPS endpoint on a unique subdomain of the
*.run.app
domain – and you can configure custom domains as well. Cloud Run manages TLS for you and supports WebSockets, HTTP/2 (end-to-end), and gRPC (end-to-end). - Fast request-based auto scaling
- Cloud Run rapidly scales out to handle all incoming requests or to handle increased CPU utilization outside requests if the billing setting is set to instance-based billing. A service can rapidly scale out to one thousand instances, or even more if you request a quota increase. If demand decreases, Cloud Run removes idle containers. If you're concerned about costs or overloading downstream systems, you can limit the maximum number of instances.
- Optional manual scaling
- By default, Cloud Run automatically scales to more instances to handle more traffic, but you can override this behavior by using manual scaling to control scaling behavior.
- Built-in traffic management
To reduce the risk of deploying a new revision, Cloud Run supports performing a gradual rollout, including routing incoming traffic to the latest revision, rolling back to a previous revision, and splitting traffic to multiple revisions at the same time.
For example, you can start with sending 1% of requests to a new revision, and increase that percentage while monitoring telemetry.
- Public and private services
A Cloud Run service can be reachable from the internet, or you can restrict access in these ways:
- Specify an access policy using Cloud IAM.
- Use ingress settings to restrict network access. This is useful if you want to allow only internal traffic from the VPC and internal services.
- Allow only authenticated users with Identity-Aware Proxy (IAP).
You can serve cacheable assets from an edge location closer to clients by fronting a Cloud Run service with a Content Delivery Network (CDN), such as Firebase Hosting and Cloud CDN.
Scale to zero and minimum instances
By default, if billing is set to instance-based billing, Cloud Run adds and removes instances automatically to handle all incoming requests or to handle increased CPU utilization outside requests.
If there are no incoming requests to your service, even the last remaining instance will be removed. This behavior is commonly referred to as scale to zero. Then, if there are no active instances when a request comes in, Cloud Run creates a new instance. This increases the response time for these first requests, depending on how fast your container becomes ready to handle requests.
To change this behavior, use one of the following methods:
- Configure Cloud Run to keep a minimum amount of instances active so that your service doesn't scale to zero instances
- Use manual scaling for more control over scaling.
Pay-per-use pricing for services
Scale to zero is attractive for economic reasons since you're charged for the CPU and memory allocated to an instance with a granularity of 100ms. If you don't configure minimum instances, you're not charged if your service is not used. There is a generous free-tier. Refer to pricing for more information.
There are two billing settings you can enable:
- Request-based
- If an instance is not processing requests, you're not charged. You pay a per-request fee.
- Instance-based
- You're charged for the entire lifetime of an instance. There's no per-request fee.
There is a generous free-tier. Refer to pricing for more information, and refer to Billing settings to learn how to enable request-based or instance-based billing for your service.
A disposable container file system
Instances on Cloud Run are disposable. Every container has an in-memory, writable file system overlay, which doesn't persist if the container shuts down. Cloud Run determines when to stop sending request to an instance and shut it down, for example when scaling in.
To receive a warning when Cloud Run is about to shut down an instance,
your application can trap the SIGTERM
signal. This enables your code to flush
local buffers and persist local data to an external datastore.
To persist files permanently, integrate with Cloud Storage or mount a network file system (NFS).
When to use Cloud Run services
Cloud Run services are great for code that handles requests, events, or functions. Example use cases include:
- Websites and web applications
- Build your web app using your favorite stack, access your SQL database, and render dynamic HTML pages.
- APIs and microservices
- You can build a REST API, a GraphQL API, or private microservices communicating over HTTP or gRPC.
- Streaming data processing
- Cloud Run services can receive messages from Pub/Sub push subscriptions and events from Eventarc.
- Asynchronous workloads
- Cloud Run functions can respond to asynchronous events, such as a message on a Pub/Sub topic, a change in a Cloud Storage bucket, or a Firebase event.
- AI inference
- Cloud Run services, with or without GPU configured, can host AI workloads such as inference models and model training.
Cloud Run Jobs
If your code performs work and then stops, for example by using a script, you can use a Cloud Run job to run your code. You can execute a job from the command line by using the Google Cloud CLI, by scheduling a recurring job, or by running it as part of a workflow.
Array jobs are a faster way to run jobs
A job can start a single instance to run your code — that's a common way to run a script or a tool.
However, you can also use an array job, starting many identical, independent instances in parallel. Array jobs are a faster way to process jobs that can be split into multiple independent tasks.
The following diagram shows how a job with seven tasks takes longer run sequentially than the same job when four instances can process independent tasks in parallel:
For example, if you are resizing and cropping 1,000 images from Cloud Storage, processing them consecutively is slower than processing them in parallel with many instances, with Cloud Run managing auto scaling.
When to use Cloud Run jobs
Cloud Run jobs are well-suited to run code that performs work (a job) and quits when the work is done. Here are a few examples:
- Script or tool
- Run a script to perform database migrations or other operational tasks.
- Array job
- Perform highly parallelized processing of all files in a Cloud Storage bucket.
- Scheduled job
- Create and send invoices at regular intervals, or save the results of a database query as XML and upload the file every few hours.
- AI workloads
- Cloud Run jobs with or without GPU configured can host AI workloads such as batch inferencing, fine tuning models, and model training.
Google Cloud integrations
Cloud Run integrates with the broader ecosystem of Google Cloud, which lets you to build full-featured applications.
Essential integrations include:
- Data storage
- Cloud Run integrates with Cloud SQL (managed MySQL, PostgreSQL, and SQL Server), Memorystore (managed Redis and Memcached), Firestore, Spanner, Cloud Storage, and more. Refer to Data storage for a complete list.
- Logging and error reporting
- Cloud Logging automatically ingests container logs. If there are exceptions in the logs, Error Reporting aggregates them, and then notifies you. The following languages are supported: Go, Java, Node.js, PHP, Python, Ruby, and .NET.
- Service identity
- Every Cloud Run revision is linked to a service account, and the Google Cloud client libraries transparently use this service account to authenticate with Google Cloud APIs.
- Continuous delivery
- If you store your source code in GitHub, Bitbucket, or Cloud Source Repositories, you can configure Cloud Run to automatically deploy new commits.
- Private networking
- Cloud Run instances can reach resources in the Virtual Private Cloud network through the Serverless VPC Accessconnector. This is how your service can connect with Compute Engine virtual machines, or products based on Compute Engine such as Google Kubernetes Engine or Memorystore.
- Google Cloud APIs
- Your service's code transparently authenticates with Google Cloud APIs. This includes the AI and Machine Learning APIs, such as the Cloud Vision API, Speech-to-Text API, AutoML Natural Language API, Cloud Translation API, and many more.
- Background tasks
- You can schedule code to run later or immediately after returning a web request. Cloud Run works well together with Cloud Tasks to provide scalable and reliable asynchronous execution.
Refer to Connecting to Google Cloud services for a list of the many Google Cloud services that work well with Cloud Run.
Code must be packaged in a container image
For your service, job, or worker pool to be deployable to Cloud Run, you must package it in a container image. In case you're not familiar with containers, here's a short conceptual introduction.
As the diagram shows, you use the source code, assets, and library dependencies to build the container image, which is a package with everything your service needs to run. That includes build artifacts, assets, system packages, and (optionally) a runtime. This makes a containerized application inherently portable – it runs anywhere a container can run. Examples of build artifacts include compiled binaries or script files, and examples of runtimes are the Node.js JavaScript runtime, or a Java virtual machine (JVM).
Advanced practitioners value the fact that Cloud Run does not impose extra burdens on running their code: you can run any binary on Cloud Run.
If you want more convenience or want to delegate containerizing their application to Google, Cloud Run integrates with the open source Google Cloud's buildpacks to offer a source-based deployment.
What's next
- Deploy a Cloud Run service
- Create and execute a Cloud Run job
- Learn how to execute jobs on a schedule
- Deploy a worker pool
- Explore the resource model
- Read more about the container runtime contract