Cargo Containers

3 ways to optimize Cloud Run response times

Serverless containerization has taken the world by storm as it gives developers a way to deploy their stateless microservices without a heavy burden of infrastructure management. Cloud Run abstracts all infrastructure management. You hand over a container image with a web server and stateless logic, and specify a combination of memory/CPU and allowed concurrency.

Cloud Run takes care of creating an HTTP endpoint, routing requests to containers, and scaling containers up and down to handle the volume of requests. While Cloud Run offers some native features to reduce response time latency, such as idle instances, much of it can be improved by writing effective services, which I’ll outline below.

Idle instances

As traffic fluctuates, Cloud Run attempts to reduce the chance of cold starts by keeping some idle instances around to handle spikes in traffic. For example, when a container instance has finished handling requests, it might remain idle for a period of time in case another request needs to be handled.

Google Cloud - Cloud Run Process

But, Cloud Run will terminate unused containers after some time if no requests need to be handled. This means a cold start can still occur. Container instances are scaled as needed, and it will initialize the execution environment completely. While you can keep idle instances permanently available using the min-instance setting, this incurs cost even when the service is not actively serving requests.

So, let’s say you want to minimize both cost and response time latency during a possible cold start. You don’t want to set a minimum number of idle instances, but you also know any additional computation needed upon container startup before it can start listening to requests means longer load times and latency.

Cloud Run container startup

There are a few tricks you can do to optimize your service for container startup times. The goal here is to minimize the latency that delays a container instance from serving requests. But first, let’s review the Cloud Run container startup routine.

At a high level, it consists of:

  1. Starting the service
    • Starting the container
    • Running the entrypoint command to start your server
  2. Checking for the open service port

You want to tune your service in order to minimize the time needed for step 1a. Let’s walk through 3 ways to optimize your service for Cloud Run response times.

#1 Create a leaner service

You want to be extra careful when it comes to applications written in dynamic languages. For example, if you’re using Node.js or Python, module loading that happens on process startup will add latency during a cold start.

Google Cloud - Optimized Cloud Run

Also be aware of some modules that run initialization code upon importing.

Google Cloud - Optimized Cloud Run 2

To build a leaner service you can:

  • Minimize the number and size of dependencies if you’re using a dynamic language.
  • Instead of computing things upon startup, compute them lazily. The initialization of global variables always occurs during startup, which increases cold start time. Use lazy initialization for infrequently used objects to defer the time cost and decrease cold start times.
  • Shorten your initializations and speed up time to start your HTTP server.
  • And use code-loading optimizations like PHP’s composer autoloader optimization.

#2 Use a leaner base image

For starters, on Cloud Run, the size of your container image does not affect cold start or request processing time. Cold starts aren’t affected by the size of the image, but by the image system complexity and initialization time. For example, you might have a base image that has 12 slow-to-start OS-level packages, and that will increase latency during a cold start. On the other hand, you could have a super simple base image, into which you copy a 500GB datafile, and that larger image won’t add to startup times.

You want to build a minimal container by working off a lean base image like: alpine, distroless.

You can also use, scratch, which is an empty image on which you can build your own runtime environment. If your app is a statically linked binary, it’s easy to use the scratch base image:

FROM scratch

COPY mybinary /mybinary
CMD [ "/mybinary" ]

These images also happen to be smaller in size, which tend to be faster to work with in other parts of your workflow, like local development or build systems. For example, the alpine:3.7 image is 71 MB smaller than the centos:7 image.

You should also only install what is strictly needed inside the image. In other words, don’t install extra packages that you don’t need.

#3 Use global variables

In Cloud Run, you can’t assume that service state is preserved between requests. But, Cloud Run does reuse individual container instances to serve ongoing traffic.

That means you can declare a global variable. Once a container is spun up, objects that are initialized and stored in a global variable in memory can be reused between requests. Moving this from the request logic to global scope means better performance when traffic is ongoing. Now this doesn’t exactly help cold start times, but once the container is initialized, cached objects can help reduce latency during subsequent ongoing requests.

For example, if you move per-request logic to global scope, it should make a cold starts last approximately the same amount of time (and if you add extra logic for caching that you wouldn’t have in a warm request, it would increase the cold start time), but any subsequent request served by that warm instance will have a lower latency.

// Global (instance-wide) scope
// This computation runs at instance cold-start
const instanceVar = heavyComputation();
* HTTP function that declares a variable.

* @param {Object} req request context.
* @param {Object} res response context.


exports.scopeDemo = (req, res) => {

// Per-function scope
// This computation runs every time this function is called

const functionVar = lightComputation();
res.send(`Per instance: ${instanceVar}, per function: ${functionVar}`);

One option that can help with cold starts is to offload global state to an in-memory datastore like Memorystore, which provides very low latency data access to application caches.


A lot of this boils down to creating a leaner service so logic that computes during container initialization is minimized, and it can start serving requests as soon as possible. While these are just a few best practices for designing a Cloud Run service, there are a number of other tips for writing effective services and optimizing performance, which you can read about here.

For more cloud content follow me on Twitter @swongful.

By Stephanie Wong. Source: Google Cloud Blog.

For enquiries, product placements, sponsorships, and collaborations, connect with us at We'd love to hear from you!

Our humans need coffee too! Your support is highly appreciated, thank you!

Previous Article
Google Cloud - Databases

Forbes embraces MongoDB on Google Cloud as part of digital-first strategy

Next Article
Google Cloud - Security AI

The cloud trust paradox: To trust cloud computing more, you need the ability to trust it less

Related Posts