Maximum concurrent requests for services

For Cloud Run services, each revision is automatically scaled to the number of instances needed to handle all incoming requests.

When more instances are processing requests, more CPU and memory will be used, resulting in higher costs.

To give you more control, Cloud Run provides a maximum concurrent requests per instance setting that specifies the maximum number of requests that can be processed simultaneously by a given instance.

Maximum concurrent requests per instance

You can configure the maximum concurrent requests per instance. By default each Cloud Run instance can receive up to 80 requests at the same time; you can increase this to a maximum of 1000.

Although you should use the default value, if needed you can lower the maximum concurrency. For example, if your code cannot process parallel requests, set concurrency to 1.

The specified concurrency value is a maximum limit. If the CPU of the instance is already highly utilized, Cloud Run might not send as many requests to a given instance. In these cases, the Cloud Run instance might show that the maximum concurrency is not being utilized. For example, if the high CPU usage is sustained, the number of instances might scale up instead.

The following diagram shows how the maximum concurrent requests per instance setting affects the number of instances needed to handle incoming concurrent requests:

maximum concurrent requests per instance diagram

Tuning concurrency for autoscaling and resource utilization

Adjusting the maximum concurrency per instance significantly influences how your service scales and utilizes resources.

  • Lower concurrency: Forces Cloud Run to use more instances for the same request volume, because each instance handles fewer requests. This can improve responsiveness for applications that are not optimized for high internal parallelism or for applications you want to scale more quickly based on request load.
  • Higher concurrency: Allows each instance to handle more requests, potentially leading to fewer active instances and reducing cost. This is suitable for applications efficient at parallel I/O-bound tasks or for applications that can truly utilize multiple vCPUs for concurrent request processing.

Start with the default concurrency (80) , monitor the performance and utilization of your application closely, and adjust as needed.

Concurrency with multi-vCPU instances

Tuning concurrency is especially critical if your service uses multiple vCPUs but your application is single-threaded or effectively single-threaded (CPU-bound).

  • vCPU hotspots: A single-threaded application on a multi-vCPU instance may max out one vCPU while others idle. The Cloud Run CPU autoscaler measures average CPU utilization across all vCPUs. The average CPU utilization can remain deceptively low in this scenario, preventing effective CPU-based scaling.
  • Using concurrency to drive scaling: If CPU-based autoscaling is ineffective due to vCPU hotspots, lowering maximum concurrency becomes an important tool. vCPU hotspots often occur where multi-vCPU is chosen for a single-threaded application due to high memory needs. Using concurrency to drive scaling forces scaling based on request throughput. This ensures that more instances are started to handle the load, reducing per-instance queuing and latency.

When to limit maximum concurrency to one request at a time.

You can limit concurrency so that only one request at a time will be sent to each running instance. You should consider doing this in cases where:

  • Each request uses most of the available CPU or memory.
  • Your container image is not designed for handling multiple requests at the same time, for example, if your container relies on global state that two requests cannot share.

Note that a concurrency of 1 is likely to negatively affect scaling performance, because many instances will have to start up to handle a spike in incoming requests. See Throughput versus latency versus tradeoffs for more considerations.

Case study

The following metrics show a use case where 400 clients are making 3 requests per second to a Cloud Run service that is set to a maximum concurrent requests per instance of 1. The green top line shows the requests over time, the bottom blue line shows the number of instances started to handle the requests.

Concurrency set to one

The following metrics show 400 clients making 3 requests per second to a Cloud Run service that is set to a maximum concurrent requests per instance of 80. The green top line shows the requests over time, the bottom blue line shows the number of instances started to handle the requests. Notice that far fewer instances are needed to handle the same request volume.

Concurrency set to 80

Concurrency for source code deployments

When concurrency is enabled, Cloud Run does not provide isolation between concurrent requests processed by the same instance. In such cases, you must ensure that your code is safe to execute concurrently. You can change this by setting a different concurrency value. We recommend starting with a lower concurrency like 8, and then moving it up. Starting with a concurrency that is too high could lead to unintended behavior due to resource constraints (such as memory or CPU).

Language runtimes can also impact concurrency. Some of these language-specific impacts are shown in the following list:

  • Node.js is inherently single-threaded. To take advantage of concurrency, use JavaScript's asynchronous code style, which is idiomatic in Node.js. See Asynchronous flow control in the official Node.js documentation for details.

  • For Python 3.8 and later, supporting high concurrency per instance requires enough threads to handle the concurrency. We recommend that you set a runtime environment variable so that the threads value is equal to the concurrency value, for example: THREADS=8.

What's next

To manage the maximum concurrent requests per instance of your Cloud Run services, see Setting maximum concurrent requests per instance.

To optimize your maximum concurrent requests per instance setting, see development tips for tuning concurrency.