Managed Service for Apache Spark quotas

This page lists Managed Service for Apache Spark API quota limits, which are enforced at the project and region level. The quotas reset every sixty seconds (one-minute).

For cluster optimization strategies to help avoid quota and resource availability issues, see Resource availability and zone strategies.

The following table lists the specific and default per-project Managed Service for Apache Spark API quota types, quota limits, and the methods to which they apply.

Quota Type Limit Applicable API Methods
AutoscalingOperationRequestsPerMinutePerProjectPerRegion 400 CreateAutoscalingPolicy, GetAutoscalingPolicy, ListAutoscalingPolicies, UpdateAutoscalingPolicy, DeleteAutoscalingPolicy
ClusterOperationRequestsPerMinutePerProjectPerRegion 200 CreateCluster, DeleteCluster, UpdateCluster, StopCluster, StartCluster, DiagnoseCluster, RepairCluster
NodeGroupOperationRequestsPerMinutePerProjectPerRegion 600 CreateNodeGroup, DeleteNodeGroup, ResizeNodeGroup, RepairNodeGroup, UpdateLabelsNodeGroup, StartNodeGroup, StopNodeGroup
GetJobRequestsPerMinutePerProjectPerRegion 7500 GetJob
JobOperationRequestsPerMinutePerProjectPerRegion 400 SubmitJob, UpdateJob, CancelJob, DeleteJob
WorkflowOperationRequestsPerMinutePerProjectPerRegion 400 CreateWorkflowTemplate, InstantiateWorkflowTemplate, InstantiateInlineWorkflowTemplate, UpdateWorkflowTemplate, DeleteWorkflowTemplate
DefaultRequestsPerMinutePerProjectPerRegion 7500 All other operations (primarily Get operations)

The following table lists additional limits on total active operations and jobs at the project and region level.

Quota type Limit Description
ActiveOperationsPerProjectPerRegion 5000 Limit on the total number of concurrent active operations of all types in a single project in a single regional database
ActiveJobsPerProjectPerRegion 5000 Limit on the total number of active jobs in NON_TERMINAL state in a single project in a single regional database

Other Google Cloud quotas

Managed Service for Apache Spark clusters use other Google Cloud products. These products have project-level quotas, which include quotas that apply to Managed Service for Apache Spark use. Some services are required to use Managed Service for Apache Spark, such as Compute Engine and Cloud Storage. Other services, such as BigQuery and Bigtable, can optionally use Managed Service for Apache Spark.

Required cluster services

The following services, which enforce quota limits, are required to create Managed Service for Apache Spark clusters.

Compute Engine

Managed Service for Apache Spark clusters use Compute Engine virtual machines. The Compute Engine quotas are split into regional and global limits. These limits apply to clusters that you create. For example, the creation of a cluster with one n1-standard-4 -m node and two n1-standard-4 -w nodes uses 12 virtual CPUs (4 * 3). This cluster usage counts against the regional quota limit of 24 virtual CPUs.

Default clusters resources

When you create a Managed Service for Apache Spark cluster with default settings, the following Compute Engine resources are used.

Resource Usage
Virtual CPUs 12
Virtual Machine (VM) Instances 3
Persistent disk 1500 GB

Cloud Logging

Managed Service for Apache Spark saves driver output and cluster logs in Cloud Logging. The Logging quota applies to Managed Service for Apache Spark clusters.

Optional cluster services

You can optionally use the following services, which have quota limits, with Managed Service for Apache Spark clusters.

BigQuery

When reading or writing data into BigQuery, the BigQuery quota applies.

Bigtable

When reading or writing data into Bigtable, the Bigtable quota applies.

Resource availability and zone strategies

To optimize clusters for resource availability and mitigate potential stockout errors, consider the following strategies:

  • Auto Zone Placement: When creating clusters, use auto zone placement. This allows Managed Service for Apache Spark to select an optimal zone within your specified region, improving the chances of successful cluster creation.

  • Regional quotas: Verify your regional Compute Engine quotas are sufficient since quotas can be exhausted even with auto zone placement if the total regional capacity is insufficient for your requests.

  • Machine type flexibility: If you experience persistent stockouts with a specific machine type, use a different, more readily available machine type for your cluster.