Sometimes a cluster is terminated unexpectedly, not as a result of a manual termination or a configured automatic termination. A cluster can be terminated for many reasons. Some terminations are initiated by Databricks and others are initiated by the cloud provider. This article describes termination reasons and steps for remediation.
To defend against API abuses, ensure quality of service, and prevent you from
accidentally creating too many large clusters, Databricks throttles all cluster up-sizing requests, including cluster creation, starting, and
resizing. The throttling uses the
token bucket algorithm
to limit the total number of nodes that anyone can launch over a defined
interval across your Databricks deployment, while allowing burst requests of
certain sizes. Requests coming from both the web UI and the APIs are subject to
rate limiting. When cluster requests exceed rate limits, the limit-exceeding request
fails with a
If you hit the limit for your legitimate workflow, Databricks recommends that you do the following:
- Retry your request a few minutes later.
- Spread out your recurring workflow evenly in the planned time frame. For example, instead of scheduling all of your jobs to run at an hourly boundary, try distributing them at different intervals within the hour.
- Consider using clusters with a larger node type and smaller number of nodes.
- Use autoscaling clusters.
If these options don’t work for you, contact Databricks Support to request a limit increase for the core instance.
For other Databricks initiated termination reasons, see Termination Code.
This article lists common cloud provider related termination reasons and remediation steps.
Databricks launches a cluster by requesting resources on behalf of your cloud account. Sometimes, these requests fail because they would exceed your cloud account’s resource limits. In AWS, common error codes include:
AWS limits the number of running instances for each node type. Possible solutions include:
- Request a cluster with fewer nodes.
- Request a cluster with a different node type.
- Ask AWS support to increase instance limits.
The cluster creation request exceeded the EBS volume limit. AWS has two types of volume limits: a limit on the total number of EBS volumes, and a limit on the total storage size of EBS volumes. Potential remediation steps:
- Request a cluster with fewer nodes.
- Check which of the two limits was exceeded. (AWS trusted advisor shows service limits for free). If the request exceeded the total number of EBS volumes, try reducing the requested number of volumes per node. If the request exceeded the total EBS storage size, try reducing the requested storage size and/or the number of EBS volumes.
- Ask AWS support to increase EBS volume limits.
AWS limits the rate of API requests made for an AWS account. Wait a while before retrying the request.
The Spark driver is a single point of failure because it holds all cluster state. If the instance hosting the driver node is shut down, Databricks terminates the cluster. In AWS, common error codes include:
Instance was terminated by a direct request to AWS which did not originate from Databricks. Contact your AWS administrator for more details.
AWS could not satisfy the instance request. Wait a while and retry the request. Contact AWS support if the problem persists.
Instance was terminated by AWS because the current spot price has exceeded the maximum bid made for this instance. Use an on-demand instance for the driver, choose a different availability zone, or specify a higher spot bid price.
For other shutdown-related error codes, refer to AWS docs.
In AWS, common error codes include:
Databricks was not authorized to launch the requested instances. Possible reasons include:
- Your AWS administrator invalidated the AWS access key or IAM role used to launch instances.
- You are trying to launch a cluster using an IAM role that Databricks does not have permission to use. Contact the AWS administrator who set up the IAM role. For more information, see Secure Access to S3 Buckets Using IAM Roles.
Unsupported with message “EBS-optimized instances are not supported for your requested configuration”
The selected instance type is not available in the selected availability zone (AZ). It does not actually have anything to do with EBS-optimization being enabled. To remediate, you can choose a different instance type or AZ.
The provided credentials do not have permission to create the service-linked role for EC2 spot instances. The Databricks administrator needs to update the credentials used to launch instances in your account. Instructions and the updated policy can be found AWS Account.
See Error Codes for a complete list of AWS error codes.
Databricks was able to launch the cluster, but lost the connection to the instance hosting the Spark driver.
Caused by an incorrect networking configuration (for example, changing security group settings for Databricks workers) or a transient AWS networking issue.