spark.databricks.aggressiveWindowDown 600
spark.databricks.aggressiveWindowDown 600
The Spark configuration property spark.databricks.aggressiveWindowDown
is used in Databricks clusters to control the frequency with which the cluster evaluates whether to scale down the number of workers.
Setting it to 600 (the maximum value) means:
- The cluster will check for potential downscaling opportunities every 600 seconds (10 minutes).
- This will slow down the downscaling process compared to lower values.
- It can be beneficial if your workloads have periods of high activity followed by lulls, as it prevents premature downscaling during brief quiet periods.
Considerations:
- Cost: Keeping workers running longer can increase costs, especially if the cluster remains underutilized for extended periods.
- Workload: The ideal value depends on the nature of your workloads. If you have consistent workloads, a lower value might be more efficient.
How to Set It:
You can set this configuration property in your Databricks cluster settings or through your job’s configuration.
Databricks Training Demo Day 1 Video:
Conclusion:
Unogeeks is the No.1 IT Training Institute for Databricks Training. Anyone Disagree? Please drop in a comment
You can check out our other latest blogs on Databricks Training here – Databricks Blogs
Please check out our Best In Class Databricks Training Details here – Databricks Training
Follow & Connect with us:
———————————-
For Training inquiries:
Call/Whatsapp: +91 73960 33555
Mail us at: info@unogeeks.com
Our Website ➜ https://unogeeks.com
Follow us:
Instagram: https://www.instagram.com/unogeeks
Facebook:https://www.facebook.com/UnogeeksSoftwareTrainingInstitute
Twitter: https://twitter.com/unogeeks