I try to follow the below guide from GCP to create a data quality task. https://cloud.google.com/dataplex/docs/check-data-quality?&_ga=2.139058355.-143783534.1669209779#before_you_begin when I try to run the task I get the error
- As per my understanding Dataplex is calling dataproc internally for running the spark jobs due to which we get the error message. But we dont have the option to adjust the number of CPUs that can be used.
- I tried to increase the CPU quota but GCP does not allow me( My account is enabled for billing for a long time) Any suggestion to correct this error will be appreciated.
As mentioned by @guillaume, there is set a limit on the number of CPUs you can use for a region (and all regions) in GCP. It is different for individuals and organizations. You can check this under IAM -> Quotas-> Compute Engine API CPU quota.
Solutions to your error:
You can request a quota increase on the number of CPUs for a region/all regions for your project.
Try changing the configuration of your Dataproc Cluster by decreasing the number of secondary workers. Reference