Do I need to restart nodes if i am running spark on yarn after changing spark-env.sh or spark-defaults?

442 Views Asked by At

I am working on changing conf for spark in order to limit the logs for my spark structured streaming log files. I have figured the properties to do so, but it is not working right now. Do i need to restart all nodes (name and worker nodes) or is restarting the jobs is enough. We are using google dataproc clusters and running spark with yarn .

1

There are 1 best solutions below

0
On

The simplest will be to set these properties during cluster creation time using Dataproc Cluster Properties:

gcloud dataproc clusters create $CLUSTER_NAME \
  --properties spark:<key>=<value>,yarn:<key>=<value>

Or set them when submitting your Spark application.