You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Dec 4, 2024. It is now read-only.
It seems that the resource allocation for apache spark jobs on DC/OS is a bit murky.
For example launching a job with:
dcos spark run --submit-args='http://foobar/prog.py'
will allocate lots of cluster resources but, when checking top you will only see a couple of processes chewing CPU. In order to get any power out of the job you need to tell the driver how much parallelism to use with spark.default.parallelism
dcos spark run --submit-args='-Dspark.default.parallelism=50 http://foobar/prog.py'
At least to me its a bit opaque what is going on here and the dcos abstraction from spark-submit does not make this any easier.
The pyspark application I am running takes a list and then does the following on it. As the driver does not know what is in the list nor able to infer it from a call to a partition maybe it does not know how many executors to use / start?
It seems that the resource allocation for apache spark jobs on DC/OS is a bit murky.
For example launching a job with:
dcos spark run --submit-args='http://foobar/prog.py'
will allocate lots of cluster resources but, when checking top you will only see a couple of processes chewing CPU. In order to get any power out of the job you need to tell the driver how much parallelism to use with spark.default.parallelism
dcos spark run --submit-args='-Dspark.default.parallelism=50 http://foobar/prog.py'
At least to me its a bit opaque what is going on here and the dcos abstraction from spark-submit does not make this any easier.
The pyspark application I am running takes a list and then does the following on it. As the driver does not know what is in the list nor able to infer it from a call to a partition maybe it does not know how many executors to use / start?
The program:
A sample of the test data in S3. One file for every day in Unix time (1970 - 2038).
The text was updated successfully, but these errors were encountered: