Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for Spark 2.1.0's native dispatcher failover #131

Open
christek91 opened this issue Mar 21, 2017 · 0 comments
Open

Add support for Spark 2.1.0's native dispatcher failover #131

christek91 opened this issue Mar 21, 2017 · 0 comments

Comments

@christek91
Copy link

According to the Spark documentation (http://spark.apache.org/docs/latest/running-on-mesos.html#cluster-mode), Spark 2.1.0 can write recovery state data to Zookeeper, which can then be read by new spark dispatchers during a dispatcher failover, or when restarting the service in Mesosphere.

Currently, if you restart a Spark dispatcher service in Mesosphere the previous dispatcher's job history is lost. It would be nice to persist this data so that it works similarly to the History Server (which persists its data in HDFS or elsewhere) after a restart.

Since the Spark package already does some zookeeper configuration (

export SPARK_DAEMON_JAVA_OPTS="$SPARK_DAEMON_JAVA_OPTS -Dspark.deploy.zookeeper.dir=/spark_mesos_dispatcher_${DCOS_SERVICE_NAME}"
)
this may only be a matter of exposing the spark.deploy.recoveryMode conf value through an env variable in the image and a config.json field.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant