Setting directory for Spark recovery

0 votes
Hello guys! I need help with my Spark application, especially with the recovery part. I have set the recovery mode to Zookeeper and have set the URL. Everything is fine and the connection is also happening as expected. I am stuck with the last part. I want to set the directory where the recovery state is stored. Please mention the commands to set this directory. Thanks
Mar 25, 2019 in Apache Spark by Rahul
1,031 views

1 answer to this question.

0 votes

You can set the directory to store the recovery state dynamically. Here is the command you can use to do it:

val sc = new SparkContext(new SparkConf())

./bin/spark-submit <all your existing options> --spark.deploy.zookeeper.dir=<path to directory>
answered Mar 25, 2019 by Hari

Related Questions In Apache Spark

0 votes
1 answer

How to connect to Zookeeper after setting Spark recovery mode?

You have set Zookeeper as the recovery ...READ MORE

answered Mar 25, 2019 in Apache Spark by Hari
1,399 views
0 votes
1 answer
0 votes
1 answer

Hadoop mandatory for Spark?

No not mandatory, but there is no ...READ MORE

answered Jun 1, 2018 in Apache Spark by kurt_cobain
• 9,350 points
578 views
0 votes
2 answers

Which cluster type should I choose for Spark?

Spark is agnostic to the underlying cluster ...READ MORE

answered Aug 21, 2018 in Apache Spark by zombie
• 3,790 points
2,022 views
+1 vote
2 answers
+1 vote
1 answer

Hadoop Mapreduce word count Program

Firstly you need to understand the concept ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 11,380 points
11,028 views
0 votes
1 answer

hadoop.mapred vs hadoop.mapreduce?

org.apache.hadoop.mapred is the Old API  org.apache.hadoop.mapreduce is the ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 11,380 points
2,536 views
+2 votes
11 answers

hadoop fs -put command?

Hi, You can create one directory in HDFS ...READ MORE

answered Mar 16, 2018 in Big Data Hadoop by nitinrawat895
• 11,380 points
108,832 views
0 votes
1 answer

Setting port for SSL for Spark application

You can set the port in the ...READ MORE

answered Mar 15, 2019 in Apache Spark by Karan
678 views
+1 vote
3 answers

Which cluster type should I choose for Spark?

According to me, start with a standalone ...READ MORE

answered Jun 27, 2018 in Apache Spark by nitinrawat895
• 11,380 points
1,568 views
webinar REGISTER FOR FREE WEBINAR X
REGISTER NOW
webinar_success Thank you for registering Join Edureka Meetup community for 100+ Free Webinars each month JOIN MEETUP GROUP