I am new to spark , I have written a spark application which interacts with Apache phoenix on kerberised environment , and deployed on multiple cluster.
My question is how can we get zookeeper quorum in spark configuration itself to avoid manual steps. Just like retrieving properties of hdfs-site.xml,yarn-site.xml etc from sparkContext.hadoopConfiguration
Also i have seen spark sometimes read its default hbase-site.xml , leading to issue while connecting to phoenix. What are the effective ways of including HBase configuration in spark?
Spark job is invoked using yarn-client mode.