Created 05-18-2016 11:40 AM
Guys,
I am referring the document : http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.2/bk_spark-guide/content/spark-kerb-access-hiv... and got a bit confused. Wanted to check with experts who already configured spark thrift server on kerberized environment.
If you are installing the Spark Thrift Server on a Kerberos-secured cluster, note the following requirements:
HiveServer2
, so that it can access the hiveserver2
keytab./var/run/spark
and /var/log/spark
to specify read/write permissions to the Hive service account.thriftserver
process. Thanks.
Created 05-26-2016 06:24 AM
You are going to use hive account to run spark thrift server. So, if it is a manual install, then
./sbin/start-thriftserver.sh --master yarn-client --executor-memory 512m --hiveconf hive.server2.thrift.port=10015
will be run as user hive (with su hive) instead of user spark in secure setup. Similarly /var/run/spark and /var/log/spark should be read/write to hive. So, just seeing contents as user hive is not enough, you need to be able to write to those folders. One good easy way is to give 77x permissions on these folders. Since spark:hadoop is owner:group and hive belongs to group hadoop, it will have write access with this setup.
Created 05-26-2016 06:24 AM
You are going to use hive account to run spark thrift server. So, if it is a manual install, then
./sbin/start-thriftserver.sh --master yarn-client --executor-memory 512m --hiveconf hive.server2.thrift.port=10015
will be run as user hive (with su hive) instead of user spark in secure setup. Similarly /var/run/spark and /var/log/spark should be read/write to hive. So, just seeing contents as user hive is not enough, you need to be able to write to those folders. One good easy way is to give 77x permissions on these folders. Since spark:hadoop is owner:group and hive belongs to group hadoop, it will have write access with this setup.