Support Questions
Find answers, ask questions, and share your expertise
Alert: Welcome to the Unified Cloudera Community. Former HCC members be sure to read and learn how to activate your account here.

Hive quit logging with installed Spark


Hive quit logging with installed Spark

New Contributor
Hello together,

I have a very mysterious problem. On a Hive-Server without Spark, hive.log will be written.
As soon as I install Spark, hive.log is no longer written.
The reason for this is the environment variable SPARK_HOME. If I'm taking the following block
out of /usr/lib/hive/bin/hive, then hive.log is written.

if [[ -z "$SPARK_HOME" ]]
bin=`dirname "$0"`
# many hadoop installs are in dir/{spark,hive,hadoop,..}
sparkHome=$(readlink -f $bin/../../spark)
if [[ -d $sparkHome ]]
export SPARK_HOME=$sparkHome

Same problem if I set SPARK_HOME manually.

Does somebody still have an idea or hint? Many thanks in advance!

Runnin CDH Version is 5.7.4 Regards, Daniel


Don't have an account?
Coming from Hortonworks? Activate your account here