Member since
12-18-2018
8
Posts
0
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1909 | 04-03-2019 12:11 PM |
04-03-2019
12:11 PM
Issue resolved. In case anyone needs to add another spark history server, so in order to control clients' config regarding spark history server address: In Ambari >>> spark >>> config >>> spark.yarn.history.addr Replace the variables {{spark_history_server_host}}:{{spark_history_ui_port}} with hardcoded server address and port for each config group. Do NOT use the variables.
... View more
03-28-2019
02:01 PM
Hi experts The scenario is this: Cluster with one Spark History Server on host Y. Using curl I added another Spark History Server to a host X. I also created a different config group for the new SHS, with it's own specific spark.yarn.history.address with host X The goal is that all clients will write their logs to the default SHS on host Y and whoever needs the new SHS will have to set it manually in his spark app arguments. Problem is, after adding the new SHS and performing "refresh configs" for spark-clients all of them are getting the new SHS address in their /etc/spark2/conf/spark-defaults.conf (spark.yarn.historyServer.address = X:18081) Where do i set which spark.yarn.historyServer.address the clients will be using ? Thx Adi
... View more
Labels:
- Labels:
-
Apache Spark
12-25-2018
07:55 AM
Thank you both for your insights! Using Ranger i can mask data in hive but it's different than actual encryption so Ranger as-is is not a solution for my scenario. @Raymond Honderdors thank you for that flow, i will give it a deeper thought.
... View more
12-24-2018
05:16 PM
I'm looking for a way to encrypt a column in hive. I know that using kms I can encrypt entire tables and/or data on hdfs, but looking for an efficient way to encrypt a specific column. I came across this https://issues.apache.org/jira/plugins/servlet/mobile#issue/HIVE-8049 Has anyone tried this or familiar with encrypting a column in hive? Thanks in advance
... View more
Labels:
12-18-2018
04:00 PM
Hello I'm looking for a smart solution for monitoring hdfs folder quotas. I know how to get the quota of a single directory (hadoop fs -count -q /path/to/directory), and can also do a recursive script, but on very large scale hdfs it is not efficient. Has anyone used or know of a smart \ efficient solution for this ? Or a way to show all folders that have quotas ? Thanks in advance Adi J.
... View more
Labels:
- Labels:
-
Apache Hadoop