Member since
09-22-2017
38
Posts
0
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
5957 | 11-16-2019 10:29 AM |
05-21-2025
11:18 AM
Spark job is writing event log to hdfs://namenode:8021/spark-history , but our job creates many events within 10 to 12 minutes. As per the document spark.eventLog.buffer.kb =100kb Does it mean when event writes to /spark-history/application_xxxxxx_xx/xx , whenever buffer gets 100kb? It means every time it is going to call fsync? spark.eventLog.buffer.kb 100k Buffer size to use when writing to output streams, in KiB unless otherwise specified.
... View more
Labels:
- Labels:
-
Apache Spark
12-01-2021
07:01 AM
I have seen some of my jobs are using only one reducer in the end. Does it mean we have one output file? Or it depends?
... View more
Labels:
- Labels:
-
Apache Hive
12-01-2021
06:07 AM
Our user is running a job and which is a hive query and number of mapper is always 6 and not changing even the data size change. It is a insert query. How do I change number of mappers ? Which parameter determine number of mappers?
... View more
Labels:
- Labels:
-
Apache Hive
10-13-2021
04:58 PM
I can do query through spark using spark.sql, but when I tried via beeline ranger is blocking the access because of the policy. How does it work via spark query? Doesn't it check HDFS level permission? I
... View more
Labels:
- Labels:
-
Apache Ranger
-
Apache Spark
-
HDFS
11-16-2019
05:02 PM
Thank you for the info. Yes, I have created with backup with another directory and I was about to boot restart the namenode from that image.
... View more
11-16-2019
10:29 AM
Thank you for the information. If I have multiple directories configured for dfs.namenode.name.dir. Defualt: /hadoop/hdfs/namenode I can configure /backup/namenode /goldbackup/namenode In this case I will have multiple copy of FSimage and Editlog and in case one the directory is corrupted, I can use another one. Does it make sense?
... View more
11-15-2019
03:35 PM
Hello,
I have a basic question about about namenode format. If we perform namenode format, will clean up all the metadata info in namenode like directory structure, etc. After it completed, if we restart namenode, would it rebuild the directory structure from datanode information? Or we are lost.
... View more
Labels:
- Labels:
-
Apache Hadoop
05-25-2018
01:24 PM
Hello, When NN writes data(block) to DN or when DN send block over to next DN, what protocol they use? RPC or HTTP?
... View more
Labels:
- Labels:
-
Apache Hadoop
05-24-2018
02:31 PM
Thanks a lot for giving great explanation. Just a last question, delegation token is given by NN, from NN what package or what tool provides the delegation token. Is it part of namenode package?
... View more