Member since
08-08-2017
1652
Posts
30
Kudos Received
11
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2003 | 06-15-2020 05:23 AM | |
| 16496 | 01-30-2020 08:04 PM | |
| 2155 | 07-07-2019 09:06 PM | |
| 8369 | 01-27-2018 10:17 PM | |
| 4744 | 12-31-2017 10:12 PM |
09-30-2018
06:54 AM
hi Sandepp . can you also advice my thred - https://community.hortonworks.com/questions/222627/how-re-balance-the-partitions-to-available-brokers.html
... View more
09-13-2018
09:49 AM
Hi @Michael Bronson , ambari will give you recommendations on the Value to be set. You can set any value as per your Cluster and business logic. Once you saved and restart the Yarn clients the changes will be synced to your yarn-site.xml if you are able to save the configuration after clickin on proceed anyway ambari will save 100 . Please see if this helps you.
... View more
09-13-2018
12:02 PM
1 Kudo
@Michael Bronson I will look to create an article about configuring the vcores for cpu scheduling when I get time. I will mention this part there.
... View more
09-07-2018
12:49 PM
@Michael Bronson In yarn master mode executors will run inside a yarn container. Spark will launch an Application Master that will be responsible of negotiating the containers with Yarn. Having that said only nodes running Nodemanager are eligible to run executors. First question: The executor logs you are looking for will be part of the yarn application logs for the container running on the specific node. (yarn logs -applicationId <appId>) Second question: Executor will notify in case heartbeat fails to reach driver for some network problem/timeout. So this should be in the executor log that is part of the application logs. HTH *** If you found this answer addressed your question, please take a moment to login and click the "accept" link on the answer.
... View more
09-06-2018
01:09 PM
@Michael Bronson By default Spark2 has log level as WARN. Set it to INFO to get more context on what is going on in the driver and executor. More over the log will be locally available in Nodemanager when the container is still running. The easiest way is to go to spark UI (yarn application master UI) -> click on executors tab -> Here you should see stderr and stdout corresponding to driver and executors. Regarding the WARN on heartbeat , we'd need to check what driver is doing at that point. I think you already have asked another question with more details on driver and executor.
... View more
09-12-2018
04:16 PM
so in case we verify the logs of gc by http://gceasy.io/ , and we see that Driver isn't doing full garbage collection , that what are the next steps that we need to do ?
... View more
09-06-2018
10:01 AM
we need the debug for Spark Thrift server , we have issue when heartbeat from datanode machine not communicated with the driver , so this is the reason that we need debug mode on for Spark Thrift server
... View more
10-25-2018
01:29 PM
@Michael Bronson
The warning message means that the Executor is unable to send the Heartbeat to the driver (might be network issue). This is just a warning message, but each failure increments heartbeat failure count and when we hit the maximum failures the executor will fail and exit with error. There are two configurations that we can tune to avoid this issue. spark.executor.heartbeat.maxFailures (default value: 60) Number of times an executor will try to send heartbeats to the driver before it gives up and exits (with exit code 56). spark.executor.heartbeatInterval ( default value: 10s ) Interval between each executor's heartbeats to the driver. Heartbeats let the driver know that the executor is still alive and update it with metrics for in-progress tasks. spark.executor.heartbeatInterval should be significantly less than spark.network.timeout
... View more
09-05-2018
08:06 AM
@Jay . please let me know if I understand it as the following let say that one of the replica spark2-hdp-yarn-archive.tar.gz , is corrupted when I run this CLI su - hdfs -c "hdfs fsck /hdp/apps/2.6.4.0-91/spark2/spark2-hdp-yarn-archive.tar.gz" dose its actually means that fsck will replace the bad one with the good one and status finally will be HEALTHY ?
... View more
09-06-2018
07:28 AM
You can do tail in namenode and datanode log, also you can redirect output to dummy log file during restart. #tailf <namenode log> >/tmp/namenode-`hostname`.log #tailf <datanode log> >/tmp/datanode-`hostname`.log
... View more