Member since
08-16-2016
642
Posts
131
Kudos Received
68
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3926 | 10-13-2017 09:42 PM | |
7343 | 09-14-2017 11:15 AM | |
3734 | 09-13-2017 10:35 PM | |
5924 | 09-13-2017 10:25 PM | |
6498 | 09-13-2017 10:05 PM |
04-14-2017
10:38 AM
It's true that you can aggreate logs to hdfs when the job is still running, however, the minimun log uploading interval (yarn.nodemanager.log-aggregation.roll-monitoring-interval-seconds) you can set is 3600 seconds which is 1 hour. The design is trying to protect namenode from being spamed. You may have to use an external service to do the log aggregation. Either write your own or find other tools. Below is the proof from yarn-default.xml in hadoop-common source code (cdh5-2.6.0_5.7.1). <property> <description>Defines how often NMs wake up to upload log files. The default value is -1. By default, the logs will be uploaded when the application is finished. By setting this configure, logs can be uploaded periodically when the application is running. The minimum rolling-interval-seconds can be set is 3600. </description> <name>yarn.nodemanager.log-aggregation.roll-monitoring-interval-seconds</name> <value>-1</value> </property>
... View more
04-12-2017
02:28 PM
There is probably an issue with the client connecting to the Datanode. It is reporting that you have one live data nodes but it is failing to place any replica on it. I would expect the client to get a different error if it was failing to write out the first replica. Check the NN UI to validate that the DN is live, and check the NN and DN logs to see if there is more information on what the issue is.
... View more
04-11-2017
06:22 AM
I moved it for you. 🙂
... View more
04-03-2017
10:51 PM
Thanks 🙂
... View more
03-22-2017
07:57 AM
We are only runnng hdfs, so we only need to upgrade that. Since it was a dev environment, we shut all of hdfs down, download hadoop-2.6.0-cdh5.8.4.tar.gz from http://archive.cloudera.com/cdh5/cdh/5/ and run with that. (We are actually running with hdfs on mesos, so the artifacts get packaged up into an uberjar with the mesos executor, but there's no real magic there. I think it just uses the stuff in hadoop/common and hadoop/hdfs and some of the run scripts.)
... View more
03-05-2017
02:26 PM
When i checked the job/the query that occur prior to the alert on the JN, i found one hive query that runs on a data of 6 months and recreate the hive table from new, which resulted in a good percentage of edit logs, i contacted the query owner and he reduced the his running window from 6 months to 2 months which solve for us the issue.
... View more
03-02-2017
04:39 PM
@Akira191 1. Go to Cloudera Manager -> Spark -> Instance -> Identify the node where you have Spark server installed 2. Login to the above identified node using CLI and go to path "/opt/cloudera/parcels/CDH-<version>/lib/spark/bin" , it will list binaries for "spark-shell, pyspark, spark-submit, etc". It helps us to login to spark & submit jobs. if it has spark-sql, then you can run the command that you have mentioned. In your case, spark-sql binary should be missing, so you are getting this error. You need to talk to your admin
... View more
03-02-2017
07:49 AM
The only suggestion I have is to try running some tests to see if you can weed out any bad disks. DFSIO and Terasort may hit on it but may not. You can use 'dd' or other software to test the raw disks. Beyond that you may be chasing ghosts (spending more time than worth it on an ephemeral problem).
... View more
03-02-2017
03:01 AM
Thanks, Removing all alternatives OpenJDK helped:)
... View more