Member since
08-16-2016
642
Posts
131
Kudos Received
68
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 3978 | 10-13-2017 09:42 PM | |
| 7477 | 09-14-2017 11:15 AM | |
| 3804 | 09-13-2017 10:35 PM | |
| 6047 | 09-13-2017 10:25 PM | |
| 6607 | 09-13-2017 10:05 PM |
02-21-2017
10:16 PM
Ok, had to get my MySQL DBA hat back out. InnoDB max key length is 767 bytes and MyISAM's is 1000. Latin1 is a character to byte equivalent while utf8 requires additional bytes per character. So the key that is trying to add to this table at a minimum is 1151 bytes. I just checked out my CDH 5.8.2 metastore and see the same index and same column sizes. So I don't have idea why it is an issue for you. Can you try upgrading to a lower CDH version first?
... View more
02-21-2017
02:56 PM
Do you have the bigtop-detect-javahome at either location? The error in the one image says it isn't there which may indicate an issue with your parcels or packages install (oh below is for parcels). /opt/cloudera/parcels/CDH/lib/bigtop-utils/bigtop-detect-javahome /opt/cloudera/parcels/CDH/bin/bigtop-detect-javahome
... View more
02-21-2017
02:02 PM
Sorry wrong setting. yarn.nodemanager.log-aggregation.roll-monitoring-interval-seconds
... View more
02-21-2017
01:50 PM
yes, that is pretty frequent though so I don't know how it will go. I'd be interested to know.
... View more
02-21-2017
01:21 PM
This got lost in my earlier reply... yarn.log-aggregation.retain-check-interval-seconds This determine when it checks if logs need to be aggregated. By default it is 0 which means it doesn't check and a job must finish. This will allow it to collect the logs for jobs that, in theory, won't end.
... View more
02-21-2017
12:23 PM
What is your CDH version?
... View more
02-21-2017
12:18 PM
I think I tried that too but it doesn't work. You need to set it in the spark-opts (where you should have your exectutor and driver memory set) like '--files hdfs:///user/hue/oozie/workspaces/hue-oozie-1463575878.15/hive-site.xml'
... View more
02-21-2017
11:54 AM
Where is your hive-site.xml located? This exception indicates that it isn't available to the job so it is launching the default embedded derby HMS database.
... View more
02-21-2017
11:46 AM
2 Kudos
In my opinion, the issue at had is that the hive-site.xml is not passed properly and there for it defaults to using the embedded derby database. This "fix" is just allowing the Spark job to use an embedded derby HMS instead of your actual HMS. Have you checked that it is properly created tables or other metadata in your actual HMS?
... View more
02-21-2017
11:32 AM
You could do this in many ways. You could just load it in Solr/ES and go to town. Hive would not be a great fit but I could see some tables being build around specific data like job counters or metrics. MR jobs could be build to pull out specific data (possible to load into a Hive table) or Spark jobs (and the Spark shell can be used to explore there raw data). And simple tools like grep, awk, etc. can be used as the individual logs, when aggregated, are available to the user. If you have CM, the YARN application screen for a cluster, I'm pretty sure, is built using an embedded Solr and gives you and idea of what could be done. This is more around metrics and job counters again.
... View more