Member since
08-13-2013
18
Posts
0
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2199 | 08-26-2013 02:33 AM |
01-21-2014
04:31 AM
We are facing issues with Open File descriptors in Task tracker from cloudera manager and most of the Task tracker are in critical health. Cloudera Manager: ver 4.7 Please find below the status of one such Task Tracker: "Open file descriptors: 24,241. File descriptor limit: 32,768. Percentage in use: 73.98%. Critical threshold: 70.00%." we have checked the File descriptor with the OS using the command "ulimit -a" and it shows "unlimited", core file size (blocks, -c) 0 data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 118683 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 1024 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 1024 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited In Cloudera documentation, we see the description for TaskTracker File Descriptors as: "The health check thresholds of the number of file descriptors used. A failure of this health check may indicate a bug in either Hadoop or Cloudera Manager. Contact Cloudera support." We found a similar issue in the issue list, but not sure whether they are talking about the same underlying problem: https://issues.cloudera.org/browse/DISTRO-530 Is this indeed a bug in the underlying framework or is there a way to increase the Open File Descriptors thorugh Cloudera Manager? Any other approach to recover from this issue is also appreciated. Thanks
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Cloudera Manager
08-26-2013
02:38 AM
When I tried to alter the table and added the external serde,its shows the error as : Added /tmp/b07e0a29-979b-49fd-8d09-a007a4a76529_resources/json-serde-1.1.6-jar-with-dependencies.jar to class path Added resource: /tmp/b07e0a29-979b-49fd-8d09-a007a4a76529_resources/json-serde-1.1.6-jar-with-dependencies.jar FAILED: RuntimeException MetaException(message:org.apache.hadoop.hive.serde2.SerDeException SerDe org.apache.hadoop.hive.contrib.serde2.JsonSerde does not exist)
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Hive
08-26-2013
02:33 AM
I have find out the solution.we need to add the core-site.xml property> <name>fs.s3n.awsAccessKeyId</name> <value></value> </property> <property> <name>fs.s3n.awsSecretAccessKey</name> <value></value> </property>
... View more
08-19-2013
09:21 PM
Thanks. Is this considered as a bug? If yes, what are some workarounds that we can follow for now? Any help is appreciated.
... View more
08-19-2013
01:46 AM
FWIW, the same job works fine as a workflow when submitted via Hue. In this case, we manually pass the input (S3) and output (hdfs) locations and the job runs successfuly - thus establishing that the problem is not with S3 support. The problem is when we let the co-ordinator pass this input (via a computed datasource) does it automatically prepend hdfs://{nameNode} in front of the s3n://<> URI. Hope this clarifies.
... View more
08-18-2013
11:14 PM
In coordinator jobs i'm passing the dataset uri template as s3n://xxx-xxx/${YEAR}${MONTH}${DAY}/${HOUR} and coord:dataOut as <property> <name>in_folder</name> <value>${coord:dataOut('in_folder')}</value> </property> and my workflow.xml input as ${in_folder} when I submit the coordinator job it automatically preappend the configuration like: ${nameNode}s3n://xxx-xxx/${YEAR}${MONTH}${DAY}/${HOUR}
... View more
08-14-2013
05:43 AM
My input path in another instance of EC2 as I specified input path as s3n://xxx-ssss/ It is showing error as: AWS Access Key ID and Secret Access Key must be specified as the username or password (respectively) of a s3n URL, or by setting the fs.s3n.awsAccessKeyId or fs.s3n.awsSecretAccessKey properties (respectively). Where could I configure the properties for fs.s3n.awsAccessKeyId and fs.s3n.awsSecretAccessKey in cloudera manager in web UI (HUE).
... View more
Labels:
- Labels:
-
Cloudera Hue
-
Cloudera Manager
08-13-2013
07:58 AM
Hello guys,
I have installed Cloudera Enterprise 4.6.2. And
I little bit stuck with oozie Input path.The Default URI for Dataset is
${nameNode} always ,I need to reconfigure the Input path to s3://xxx
.pls help me
... View more
Labels:
- Labels:
-
Apache Oozie
08-13-2013
07:33 AM
Sorry, my question is through Hue in cloudera manager i'm running the oozie job .And I can able to access the hdfs,my question is to connect the another instance Amazon as s3n://xxx to connect ..
... View more