Member since
08-12-2014
34
Posts
2
Kudos Received
2
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2331 | 09-22-2014 03:47 PM | |
12593 | 09-15-2014 04:04 PM |
01-10-2016
08:49 AM
Hi Harsh, Thank you very much for your assistance. We are using CM 5.3.3, and plan to upgrade later. I will explore the API docs, about using the CLI. Michael
... View more
01-08-2016
03:02 PM
We have numerous backup replications defined and running, causing access via a web browser to be very slow. Is there a way to access, create, and modify schedules via the Linux shell? Michael Reynolds
... View more
Labels:
- Labels:
-
HDFS
12-21-2015
12:35 PM
Thank you very much for your reply. According to the agent log, the service account responsible for running the Impala service does not exist on the host. I appreciate your pointing me in the right direction.
... View more
12-21-2015
09:02 AM
1 Kudo
One of our Impala daemons will not start. There is no directory structure being created at: /var/run/cloudera-scm-agent/process/nnnnnn-impala-IMPALAD After the daemon fails to start, the stderr and stdout logs cannot be accessed because the directory structure does not exist. Instead, there is an HTTP 403 error, referencing the expected path, indicating "The server declined access to the page or resource." In addition, nothing is being written to the error, warning, or info logs at /hadoop/log/impalad. What can be done to enable the creation of the directory structure? Deleting and reinstalling the Impala role does not resolve it. There are no issues with other roles on the host. I am wondering whether it is necessary to re-initialize the host and re-install the roles. Michael Reynolds
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Impala
09-22-2014
03:47 PM
I found the running coordinator in the Oozie Web Console. By referencing the job ID in the console, I was able to kill the coodinator via the Oozie -kill command. Michael Reynolds
... View more
09-19-2014
09:54 AM
I want to change the schedule for a running Oozie Coordinator, however it is missing from the list of running coordinators in Hue. Further, it is not listed if I run the following shell command: sudo -u oozie oozie jobs -oozie http://<server-url>:11000/oozie/ -jobtype coordinator | grep RUNNING How is it possible to locate the coodinator, so I can kill it? Thank you, Michael Reynolds
... View more
Labels:
- Labels:
-
Apache Oozie
09-15-2014
04:04 PM
1 Kudo
I was able to get this working. The solution is to submit all of the query elements as separate arguments. Nothing should be in the "Command" window. Instead, starting with "import" as the first argument, enter each part of the query as a separate argument. Properties and values for each element are entered as separate arguments. For example: arg: import
arg: --connect
arg: jdbc:mysql....
arg: --username
arg: [username]
arg: --password-file
arg: [password file]
arg: --query
arg: select .....
arg: --target-dir
arg: [target] The workflow performs as expected. Michael Reynolds
... View more
08-18-2014
03:29 PM
We are still experiencing periodic problems with applications hanging when a number of jobs are submitted in parallel. We have reduced 'maxRunningApps', increased the virtual core count, and also increased 'oozie.service.callablequeueservice.threads' to 40. In many cases, the applications do not hang, however this is not consistent. Regarding YARN issue number 1913 (https://issues.apache.org/jira/browse/YARN-1913), is this patch incorporated in CDH 5.1.0, the version we are using? YARN-1913 indicates the affected version is 2.3.0, and is fixed in 2.5.0. Our Hadoop version in 5.1.0 is 2.3.0. Thank you, Michael Reynolds
... View more
08-12-2014
02:34 PM
Thank you very much for your assistance! It is now working fine. Michael Reynolds
... View more
08-12-2014
10:19 AM
I am experiencing the same problem stated earlier. We have a 4-node cluster using YARN on v5.1.0. I have an Oozie workflow that uses Sqoop to import from MySQL, which is sharded with 10 tables. Therefore, I have a coordinator that executes the same workflow with 10 simultaneous (parallel) sessions, to pull from each sharded table. However, sometime after the workflows reach the Sqoop action step, they stop running. The jobs are not failing, rather they stop processing, even though their status shows "Running" in the Hue workflow dashboard. None of the jobs have had any updated status in the SysLog for more than 12 hours. Further, if other, unrelated jobs are submitted, they also appear to hang. I have had a job running successfully for several days, which is executing a DISTCP command to import S3 data. This job has also hung after submitting the 10 parallel workflows. Is there a configuration that must be set to allows the same workflow to be processed in parallel? Thank you! Michael Reynolds
... View more