Member since
11-02-2016
31
Posts
4
Kudos Received
0
Solutions
12-15-2017
06:51 PM
Thank you bkosaraju
... View more
12-14-2017
09:56 PM
Hello all, I want to ingest data logs from application server logs into HDFS using Flume version 1.5 . Do I need to install Flume agent (client) on these application servers? How can I pull these application logs without install Flume agent? However, these servers are not part of Hadoop cluster. Can you please help? Thanks JN
... View more
Labels:
- Labels:
-
Apache Flume
08-09-2017
01:38 PM
hi Sahi, Thanks for your posts. I am able to use to_utc_timestamp() and from_utc_timestamp() functions. Thanks
... View more
08-07-2017
09:02 PM
Hello experts; I have this query is getting PST time in Hive but how can I convert this PST time to UTC time? Please help select from_unixtime(unix_timestamp()-1*60*60*4, 'yyyyMMddHH') as pst_time; Thanks in advance JN
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Hive
08-07-2017
08:58 PM
Hello experts; I have this query is getting PST time in Hive but how can I convert this PST time to UTC time? Please help select from_unixtime(unix_timestamp()-1*60*60*4, 'yyyyMMddHH') as pst_time; Thanks in advance JN
... View more
Labels:
07-20-2017
04:11 PM
Hello all, I am getting core-site.xml (Permission denied) below when using Oozie-Spark workflow. I have access to this directory /opt/data/yarn/local/usercache/oozie/appcache drwxrwxrwx 2 yarn hadoop 10 Jul 20 11:30 appcache Could anyone please help me why I am getting the Permission denied? I really appreciated. Thanks Failing Oozie Launcher, Main class [org.apache.oozie.action.hadoop.SparkMain], main() threw exception, /opt/data/yarn/local/usercache/oozie/appcache/application_1500495991334_1431/container_e04_1500495991334_1431_01_000002/core-site.xml (Permission denied) java.io.FileNotFoundException: /opt/data/yarn/local/usercache/oozie/appcache/application_1500495991334_1431/container_e04_1500495991334_1431_01_000002/core-site.xml (Permission denied)
... View more
Labels:
- Labels:
-
Apache Oozie
07-20-2017
04:03 PM
Hello all, I am getting Permission denied error below when launching Oozie-Spark workflow. Also the directory with read, write access /opt/data/yarn/local/usercache/oozie/appcache drwxrwxrwx 2 yarn hadoop 10 Jul 20 11:30 appcache Could anyone please help me with below error? I really appreciated. Thanks Failing Oozie Launcher, Main class [org.apache.oozie.action.hadoop.SparkMain], main() threw exception, /opt/data/yarn/local/usercache/oozie/appcache/application_1500495991334_1431/container_e04_1500495991334_1431_01_000002/core-site.xml (Permission denied) java.io.FileNotFoundException: /opt/data/yarn/local/usercache/oozie/appcache/application_1500495991334_1431/container_e04_1500495991334_1431_01_000002/core-site.xml (Permission denied)
... View more
Labels:
- Labels:
-
Apache Oozie
07-12-2017
05:43 PM
Hello all, How to set variable in hive and insert value as partition? I have below commands and the partition is not getting from the variable that I set but getting constant "var" instead datetime. set hiveVar:var= from_unixtime(unix_timestamp()-1*60*60*4, 'yyyyMMddHH');
INSERT INTO TABLE tmp Partition (datehour='${hiveVar:var}') SELECT * FROM tmp2; Please help. Thanks JN
... View more
Labels:
- Labels:
-
Apache Hive
07-06-2017
01:54 PM
Thanks Kuldeep. Its working now
... View more
07-05-2017
09:29 PM
Hello all, I am getting below error when I tried to submit Workflow from Workflow Manager View in Ambari. at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.hadoop.security.authorize.AuthorizationException: User: root is not allowed to impersonate admin I also added to core-site properties: hadoop.proxyuser.root.groups="users"
hadoop.proxyuser.root.hosts=ambari-server.hostname I have Ambari runs as root but I login to Ambari as admin. Could you please help me how to avoid the AuthorizationException. Many Thanks JT Ng
... View more
Labels:
- Labels:
-
Apache Ambari
06-12-2017
05:47 PM
hi Shashank, I added the args from your recommended. My job is killed but I dont see any ERROR in the log. Any idea ? Thanks JT
... View more
06-09-2017
03:32 PM
hi all, I am getting this error in Oozie by using sqoop and Hive import. [main] ERROR org.apache.sqoop.tool.ImportTool - Encountered IOException running import job: java.io.IOException: Cannot run program "hive": error=2, No such file or directory
Caused by: java.io.IOException: error=2, No such file or directory
Below is my Oozie workflow:
<arg>import</arg>
<arg>--connect</arg>
<arg>jdbc:oracle:thin:@xxx</arg>
<arg>--username</arg>
<arg>abc</arg>
<arg>--password</arg>
<arg>abc</arg>
<arg>--query</arg>
<arg>SELECT STARTTIME FROM TEST WHERE STARTTIME BETWEEN TO_DATE('2017-06-08 14:00:00', 'YYYY-MM-DD HH24:MI:SS') AND TO_DATE('2017-06-08 14:10:59','YYYY-MM-DD HH24:MI:SS') AND $CONDITIONS)</arg>
<arg>--target-dir</arg>
<arg>/${examplesRoot}/output-data/sqoop/mytest</arg>
<arg>--hive-import</arg>
<arg>--hive-overwrite</arg>
<arg>--hive-table</arg>
<arg>stest_tmp</arg>
<arg>-m</arg>
<arg>1</arg>
<file>hdfs://localhost/sqoop/hive-site.xml</file>
Could you please help me with the error? What files that I am missing?
Thanks
... View more
Labels:
- Labels:
-
Apache Oozie
06-09-2017
02:04 AM
Hello all, I have this below action Oozie workflow but I keep getting these error: ERROR org.apache.sqoop.tool.BaseSqoopTool - Error parsing arguments for import: ERROR org.apache.sqoop.tool.BaseSqoopTool - Unrecognized argument: --hive-drop-import-delimns
ERROR org.apache.sqoop.tool.BaseSqoopTool - Unrecognized argument: --hive-overwrite
ERROR org.apache.sqoop.tool.BaseSqoopTool - Unrecognized argument: --hive-table ERROR org.apache.sqoop.tool.BaseSqoopTool - Unrecognized argument: test_tmp
ERROR org.apache.sqoop.tool.BaseSqoopTool - Unrecognized argument: --target-dir ERROR org.apache.sqoop.tool.BaseSqoopTool - Unrecognized argument: /apps/hive/warehouse/test
ERROR org.apache.sqoop.tool.BaseSqoopTool - Unrecognized argument: -m ERROR org.apache.sqoop.tool.BaseSqoopTool - Unrecognized argument: 1 ----------------------------- Oozie Workflow ---------------------------- <action name="sqoop-node"> <sqoop xmlns="uri:oozie:sqoop-action:0.2"> <job-tracker>${jobTracker}</job-tracker> <name-node>${nameNode}</name-node> <prepare> <delete path="hdfs://localhost/test-oozie/examples/output-data/sqoop"/>
<mkdir path="hdfs://localhost/test-oozie/examples/output-data"/>
</prepare> <configuration> <property>
<name>mapred.job.queue.name</name> <value>${queueName}</value>
</property>
</configuration>
<arg>import</arg> <arg>--connect</arg> <arg>jdbc:oracle:thin:</arg> <arg>--username</arg> <arg>user</arg> <arg>--password</arg>
<arg>pwd</arg>
<arg>--query</arg> <arg>SELECT * FROM MYTASK WHERE STARTTIME BETWEEN TO_DATE('2017-06-08 14:00:00', 'YYYY-MM-DD HH24:MI:SS') AND TO_DATE('2017-06-08 14:10:59','YYYY-MM-DD HH24:MI:SS') AND $CONDITIONS)</arg> <arg>--hive-import</arg> <arg>--hive-drop-import-delimns</arg>
<arg>--hive-overwrite</arg> <arg>--hive-table</arg>
<arg>test_tmp</arg> <arg>--target-dir</arg> <arg>/apps/hive/warehouse/test</arg> <arg>-m</arg> <arg>1</arg> Would you please help me why I got the Unrecognized argument ? Thank you for your help. JT
... View more
Labels:
- Labels:
-
Apache Oozie
06-07-2017
06:58 PM
2 Kudos
Hi all, I am getting this error in Hive info=[Error: Failure while running task:java.lang.IllegalArgumentException: tez.runtime.io.sort.mb 2027 should be larger than 0 and should be lessthan the available task memory (MB):538 Vertex did not succeed due to OWN_TASK_FAILURE, failedTasks:1 killedTasks:12, Vertex vertex_1496756949291_0733_1_00 [Map 1] killed/failed due to:OWN_TASK_FAILURE]Verte killed, vertexName=Reducer 2, vertexId=vertex_1496756949291_0733_1_01, diagnostics=[Vertex received Kill while in RUNNING state., Vertex did not succeed due to OTHER_VERTX_FAILURE, failedTasks:0 killedTasks:101, Vertex vertex_1496756949291_0733_1_01 [Reducer 2] killed/failed due to:OTHER_VERTEX_FAILURE]DAG did not succeed due to VERTEX_FAIURE. failedVertices:1 killedVertices:1 I am getting the error by using this "select ordernum, count(*) from status group by ordernum having count(*) > 1;" I've set tez.runtime.io.sort.mb = 2027 to tez.runtime.io.sort.mb = 409 but I still getting the same error tez.runtime.io.sort.mb 2027 Please help. Thanks
... View more
Labels:
- Labels:
-
Apache Hive
-
Apache Tez
06-07-2017
02:12 PM
Hello all, I have around 20 workflows with a lot of dependencies between them. Can I have one job.properties for 20 Workflows? How about coordinator and bundle? Should I use the same job.properties? Or Do I need to have job.properties for each Workflow, each Coordinator and Bundle? for example : /bundle.xml /job.properties (for bundle properties) /workflowA /coordinator.xml /workflow.xml /job.properties /workflowB /coordinator.xml /workflow.xml /job.properties Please give me advise. Thanks
... View more
Labels:
- Labels:
-
Apache Oozie
06-06-2017
08:41 PM
HI all, Does anyone know is the Workflow Manager to create Oozie jobs exist in Ambari verions 2.5.0.3 ? Please let me know Thanks
... View more
Labels:
- Labels:
-
Apache Ambari
06-06-2017
01:40 PM
Hi Eyad, I have Ambari Version 2.4.0.1 and HDP Version 2.5.3.0-37. Is there anyway I can upgrade Ambari to 2.5+ without upgrading whole HDP? Please advise. Thanks
... View more
06-06-2017
12:43 AM
You answered my question. Many Thanks
... View more
06-05-2017
07:16 PM
1 Kudo
Hello all; I have separate sqoop workflow, hive workflow and spark workflow. Is there anyway I can combine these three workflows into one Oozie job? if so, What should I use? Oozie Coordinator or Bundle? Please advise Thanks
... View more
Labels:
- Labels:
-
Apache Oozie
06-04-2017
01:38 PM
Hello, I have these file in my workflow now <file>/test-oozie/examples/apps/sqoop/hive-site.xml</file>
<file>/test-oozie/examples/apps/sqoop/tez.xml</file> But the errror now is "JA008: File does not exist hdfs://localhost/test-oozie/examples/apps/sqoop/tez.xml#tez.xml" Even though I copied the Tez jars and tez.tar.gz into hdfs. Which jar am I missing? Thanks
... View more
06-04-2017
02:18 AM
Hello all, I am new to oozie. I have oozie wf working successful with sqoop command but when I added hive import into oozie workflow I got it failed as "Main class [org.apache.oozie.action.hadoop.SqoopMain], exit code [1]" Can anyone please help me? I really appreciated. Thanks Here is my workflow: /************ <?xml version="1.0" encoding="UTF-8"?> <workflow-app xmlns="uri:oozie:workflow:0.2" name="sqoop-wf"> <start to="sqoop-node"/>
<action name="sqoop-node"> <sqoop xmlns="uri:oozie:sqoop-action:0.2">
<job-tracker>${jobTracker}</job-tracker> <name-node>${nameNode}</name-node>
<prepare>
<delete path="test-oozie/examples/output-data/sqoop"/>
<mkdir path="test-oozie/examples/output-data"/>
</prepare>
<configuration>
<property>
<name>mapred.job.queue.name</name>
<value>${queueName}</value>
</property>
<property>
<name>oozie.hive.defaults</name>
<value>/examples/apps/sqoop/hive-site.xml</value>
</property>
</configuration>
<arg>import</arg>
<arg>--connect</arg>
<arg>jdbc:oracle:thin:@hostname</arg>
<arg>--username</arg>
<arg>usename</arg>
<arg>--password</arg>
<arg>pwd</arg>
<arg>--query</arg>
<arg><![CDATA[SELECT ID, STARTTIME FROM TEST WHERE STARTTIME BETWEEN TO_DATE('2017-05-30 13:00:00', 'YYYY-MM-DD HH24:MI:SS') AND TO_DATE('2017-05-30 13:15:59','YYYY-MM-DD HH24:MI:SS') AND (1=1 OR $CONDITIONS)]]></arg> <arg>--hive-import</arg>
<arg>--hive-drop-import-delims</arg>
<arg>--hive-overwrite</arg>
<arg>--hive-table</arg>
<arg>task_detailstest_tmp</arg>
<arg>--target-dir</arg>
<arg>/apps/hive/warehouse/task_detailstest</arg>
<arg>-m</arg>
<arg>1</arg>
</sqoop>
<ok to="end"/>
<error to="fail"/>
</action>
<kill name="fail">
<message>Sqoop failed, error message[${wf:errorMessage(wf:lastErrorNode())}]</message>
</kill>
<end name="end"/>
</workflow-app>
... View more
Labels:
- Labels:
-
Apache Oozie
02-20-2017
06:11 PM
hi Arvits; I am using HDP 2.4.2.0. I am trying to import data from database into hdfs using oozie workflow. I have questions to ask you, that I execute the oozie command as mapred user but I keep getting permission denied: user=mapred, access=READ. I have directory owned by mapred (drwxrwxrwx - mapred hadoop 0 2016-11-01 09:40 /mr-history) however, when oozie job started somewhere generate the job by hdfs and changed mapred user to read only. Can you please show me where I can set or change the hdfs permission to mapred? Pls see below log: Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): Permission denied: user=mapred, access=READ, inode="/mr-history/tmp/hdfs/job_1487105716370_1426.summary":hdfs:hdfs:-rwxrwx---
at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:319)
at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:219)
at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:190)
... View more
02-20-2017
03:28 PM
Hello all, I am successfully able to run the sqoop command. But when I added to oozie workflow xml I am not able to do with oozie. I received STATUS as KILLED. The error code is JA018. Here is my workflow: <workflow-app xmlns="uri:oozie:workflow:0.4" name="oozie-wf"> <start to="sqoop-wf"/>
<action name="sqoop-wf">
<sqoop xmlns="uri:oozie:sqoop-action:0.2">
<job-tracker>${jobTracker}</job-tracker>
<name-node>${nameNode}</name-node> <command>import --connect jdbc:oracle:thin:@dbservername:1521:database --username user --password passwprd --query "SELECT TASKID FROM TASKS WHERE TASKTIMESTAMP BETWEEN TO_DATE('2017-02-13 00:00:00', 'yyyy-mm-dd hh24:mi:ss') AND TO_DATE('2017-02-13 23:59:59', 'yyyy-mm-dd hh24:mi:ss') AND \$CONDITIONS" --fields-terminated-by '|' -m 1 --target-dir /dbdata/mytest</command>
</sqoop>
<ok to="end"/>
<error to="fail"/>
</action>
<kill name="fail"> <message>Failed, Error Message[${wf:errorMessage(wf:lastErrorNode())}]</message>
</kill>
<end name="end"/>
</workflow-app> Could someone please help me? I am new to oozie and I stuck with this error for awhile. Thanks in advance
... View more
Labels:
- Labels:
-
Apache Oozie