Member since
07-19-2018
613
Posts
101
Kudos Received
117
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 5094 | 01-11-2021 05:54 AM | |
| 3421 | 01-11-2021 05:52 AM | |
| 8789 | 01-08-2021 05:23 AM | |
| 8385 | 01-04-2021 04:08 AM | |
| 36687 | 12-18-2020 05:42 AM |
04-02-2020
11:45 AM
@Gubbi I think your ListFile proc is still executing 0 sec. Reference our private message.
... View more
04-02-2020
06:32 AM
1 Kudo
@Gubbi The next solution here is to just add each route for today, yesterday, day before yesterday. Then route all 3 to the next proc. Anything not matching won't be routed.
... View more
04-02-2020
06:29 AM
1 Kudo
@Gubbi The solution here is a now minute 24 hours: Yesterday: ${now():minus(86400000):format('MM-dd-yyyy hh:mm:ss') } Day Before Yesterday: ${now():minus(86400000):minus(86400000):format('MM-dd-yyyy hh:mm:ss') }
... View more
04-01-2020
12:50 PM
@Saisreenath According to this article: you should be in the folder containing the script you want to execute to set the execution permissions. Nifi needs permissions to access the script. It also needs permission to be able to execute the command you gave it. In your error it is saying it cannot run the python command. You may need to adjust that command path. In summary make sure you give correct permissions to everything you put in Properties tab of the NiFi Processor, and use the correct paths for python.
... View more
03-30-2020
07:51 PM
Thanks Steven
... View more
03-29-2020
10:38 AM
Yes, by adding jar in classpath, i am able to import without using -libjars option also as given below.
export HADOOP_CLASSPATH=/app/hadoop_users/Mahfooz/sqoop/mysql-connector-java-5.1.48.jar
sqoop import \
--connect "jdbc:mysql://localhost:3306/move?verifyServerCertificate=false&zeroDateTimeBehavior=round" \
--username "something" \
--password "something" \
--delete-target-dir \
--table users \
--fields-terminated-by "," \
--hive-import \
--create-hive-table \
--hive-table test.users \
-- \
--schema "move"
But what is the use of this -libjars option. I am still confused.
... View more
03-29-2020
05:58 AM
@saaga119 The solution here is to create the external table in the format you need to query the raw xml data. Seems like you have this done already. Next create a native hive table. It could still be external too. The new table should be the schema for the fields you want with required ORC lines. You can also choose some other text format; for example: csv. Last INSERT INTO new_table SELECT * FROM external_table. The main idea here is to use external and staging tables, then INSERT INTO SELECT to fill the final table you want. Final table could then be optimized for performance (ORC). This kind of idea also allows multiple external/raw/staging tables able to be combined (select w/ join) into a single final optimized table (orc, compression, partitions, buckets, etc).
... View more
03-28-2020
09:26 AM
1 Kudo
great and Thanks it was very useful for my work. Here I will be having data in the excel, I need to extract the column names(1st row) and form the create the table with primary key and foreign key and data(2nd row) in the database(mysql) from nifi.
... View more
03-27-2020
06:01 AM
@Savi I think you only need one date format in your nifi expression language. Next, it seems you know what format it is upstream (source), and what it needs to be for downstream. Based on your expression need, you should play the flow up to the csv reader processor and then List Queue on the FlowFiles and inspect the attributes. Adjust the expression, and retest until the attribute format meets the required format. Inspecting the dates upstream won't show you what the date looks like after expression and before the reader. Often times I find myself testing a flow step by step, each step confirming the attributes needed are correct for the next step. Never moving forward to the next step until I am sure the attributes are correct after visual inspection.
... View more
03-26-2020
07:18 AM
Hi @Shelton, We have disabled ranger authorization in Ambari & allow to run hive as end user instead of hive user. Still hiveserver2 is not coming up. 2020-03-26 05:36:05,838 - call['/usr/hdp/current/zookeeper-client/bin/zkCli.sh -server <master node>:2181,<datanode1>:2181,<data node3>:2181 ls /hiveserver2 | grep 'serverUri=''] {} 2020-03-26 05:36:06,497 - call returned (1, 'Node does not exist: /hiveserver2') 2020-03-26 05:36:06,498 - Will retry 1 time(s), caught exception: ZooKeeper node /hiveserver2 is not ready yet. Sleeping for 10 sec(s) Any clue on this ?
... View more