Member since
07-08-2016
260
Posts
43
Kudos Received
9
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
619 | 05-02-2018 06:03 PM | |
1020 | 10-18-2017 04:02 PM | |
375 | 08-25-2017 08:59 PM | |
524 | 07-21-2017 08:13 PM | |
4462 | 04-06-2017 09:54 PM |
04-15-2021
03:58 PM
Thanks @wsalazar for the insights. I know it is an older article, but it is worth revisiting. For real time data need what would approach would you take to connect from NiFi?
... View more
01-26-2021
08:51 AM
https://nifi.apache.org/docs/nifi-docs/components/nifi-docs/components/org.apache.nifi/nifi-gcp-nar/1.9.0/org.apache.nifi.processors.gcp.bigquery.PutBigQueryBatch/index.html With this processor, you can do Batch loads flow files content to a Google BigQuery table.
... View more
04-27-2020
12:57 PM
The answer @mpayne is correct. Only that setting the header in MergeContent doesn't include a line break at the end of header and the records. So as EL is supported please include line break at the end.
... View more
04-19-2020
05:13 PM
Hi @DarkStar
As this thread was marked 'Solved' in March of 2018 you would have a better chance of receiving a resolution by starting a new thread. This will also provide the opportunity to provide details specific to your XML source that could aid others in providing a more targeted answer to your question.
... View more
02-18-2020
08:03 AM
Hi All,
I am facing issue still after adding below both hive-site.xml and core-site.xml.
<property> <name>hadoop.security.authentication</name> <value>kerberos</value> </property>
I am facing below error
org.apache.commons.dbcp.SQLNestedException: Cannot create JDBC driver of class 'org.apache.hive.jdbc.HiveDriver' for connect URL jdbc:hive2://ux329tas101.ux.hostname.net:10000/default;principal=<principal name>;ssl=true
Could you please help me regarding this.
Regards,
Swadesh Mondal
... View more
02-01-2020
08:36 AM
Hello, did you make this work? Is it possible to run the Nifi as a service on Windows in some way? Please give us more detailed information about it, if you succeeded... Can't apps like AlwaysUp or JSL work? Java Service Wrappers... Thanks.
... View more
01-23-2020
06:23 AM
@alexmc As this thread is a couple of years old, you would be better served by creating a new thread.
... View more
01-07-2020
11:56 PM
Hello @saikrishna_tara @bbende @emaxwell . Thanks for the solution and it worked well for me. I am new to Nifi and got the same problem statement as @saikrishna_tara. I am able to make it till merge content, i can see my files are in parent flow files. but in parent name of files are uuid of the flow files and not the actual name of the file which is processed. I need to put all the parent flow file's actual name via email processor to the outside world. Please let me know in case more details are required. Regards Nitin
... View more
11-20-2019
04:11 PM
How to do the file count if we are using Compress Content Processor ?
... View more
11-13-2019
11:01 AM
Hi @mburgess I tried adding attribute name “auth” in Attributes to Send property and had this “auth” as a custom property in my invokeHTTP processor with value 'Authorization: Bearer ${access_token}' where the value of variable access_token gets populated from EvaluateJsonPath processor in front of my invokeHTTP, but somehow I am still getting unauthorized as the response. Is there anything else that I’m missing for token to get passed on as a header in this call?
... View more
09-29-2019
11:51 PM
Hi, is it possible to disabled authentication on schema registry? I am trying to do that but service fails to start with error message: Configuration parameter 'registry_ui_principal_name' was not found in configurations dictionary
... View more
01-15-2019
10:55 AM
In short: Please ensure the maximum log size is configured properly. Here is an existing answer that should help: https://community.hortonworks.com/answers/144063/view.html
... View more
07-30-2018
06:26 PM
ValidateRecord is more about validating the individual records than it is about validating the entire flow file. If some records are valid and some are invalid, each type will be routed to the corresponding relationship. However, for invalid records, we can't use the same record writer as valid records, or else we know it will fail (because we know they're invalid), so there is a second RecordWriter for invalid records (you might use this to try to record the field names or something, but by the time that ValidateRecord knows the individual record is invalid, it doesn't know that it came in as Avro (for example), nor does it know that you might want it to go out as Avro. That's the flexibility and power of the Record Reader/Writer paradigm, but in this case the tradeoff is that you can't currently treat the entire flow file as valid or invalid. It may make sense to have a "Invalid Record Strategy" property, to choose between "Individual Records" using the RecordWriters (the current behavior), or "Original FlowFile" which would ignore the RecordWriters and instead transfer the entire incoming flow file as-is to the 'invalid' relationship. Please feel free to file an improvement Jira for this capability.
... View more
08-01-2018
04:46 PM
Hi , just trying to get some attention from experts.. any idea on how to decrypt PGP file using Passphrase and secrect key (Private Key). Regards, Sai
... View more
07-17-2018
06:29 PM
@Saikrishna Tarapareddy - The only NiFi configuration file you can edit that will take affect without requiring a NiFi restart is the logback.xml file. - As far as what is an acceptable search base, best to test your search base command on command line using ldapsearch. If it doesn't work there, it will not work in NiFi either. - Thank you, Matt - If you found this Answer addressed your original question, please take a moment to login and click "Accept" below the answer.
... View more
07-03-2018
09:21 PM
Jeez I would hope not, I'm not aware of any platform differences for Jayway (the underlying library used to do JSONPath stuff in NiFi)
... View more
07-02-2018
08:21 PM
1 Kudo
@Saikrishna Tarapareddy - You can increase the number of concurrent tasks on each remote port you have established a connection to: Once you have your flow established/connected to your Remote Process Group (RPG), right click on the RPG to bring up teh following context menu: Select "Manage Remote Ports" to open a new UI. You will then be able to edit each Remote Input/Output you have established connection to: You will also want to increase the number concurrent tasks on your remote ports as well. - Thanks, Matt - When an "Answer" addresses/solves your question, please select "Accept" beneath that answer. This encourages user participation in this forum.
... View more
06-19-2018
05:36 PM
I recommend using MergeRecord before the JoltTransformJSON as the Jolt transform should be able to be applied to the whole JSON array (after your smaller JSON objects have been merged). You'll want to use a JsonTreeRecordReader and provide an Avro schema that matches your input data above. mergerecord-example.xml is an example template where I generate data similar to yours, use MergeRecord to bundle them 20 at a time, then run the Jolt spec on it, it includes the associated Avro schema and hopefully all config to get you up and going.
... View more
01-15-2019
11:24 AM
First of all doublecheck all configurations (incl. password). Just to avoid moving in the right direction. Secondly confirm that you do not need TLS enabled. If these don't help, the following might help with troubleshooting: 1. Become nifi on the node where nifi is running 2. Send the message via Python 3. Share the python command here Note: Please explicity specify all things that you configure in nify when executing python (even if they are not needed because of good defaults for instance).
... View more
05-29-2018
09:09 PM
Hi guys, thanks so much for the fast support and thanks to the Matts Team @Matt Burgess and @Matt Clarke I finally understood how the processor works. He emits a flow file with no payload and in the meta attributes are the file details like path and filename. Those are used by the HDFSFetch to fetch the correspondent files. Kind regards, Paul
... View more
01-31-2019
02:22 PM
@Shu , can you upload xml for recent solution , @shu can you upload .xml file here for recent flow
... View more
05-17-2018
08:09 PM
@Chandan Singh If you would like to investigate this further, you can get a NiFi thread dump and look for the "putJMS" related threads to see what those threads are waiting on. # ./nifi.sh dump <dump-file-name> Unfortunately right now there is no way to interrupt/kill these threads from within NiFi's UI. NiFi will be adding that ability in a future release. For now, as Saikrishna mentioned, a NiFi restart will be required to kill these threads if they never complete on their own. Thanks, Matt
... View more
04-17-2018
02:01 PM
I am working on NIFI-4456 which will allow the JSON reader/writer to support the "one JSON per line" format as well as the "JSON array" format for input and output, so you will be able to read in one JSON per line and output a JSON array, using ConvertRecord (or any other record-aware processor). In the meantime, you can use the following crude script in an ExecuteGroovyScript processor to process your entire file (avoiding the Split/Merge pattern), it should get you what you want: def flowFile = session.get()
if(!flowFile) return
flowFile = session.write(flowFile, {inStream, outStream ->
outStream.write('['.bytes)
inStream.eachLine { line, i ->
if(i > 1) outStream.write(','.bytes)
outStream.write(line.bytes)
}
outStream.write(']'.bytes)
} as StreamCallback)
session.transfer(flowFile, REL_SUCCESS) The script just adds array brackets around the whole doc, and separates the lines by a comma. I did the crude version because it doesn't need to load the entire input content into memory. If you need more control over the JSON objects, you could iterate over the lines (still with eachLine), use JsonSlurper to deserialize each string into a JSON object, then add each object to an array, then use JsonOutput to serialize the whole thing back to a string. However that involves having the entire content in memory and could get unwieldy for large input flow files.
... View more
04-06-2018
02:32 PM
Hi, has anyone accessed GCS buckets thru NiFi.? if so what am I doing wrong.? Regards, Sai
... View more
03-27-2018
09:18 PM
are there any logs errors or anything else in the log? how much ram on the machine?
... View more
03-21-2018
06:38 PM
@Saikrishna Tarapareddy
Flatten json processor doesn't work if you are having arrays,nested arrays in the json content and the flowfile will route to failure if you are having array,nested arrays in the content. we still need to use splitjson (or) jolt transform processors to split the array. As this processor joins all The keys are combined at each level with a user-defined separator that we have specified in the processor configs. Input Json:- {
"id": 17,
"name": "John",
"child": {
"id": "1"
}} Output json:- {"id":17,"name":"John","child.id":"1"} As you can see the nested json message has been joined with .(period) in the output json content.
... View more
03-20-2018
07:38 PM
@Saikrishna Tarapareddy Which HDF components have you installed? Did you do an Ambari or a manual install?
... View more
08-27-2018
07:12 AM
@Scott Shaw , I am not able to access your gitub url . here I have a requirement like need to create the hive table on nested JSON file .please help me how to do it using jsonserde serde..thank you in advance
... View more
03-09-2018
08:24 PM
Hi, We are trying secure our NiFi instance by using our company signed certificates instead of self signed. The instance is working fine with self-signed certs. after adding the new company CA certs to keystore and truststores , we changed the nifi.cluster.node.address and nifi.web.https.host to match the host from certificate and restarted NiFi. It looks like NiFi has restarted successfully but i am not able to log in . it shows Success you are already logged in. The hostname of the server is different than the host name of the certificate.
... View more
Labels:
03-01-2018
08:56 PM
@Bryan Bende looks like I have an option to use SegmentContent and MergeContent after I use ReplaceText on each segment. I tried it with 10mb segment size on a 120 mb file and it worked..now will try on the bigger file.
... View more