Member since
08-23-2016
261
Posts
201
Kudos Received
106
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1792 | 01-26-2018 07:28 PM | |
1416 | 11-29-2017 04:02 PM | |
35474 | 11-29-2017 03:56 PM | |
3577 | 11-28-2017 01:01 AM | |
988 | 11-22-2017 04:08 PM |
05-29-2017
03:05 PM
1 Kudo
Hi @heta desai You could be using HDF (NiFi) as your primary ingestion tool and not have to worry about the other options necessarily. That said, sqoop is primarily used to move data from an existing RDBMS to Hadoop (or vice versa). Flume was the main tool previously to ingest flat files, csv, etc, but has fallen out of favour and is often being replaced with HDF/NiFi now. Kafka is a distributed messaging system which can be used as a pub/sub model for data ingest, including streaming. So all three are a bit different. The right tool for the job depends on your use case, but as I said, HDF/NiFi can pretty much cover the gambit, so if you are starting out now, you may want to look at that first. Here is another good write up on the same subject: https://community.hortonworks.com/questions/23337/best-tools-to-ingest-data-to-hadoop.html As always, if you find this post useful, please accept the answer.
... View more
05-29-2017
02:54 PM
3 Kudos
Hi @Narasimma varman the InvokeHTTP Processor in NiFi can be used. More on the InvokeHTTP processor here: https://nifi.apache.org/docs/nifi-docs/components/org.apache.nifi/nifi-standard-nar/1.2.0/org.apache.nifi.processors.standard.InvokeHTTP/ In fact, there is a good tutorial walking through how to get data from an API endpoint here: https://hortonworks.com/hadoop-tutorial/learning-ropes-apache-nifi/#section_5 As always, if you find this source useful, please accept the answer.
... View more
05-29-2017
02:29 PM
1 Kudo
Hi @Anirban Das Deb I don't think I have access to the same course, but, all of the Sandbox verisons I have used have a different port for SSH access. Try using port 2222 instead and see if that works: ssh root@127.0.0.1 -p 2222 As always, if you find this post useful, please accept the answer.
... View more
05-25-2017
01:21 AM
1 Kudo
Hi @Cam Mach There isn't one for Kudu out of the box. However, you could either create a custom one, or create a flow that would land data on the box and then call the Kudu APIs (ExecuteScript processor maybe?) for the ingest. As always if you find this post useful, please accept the answer.
... View more
05-24-2017
10:20 PM
1 Kudo
Hi @Carlos Barichello I haven't seen anything like that yet. The closest would the API or GUI feature to clear the results from the paragraph, but, that doesn't erase the cache from previous results. The only way I've seen is to edit the JSON file. I suppose though that you could also just reuse the variable in another paragraph to make it useful again?
... View more
05-23-2017
08:59 PM
1 Kudo
Hi @Carlos Barichello Zeppelin stores these values in a file called note.json. This file is usually stored inside of a directory (the note's ID is usually in the URL for the note, and is the subdirectory name) on the Zeppelin home server, and can be edited in a text editor where you can remove the values/paragraphs if you so require. Here is a good post showing how to do it: https://community.hortonworks.com/questions/24254/zeppelin-how-to-remove-results-cache-from-notebook.html As always, if you find this post useful, please accept the answer.
... View more
05-23-2017
07:34 PM
No, I don't have this issue. I'm able to hit the nifi UI at http://127.0.0.1:19090/nifi/ without a problem
... View more
05-23-2017
04:06 PM
1 Kudo
Hi @brian adi You can use the SplitJson processor to get the array into individual flowfile records, then use EvaluateJsonPath to pull out the values of interest, and ReplaceText to get the content back into a CSV format. Here is a good example with screenshots showing how to do this using NiFi https://community.hortonworks.com/articles/64069/converting-a-large-json-file-into-csv.html As always, if you find this post useful, don't forget to accept the answer.
... View more
05-23-2017
02:16 PM
1 Kudo
Can you get to the main welcome page? http://127.0.0.1:18888 or try also http://127.0.0.1:9080
... View more