Member since
07-30-2019
333
Posts
357
Kudos Received
76
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
10197 | 02-17-2017 10:58 PM | |
2430 | 02-16-2017 07:55 PM | |
8191 | 12-21-2016 06:24 PM | |
1827 | 12-20-2016 01:29 PM | |
1289 | 12-16-2016 01:21 PM |
10-08-2015
02:58 PM
1 Kudo
If you have an SSD in the node (or, oftentimes, a RAID 1 mirror) and it's large enough, YTS database would be a good candidate for putting on it.
... View more
10-08-2015
02:53 PM
1 Kudo
Sure! Hosts -> click into the node -> Host Actions -> Delete Host. It may take some time if data blocks for HDFS need to be moved away from that host.
... View more
10-08-2015
02:50 PM
1 Kudo
Linking your cross-post in another space, there was a discussion going. http://community.hortonworks.com/questions/953/can-nifi-be-used-to-pipe-the-data-from-oracle-data.html
... View more
10-08-2015
02:46 PM
1 Kudo
Alex, I think API is a no-go for bulk ingest. You should be looking at native admin tools in SP and Documentum to do the bulk. Otherwise their APIs (REST, SOAP, Java) may not perform at the desired level.
... View more
10-08-2015
01:04 PM
5 Kudos
We have created this write-up some time ago, might be useful: https://cwiki.apache.org/confluence/display/TEZ/How+initial+task+parallelism+works
... View more
10-08-2015
01:00 PM
Thanks, Bosco. Definitely, worth mentioning it in the docs. Even simple stuff, e.g. if port and path are the same or different, etc. Ping me offline so we can track this update, please.
... View more
10-06-2015
12:47 PM
14 Kudos
A series of examples and flow files: https://github.com/xmlking/nifi-examples NiFi Examples Apache NiFi example flows. collect-stream-logs This flow shows workflow for log collection, aggregation, store and display. Ingest logs from folders. Listen for syslogs on UDP port. Merge syslogs and drop-in logs and persist merged logs to Solr for historical search. Dashboard: stream real-time log events to dashboard and enable cross-filter search on historical logs data. csv-to-json This flow shows how to convert a CSV entry to a JSON document using ExtractText and ReplaceText. decompression This flow demonstrates taking an archive that is created with several levels of compression and then continuously decompressing it using a loop until the archived file is extracted out. http-get-route his flow pulls from a web service (example is nifi itself), extracts text from a specific section, makes a routing decision on that extracted value, prepares to write to disk using PutFile. invoke-http-route This flow demonstrates how to call an HTTP service based on an incoming FlowFile, and route the original FlowFile based on the status code returned from the invocation. In this example, every 30 seconds a FlowFile is produced, an attribute is added to the FlowFile that sets q=nifi, the google.com is invoked for that FlowFile, and any response with a 200 is routed to a relationship called 200. retry-count-loop This process group can be used to maintain a count of how many times a flowfile goes through it. If it reaches some configured threshold it will route to a 'Limit Exceeded' relationship otherwise it will route to 'retry'. Great for processes which you only want to run X number of times before you give up. split-route This flow demonstrates splitting a file on line boundaries, routing the splits based on a regex in the content, merging the less important files together for storage somewhere, and sending the higher priority files down another path to take immediate action. twitter-garden-hose This flow pulls from Twitter using the garden hose setting; it pulls out some basic attributes from the Json and then routes only those items that are actually tweets. twitter-solr This flow shows how to index tweets with Solr using NiFi. Pre-requisites for this flow are NiFi 0.3.0 or later, the creation of a Twitter application, and a running instance of Solr 5.1 or later with a tweets collection. Here are sample steps to set this up (along with Banana dashboard) on HDP Sandbox. Other examples https://github.com/hortonworks-gallery/nifi-templates
... View more
Labels:
10-06-2015
12:43 PM
3 Kudos
Hi, let me make sure I understand the environment. According to https://www.elastic.co/guide/en/logstash/2.0/plugins-inputs-jdbc.html there's nothing in the jdbc plugin to track incremental runs (e.g. only pick up data added/changed since the last run). This has to be built into the query itself (a generic mechanism). For this matter NiFi can absolutely serve the purpose, not sure there would be much that LogStash would bring on top. There is also another use case, when Oracle transaction log is being followed for real-time replication, which is a very different architecture. E.g. Oracle GoldenGate is one of the best products in its class (I think it works with other DBs as well).
... View more
10-06-2015
12:35 PM
1 Kudo
Actually memory is #3 in this list. You will probably be concerned with available storage space for content repository before running out of memory. E.g. NiFi, even in big deployments feels comfortable with 96GB available memory, which is below an average server RAM today.
... View more
10-06-2015
02:26 AM
Hi, the reference Hadoop KMS implementation has a REST API https://hadoop.apache.org/docs/current/hadoop-kms/index.html Is there anything like that for Ranger KMS? Given that Ranger itself has a complete REST API, I would expect the same for KMS, but I don't see any mention in here http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.0/bk_Ranger_KMS_Admin_Guide/content/ch_ranger_kms_overview.html
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Ranger