Member since
09-18-2015
3274
Posts
1159
Kudos Received
426
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2140 | 11-01-2016 05:43 PM | |
6518 | 11-01-2016 05:36 PM | |
4157 | 07-01-2016 03:20 PM | |
7104 | 05-25-2016 11:36 AM | |
3446 | 05-24-2016 05:27 PM |
03-18-2016
08:40 PM
Is it possible to flat xml file using NiFi? For example: Link
... View more
Labels:
- Labels:
-
Apache NiFi
-
Cloudera DataFlow (CDF)
03-16-2016
06:25 PM
1 Kudo
@Sree Venkata One stop tool is NiFi/HDF http://hortonworks.com/webinar/introducing-hortonworks-dataflow/ for example: Does HDF address delta load from Oracle database to HDFS? A: HDF powered by Apache NiFi does support interaction with databases though it is narrowly focused. The SQL Processor set available today does not yet offer a complete change data capture solution. At a framework level, this use case is readily supportable. We expect to see increasing priority on providing a high quality user experience around database-oriented change data capture as we move forward. How different is this from Flume, Kafka, or other data ingestion frameworks? A: Kafka is a messaging system. Messaging systems generally are focused on providing mail-box like semantics whereby the ‘provider’ of data is decoupled from the ‘consumer’ of that data at least on a physical connectivity level. In enterprise dataflows, however, there are many other forms of decoupling to consider that are also critical. Protocol, format, schema, priority, and interest are all examples of important ‘separations of concern’ to consider. HDF powered by Apache NiFi is designed to address all of these forms of decoupling. In so doing, NiFi is often used with a system like Kafka, which is aimed at addressing one of those forms of decoupling but does so in a manner that can lead to very high performance under specific usage patterns. Kafka doesn’t address the user experience and real-time command and control aspects of the data lineage capabilities offered by HDF powered by Apache NiFi. The type of security that can be offered by messaging based systems will be largely limited to transport security, encryption of data at rest, and white-list style authorization to topics. HDF offers similar approaches as well but since it actually operates on and with the data it can also perform fine-grained security checks and rule-based contextual authorization. In the end, these systems are designed to tackle different parts of the data flow problem are often used together as a more powerful whole. The comparison of HDF (and a flow using the “GetFile” processor along with a “PutHDFS” processor) to Flume is a more direct comparison in that they were designed to address very similar use cases. HDF offers data provenance as well as a powerful and intuitive user experience with a drag-and-drop UI for interactive command and control. From the management and data tracking perspectives HDF and Flume offer quite a different feature set. That said, Flume has been used considerably for some time now and, as is true with any system, the goal of HDF is to integrate with it in the best manner possible. As a result, HDF powered by Apache NiFi supports running Flume sources and sinks right in the flow itself. You can now wire many Flume sources and do so in a way that combines Flume’s configuration file approach with NiFi’s UI driven approach offering a best of both worlds solution.
... View more
03-16-2016
10:34 AM
@Roberto Sancho See this https://github.com/hortonworks/HDP-Public-Utilities/blob/master/Backup/qa.backup_etc_dbs.sh#L65
... View more
03-15-2016
06:33 PM
1 Kudo
I have installed only zookeeper.
... View more
03-15-2016
06:32 PM
1 Kudo
@Sunile Manjee You can ..see this
... View more
03-15-2016
02:41 PM
1 Kudo
@Ram D Clone the vm if ambari is running in vm If it's on baremetal then backup the database and follow this http://docs.hortonworks.com/HDPDocuments/Ambari-2.1.1.0/bk_ambari_reference_guide/content/_back_up_current_data.html AMS stores the data in embeded HBASE instance. See this https://cwiki.apache.org/confluence/display/AMBARI/AMS+-+distributed+mode You need to backup AMS
... View more
03-15-2016
10:23 AM
1 Kudo
@Kuldeep Kulkarni https://hadoop.apache.org/docs/r2.7.1/hadoop-project-dist/hadoop-hdfs/HDFSHighAvailabilityWithQJM.html You can try this shell - run an arbitrary shell command to fence the Active NameNode The shell fencing method runs an arbitrary shell command. It may be configured like so: <property>
<name>dfs.ha.fencing.methods</name>
<value>shell(/path/to/my/script.sh arg1 arg2 ...)</value>
</property>
The string between ‘(’ and ‘)’ is passed directly to a bash shell and may not include any closing parentheses. The shell command will be run with an environment set up to contain all of the current Hadoop configuration variables, with the ‘_’ character replacing any ‘.’ characters in the configuration keys. The configuration used has already had any namenode-specific configurations promoted to their generic forms – for example dfs_namenode_rpc-address will contain the RPC address of the target node, even though the configuration may specify that variable asdfs.namenode.rpc-address.ns1.nn1. You can write your own custom scripts. Also, you can check with the OS vendor like "https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/5/html/Cluster_Suite_Overview/s2-fencing-overview-CSO.html"
... View more
03-14-2016
07:55 PM
1 Kudo
@Saurabh Kumar That's exactly my point was 🙂 https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/6/html/Deployment_Guide/about-sssd.conf.html I use this https://github.com/hortonworks-gallery/ambari-nslcd-service So..As you can see in the demo, we can map AD groups. https://community.hortonworks.com/articles/17135/yarn-queues-and-ad-group-mapping.html
... View more
03-14-2016
07:33 PM
1 Kudo
@bganesan Thanks!
... View more
03-14-2016
07:25 PM
1 Kudo
@Mark Lavin odbc.ini The DataDirect ODBC driver manager looks for a property that you can add to the Data Source Name (DSN) entry for the 3rd party driver in the .odbc.ini file to indicate if the driver is using 2 or 4 bytes for unicode.
Add the following property after the "driver=" line for the failing DSN in the .odbc.ini file:
DriverUnicodeType=1
then test the failing scenario. If the error still occurs, then change the line to:
DriverUnicodeType=2
... View more