Member since
02-09-2016
40
Posts
14
Kudos Received
0
Solutions
05-09-2020
01:20 AM
Hi bigdatasuperno, I need a beeline shell script which should connect hive url and print the no. of tables in the db and print the record count of each table in the output file. Your help will greatly appreciated. I belongs to different technology. Regards, Vin
... View more
05-29-2018
02:16 PM
Thanks @Aditya Sirna for the swift response.
... View more
06-07-2017
09:09 PM
1 Kudo
@bigdata.neophyte Here are a few answers for you: NiFi can be interacted with via the UI as well as its REST API. The API is documented at https://nifi.apache.org/docs/nifi-docs/rest-api/index.html. NiFi is primarily a data flow tool whereas Kafka is a broker for a pub/sub type of use pattern. Kafka is frequently used as the backing mechanism for NiFi flows in a pub/sub architecture, so while they work well together they provide two different functions in a given solution. NiFi does have a visual command and control mechanism, while Kafka does not have a native command and control GUI Apache Atlas, Kafka, and NiFi all can work together to provide a comprehensive lineage / governance solution. There is a high level architecture slide at https://hortonworks.com/apache/atlas/#section_2 as well as a tutorial that might help this make more sense at https://hortonworks.com/hadoop-tutorial/cross-component-lineage-apache-atlas/. Data prioritization, back pressure, and balancing latency and throughput are all within NiFi's many strong points and these can be leveraged easily. Kafka does really not provide data prioritization. Security aspects of both Kafka and NiFi are tightly integrated with Apache Ranger, take a look at https://hortonworks.com/apache/ranger/ for additional details. Hope this helps, and please accept the answer if this was helpful.
... View more
08-24-2018
11:58 AM
Hi @sunile.manjee I see you mention hive2 in connection but it says connected to Connected to: Apache Hive (version 1.2...) How can we connect to Hive-2.1.*
... View more
10-31-2016
05:07 PM
Thanks @Kuldeep Kulkarni
... View more
10-26-2016
04:19 PM
Another alternative would be to use the YARN REST API to submit the application: With the New Application API, you can obtain an application-id which can then be used as part of the Cluster Submit Applications API to submit applications. curl -X POST http://<resource_manager>:8088/ws/v1/cluster/apps/new-application Reference: Resource Manager REST API Documentation
... View more
09-23-2016
12:36 PM
3 Kudos
This is a great guide to what gets installed where on HDP: https://community.hortonworks.com/articles/16763/cheat-sheet-and-tips-for-a-custom-install-of-horto.html You will notice that Kafka should be installed within the cluster and is best dedicated to its own nodes. As a side note, Hortwonworks Data Flow (HDF) is a separate distribution/product provided by Hortonworks. It packages Kafka along with NiFi, Storm and Ambari and excels at acquiring, inspecting, routing, transforming, analyizing data in motion from a diverse number of sources (ranging from sensors to databases), which is typically outputted in Hadoop. Exciting technology and a lot to talk ... check it out: http://hortonworks.com/products/data-center/hdf/
... View more
08-22-2016
02:18 PM
@bigdata.neophyte -t option seems to not available in 2.7.1 version which is included with HDP 2.4.2. This option is included in hadoop 2.7.3 which is available with HDP 2.5 Tech preview . [hdfs@hdp1 ~]$ hadoop version
Hadoop 2.7.3.2.5.0.0-1133
Subversion git@github.com:hortonworks/hadoop.git -r 93bf28063ef319be6833d3d6f117d44e0b6b8fa9
Compiled by jenkins on 2016-08-03T11:38Z
Compiled with protoc 2.5.0
From source with checksum 1aed9e48ca6f7cd4ada3a36b1cd5feb
This command was run using /usr/hdp/2.5.0.0-1133/hadoop/hadoop-common-2.7.3.2.5.0.0-1133.jar
[hdfs@hdp1 ~]$ hdfs dfs -ls -t -help
-ls: Illegal option -help
Usage: hadoop fs [generic options] -ls [-C] [-d] [-h] [-q] [-R] [-t] [-S] [-r] [-u] [<path> ...]
[hdfs@hdp1 ~]$ hdp-select versions
2.5.0.0-1133
... View more
08-14-2018
08:42 PM
Hi Team , Is there any way to test multiple files exist or not in single command?? like below.. pls suggest hadoop fs -test -e filename1 filename2
... View more