Member since
02-17-2017
14
Posts
2
Kudos Received
0
Solutions
03-13-2018
05:08 PM
Download HDF 3.1 HDF will never have HBase. HBase requires the infrastructure of HDP. Download HDP 2.6 and you can combine HDF and HDP to have all your purposes. HDP 2.6 has Kafka, Storm and HBase. You can easily add NiFi to that. https://docs.hortonworks.com/HDPDocuments/Ambari-2.6.1.5/bk_ambari-installation/content/ch_Getting_Ready.html then add nifi https://docs.hortonworks.com/HDPDocuments/HDF3/HDF-3.1.1/bk_installing-hdf-and-hdp/content/ch_install-ambari.html
... View more
10-18-2017
06:51 AM
@Ramesh Prasad Some more suggestions 1) For JS code use :WebStorm ( you will get EAP version from official website with 30 days trial) 2) Remember to use Coding guidelines as mentioned here https://cwiki.apache.org/confluence/display/AMBARI/Coding+Guidelines+for+Ambari 3) when you are running the project always skip rat checks or it might fail in most of cases and takes lot of time: mvn clean install package -Dmaven.test.skip=true -Drat.skip -DskipTests 4) if you are changing only client side changes( Javascript in ambari-web folder ) , you can only run the project on ambari-web folder using the abouve command and the output will be generated on ambari/ambari/ambari-web/public/javascripts folder you can copy that for debbuging in your server folder /usr/lib/ambari-server/web/javascripts
... View more
10-13-2017
05:27 PM
Ah, i was making mistake with the keytabs and other settings. thank you
... View more
09-06-2017
02:09 PM
@Ramesh Prasad
It can't be done via any regular/standard and automated way, also any other procedure is not recommended. So I think the interviewer just wanted to make sure you understand the basic MR/Hadoop concepts.
... View more
07-28-2017
10:51 AM
1 Kudo
Time taking for Query execution depends on multiple factors 1. Mainly the Hive query design, joins and the columns being pulled 2. YARN/TEZ container size allocated, depends where you are running 3. Check the queue you are running your job, check if queue is free to answer your question on why one of the reducer is taking 1000 tasks please the hive.exec.reducers.max value defined If you want to play and modify the number of reducers, try changing the value of hive.exec.reducers.bytes.per.reducer(preferably assign a smaller, as this value is inversely proportional to number of reducers)
... View more
07-02-2017
01:25 PM
Hi @Ramesh Prasad, You may want to start looking at Apache Calcite. Calcite is the primary CBO behind Hive execution.
... View more
03-21-2017
03:13 PM
Amazing answer!
... View more
02-17-2017
07:10 PM
Thank you
... View more