Member since
01-19-2017
3679
Posts
632
Kudos Received
372
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 908 | 06-04-2025 11:36 PM | |
| 1509 | 03-23-2025 05:23 AM | |
| 744 | 03-17-2025 10:18 AM | |
| 2680 | 03-05-2025 01:34 PM | |
| 1786 | 03-03-2025 01:09 PM |
05-28-2018
08:30 PM
@Knows NotMuch That's exactly the desired output "Ambari Server 'setup' completed successfully". Now if you start your hive 2server you shouldn't get the previous error. Please let me know after attempting the H2S startup.
... View more
05-27-2018
11:25 AM
1 Kudo
@Kenneth K Below are 2 links that could help you do a semi-automated deployment of HDF aka Nifi Automate Deployment of HDF 2.x/3.0 clusters using Ambari blueprints Setting Nifi on AWS using blueprints Hope that helps
... View more
05-26-2018
10:46 PM
@Tanu Sood Indeed this has become a very unbearable condition for anyone taking an HW exam. I am just surprised and in astonishment that despite all the complaints HW exam team have NOT resolved this damn issue. Every week there is a minimum of 4 complaints and the numbers are growing I sat for my Oracle Certified Professional (OCP), CCNA, CCNP but I never got the issues I encountered when I tried HDPCA which after 1 hour of tussling I gave up, the environment was horribly slow and unstable loss of connectivity and the Proctor was an asshole to crown it all. Panning my office desk almost 4 times,I had to remove other screens and keyboards, pens just name it ..... Thank God I passed it despite not completing the questions because for sure HW wasn't going to give me a second eCredit !!! I really don't know whether HW is aware of the frustration and how such situations can sway someone to go Cloudera....basically the underlying technology is the same save for a few components and it's not with such comments that HW can compete with others !!:-) I pray this time some HW employees in HCC take it seriously what a good product and a poor disastrous exam setup.
... View more
05-26-2018
09:39 PM
@Knows NotMuch When you install HDP and you don't intend to use the derby database as in most cases with MySQL you download mysql-connector-java.jar thats one of the preparatory steps on the node where you will be running Hive etc # yum install -y mysql-connector-java The location is /usr/share/java/mysql-connector-java.jar and when running the Ambari setup the jdbc location must be appended # ambari-server setup --jdbc-db=mysql --jdbc-driver=/usr/share/java/mysql-connector-java.jar Once you have done the above then Hive, Oozie,Ranger etc will function correctly Please have a look at this HCC How to Sync MySQL Connector jar across HDP components and validate that you didn't miss a step. If you found this answer addressed your question, please take a moment to log in and click the "Accept" link on the answer.
... View more
05-26-2018
07:59 AM
@sudi ts You need to copy the connector into the hadoop-client and hive-client location otherwise you will hit an error cp gcs-connector-latest-hadoop2.jar /usr/hdp/current/hadoop-client/lib/
cp gcs-connector-latest-hadoop2.jar /usr/hdp/current/hive-client/lib The below command should run successfully $ hdfs dfs -ls gs://bucket/ This should run fine, but the issue you are having is with permission for hdpuser1 you will need to correct by running $ hdfs dfs -chown hdpuser1 gs://bucket/ Now your create table should work, while logged in as hdpuser1 CREATE EXTERNAL TABLE test1256(name string,id int) LOCATION 'gs://bucket/'; Please let me know. If you found this answer addressed your question, please take a moment to log in and click the "Accept" link on the answer.
... View more
05-25-2018
11:36 PM
@Saurabh Srivastava I noticed you created 2 similar threads, please can you delete the other one.
... View more
05-25-2018
11:34 PM
@Saurabh Srivastava The maximum transmission unit (MTU) could have been set too low the default is 1500 bytes, change it to 9000 recommended see HCC document Typical HDP Cluster Network Configuration Best Practices do this for all the nodes in the cluster Check the current MTU setting with ifconfig or ip link list command under Linux look at the fourth line # /sbin/ifconfig
eth1 Link encap:Ethernet HWaddr 08:00:27:D8:06:8F
inet addr:192.168.0.171 Bcast:192.168.0.255 Mask:255.255.255.0
inet6 addr: fe80::a00:27ff:fed8:68f/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:786 errors:0 dropped:0 overruns:0 frame:0
TX packets:444 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1000
RX bytes:455353 (444.6 KiB) TX bytes:57860 (56.5 KiB) You can also use $ ip link list
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
2: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP qlen 1000
link/ether 08:00:27:d8:06:8f brd ff:ff:ff:ff:ff:ff As you see, MTU set to 1500 for eth1. You should optimize to let's say 9000 then you can use any one of the following commands to setup MTU: # ifconfig eth1 mtu 9000 or # ip link set dev eth1 mtu 9000 Verify that new MTU is setup with the following command: $ ip link list or $ /sbin/ifconfig Edit /etc/sysconfig/network-scripts/ifcfg-ethx (Red Hat Linux ) to permanently change the MTU,notice I have added MTU=9000 DEVICE=eth0
HWADDR=08:00:27:FF:AF:39
TYPE=Ethernet
UUID=4df8046e-59c0-4667-9a6e-daa61b83682c
ONBOOT=no
MTU=9000
NM_CONTROLLED=yes
BOOTPROTO=dhcp Restart the network on Redhat/Centos # service network restart Reference: https://www.pcmech.com/article/jumbo-frames/ Now retry you copy, Hadoop 2.0 should be good with big files Please let me know if the copy timing improved drastically
... View more
05-25-2018
04:08 PM
1 Kudo
@Winnie Philip Check the current limit cat /proc/sys/fs/file-max Run this command as root set it for example to 1million # sysctl -w fs.file-max=1000000 Then to permanently update it edit /etc/sysctl.conf and the below line fs.file-max = 1000000 Retry and the "Too many open files" shouldn't appear
... View more
05-24-2018
02:20 PM
@Mokkan Mok Edge nodes are the interface between the Hadoop cluster and the outside network. For this reason, at times referred to as gateway nodes. Most commonly, edge nodes are used to run client applications and cluster administration tools. The edge node should be the access point to your cluster for your developers/data analysts. Usually, you install all the client software on the edge node some of them are listed below MAPREDUCE2_CLIENT HDFS_CLIENT YARN_CLIENT ZOOKEEPER_CLIENT KERBEROS CLIENT HIVE CLIENT PIG CLIENT SQOOP CLIENT OOZIE CLIENT FALCON CLIENT INFRA SOLR CLIENT ATLAS METADATA SPARK2 CLIENT SPARK CLIENT With all the user's confined to the edge node securing the edge node is very important and easier to manage. Hope that helps in your decision
... View more