Member since
01-19-2017
3679
Posts
632
Kudos Received
372
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1025 | 06-04-2025 11:36 PM | |
| 1582 | 03-23-2025 05:23 AM | |
| 798 | 03-17-2025 10:18 AM | |
| 2862 | 03-05-2025 01:34 PM | |
| 1874 | 03-03-2025 01:09 PM |
08-02-2019
11:18 PM
@FA Use a ReplaceText processor which replaces \A\n|\n*\s*(?=\n) with '' (empty replacement value). The search regex looks for: \n*\s*(?=\n) - newline (0 or more) followed by whitespace (0 or more) followed by a newline.
... View more
08-02-2019
10:18 PM
@Matas Mockus This is a duplicate posting I responded to the initial thread. http://community.hortonworks.com/answers/249938/view.html Did you check the response please either merge or delete this post as it will be difficult to follow the 2 threads!
... View more
08-02-2019
08:46 PM
@Matas Mockus The below error comes up in your ambari-agent logs , can you check your hiveServer2 hostname, you might need to adjust in Ambari UI Execute['! beeline -u 'jdbc:hive2://:10000/;transportMode=binary;auth=noSasl' It should look like this Execute['! beeline -u 'jdbc:hive2://{some_host_FQDN}:10000/;transportMode=binary;auth=noSasl' Please try that and revert
... View more
07-31-2019
04:06 PM
@Madhura Mhatre The hdfs balancer utility that analyzes block placement and balances data across the DataNodes. The balancer moves blocks until the cluster is deemed to be balanced. A threshold parameter is a float number between 0 and 100 (12.5 for instance). From the average cluster utilization (about 50% in the graph below), the balancer process will try to converge all data nodes' usage in the range [average - threshold, average + threshold]. In the current example: - Higher (average + threshold): 60% if run with the default threshold (10%) - Lower (average - threshold): 40% You can easily notice that the smaller your threshold, the more balanced your data nodes will be. For very small threshold, the cluster may not be able to reach the balanced state if other clients concurrently write and delete data in the cluster. A threshold of 15 should be okay $ hdfs balancer -threshold 15 or by giving the list of datanodes $ hdfs balancer -threshold 15 -include hostname1,hostname2,hostname3 HTH
... View more
07-31-2019
01:36 PM
@Ray Teruya How many hosts do you have in your cluster? Can you share your zookeeper logs and your /etc/hosts? HTH
... View more
07-30-2019
11:00 PM
1 Kudo
@Haijin Li To be able to help, it's a good idea to always share your cluster config and version. I can see you are referencing HDP 2.3 documentation which is obsolete and personally I am wondering why you are running this version? Some parameters like hive.users.in.admin.role are not default values in hive_site.xml , so you will need to add them in Custom hive-site these are considered custom site values see attached screenshot Haijin2.png Authorization: SQL-Standard Based (SQLStdAuth) Custom values The above UI is specific to HDP 3.1.0.0 running Ambari Version 2.7.3.0 but even with earlier versions of Ambari you can filter see the ARROW HTH
... View more
07-30-2019
09:05 PM
@Ray Teruya Start-all-services-from-Ambari Start all services. Use Ambari UI > Services > Start All to start all services at once. In Ambari UI > Services you can start, stop, and restart all listed services simultaneously. In Services, click ... and then click Start All. The first place to check for start failures or success in /var/logs/zookeeper/zookeeper.log or zookeeper-zookeeper-server-[hostname].out According to HWX documentation make sure to manually start the Hadoop services in this prescribed order 1. How do I check what services need to be "up and running" before restarting the next one? Is there any place where I can see the dependency? The above gives you the list and order of dependency 2. Do I need 2 ZooKeeper servers up and running? The first one is running in localhost but the 2nd one runs in a different machine. If I actually need them both, how can I check what was wrong in the second one? If you are not run an HA configuration a single zookeeper suffice, but if you want to emulate a production environment with many data nodes to enable [HA Namenode or RM] you MUST have at least 3 zookeepers to avoid the split-brain phenomenon Hope that helps
... View more
07-28-2019
09:15 PM
2 Kudos
@Figo C The reason is by design NiFi as a client communicates with HDFS Namenode on port 8020 and it returns the location of the files using the data node which is a private address. Now that both your HDF and HDF are sandboxes I think you should switch both to host-only-adapter your stack trace will be a statement that the client can’t connect to the data node, and it will list the internal IP instead of 127.0.0.1. That causes the minReplication issue, etc. Change the HDP and HDF sandbox VM network settings from NAT to Host-only Adapter. Here are the steps: 1. Shutdown gracefully the HDF sandbox 2. Change Sandbox VM network from NAT to Host-only Adapter It will automatically pick your LAN or wireless save the config. 3. Restart Sandbox VM 4. Log in to the Sandbox VM and use ifconfig command to get its IP address, in my case 192.168.0.45 5. Add the entry in /etc/hosts on my host machine, in my case: 192.168.0.45 sandbox.hortonworks.com 6. Check connectivity by telnet: telnet sandbox.hortonworks.com 8020 7. Restart NiFi (HDF) By default HDFS clients connect to DataNodes using the IP address provided by the NameNode. Depending on the network configuration this IP address may be unreachable by the clients. The fix is letting clients perform their own DNS resolution of the DataNode hostname. The following setting enables this behavior. If the above still fails make the below changes in the hdfs-site.xml that NiFi is using set dfs.client.use.datanode.hostname to true in your <property>
<name>dfs.client.use.datanode.hostname</name>
<value>true</value>
<description>Whether clients should use datanode hostnames when
connecting to datanodes.
</description>
</property> Hope that helps
... View more
07-26-2019
08:46 PM
@Figo C Can you check running status/logs of datanode/namenode and copy-paste it here. Did ou add these 2 files to your nifi config Core-site.xml and hdfs-site.xml
... View more
07-24-2019
11:08 PM
1 Kudo
@Michael Bronson Here is an HCC doc that could help uninstall completely HDP for clean Fresh Install Hope that helps
... View more