Member since
08-08-2017
1652
Posts
30
Kudos Received
11
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1998 | 06-15-2020 05:23 AM | |
| 16442 | 01-30-2020 08:04 PM | |
| 2145 | 07-07-2019 09:06 PM | |
| 8340 | 01-27-2018 10:17 PM | |
| 4728 | 12-31-2017 10:12 PM |
01-17-2021
09:21 AM
we have HDP cluster version `2.6.5` with `8` data nodes , all machines are installed on rhel 7.6 version HDP cluster is based amabri platform version - `2.6.1` each data-node ( worker machine ) include two disks and each disk size is 1.8T when we access the data-node machines we can see differences between the size of the disks for example on the first data-node the size is : ( by `df -h` ) /dev/sdb 1.8T 839G 996G 46% /grid/sdc /dev/sda 1.8T 1014G 821G 56% /grid/sdb on the second data-node the size is: /dev/sdb 1.8T 1.5T 390G 79% /grid/sdc /dev/sda 1.8T 1.5T 400G 79% /grid/sdb on the third data-node th size is: /dev/sdb 1.8T 1.7T 170G 91% /grid/sdc /dev/sda 1.8T 1.7T 169G 91% /grid/sdb and so on the big question is why HDFS not perform the re-balance on the HDFS disks? *for example expected results on all disks should be with the same size on all datanodes machines* why is the used size differences between `datanode1` to `datanode2` to `datanode3` etc ? any advice about the tune parameters in HDFS that can help us? *because its very critical when one disk is reached `100%` size and the other are more small as `50%`*
... View more
Labels:
- Labels:
-
HDFS
01-17-2021
09:19 AM
we installed small HDP cluster with one data-node machine HDP version is `2.6.5` and ambari version is `2.6.1` so this is new cluster that contain two name-node with only one data-node ( worker machine ) the interesting behavior that we see is that increasing of `under replica` on ambari dashboard , for now the number is `15000` under replica blocks as we know the most root cause of this problem is network issues between name node to data-node but this isn't the case in our hadoop cluster we can also decrease the under replica by the following procedure su - <$hdfs_user> bash-4.1$ hdfs fsck / | grep 'Under replicated' | awk -F':' '{print $1}' >> /tmp/under_replicated_files -bash-4.1$ for hdfsfile in `cat /tmp/under_replicated_files`; do echo "Fixing $hdfsfile :" ; hadoop fs -setrep 3 $hdfsfile; done but we not want to do it because under replica problem should not happens from beginning and maybe need to tune some HDFS parameters , but we not sure about this please let us know about any advice that can help us
... View more
Labels:
- Labels:
-
HDFS
10-14-2020
08:48 AM
now I want to add the component as HBASE master and RegionServers and Phoenix Query Servers
... View more
10-14-2020
08:47 AM
this is what I have now in ambari
... View more
10-14-2020
08:44 AM
actually what I want to do is to use API to install HBASE , with all hbase component so when you wrote the OOZIE procedure , what is the API to add HBASE insted OOZIE?
... View more
10-14-2020
08:42 AM
I have already the files ls /etc/hbase/conf/ hadoop-metrics2-hbase.properties hbase-env.cmd hbase-env.sh hbase-policy.xml hbase-site.xml log4j.properties regionservers
... View more
10-14-2020
08:24 AM
I have another issue that is related but it happens when I install component from amabri from host I do --> Add --> install hbase master but I get the following Traceback (most recent call last):
File "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/before-INSTALL/scripts/hook.py", line 37, in <module>
BeforeInstallHook().execute()
File "/usr/lib/ambari-agent/lib/resource_management/libraries/script/script.py", line 375, in execute
method(env)
File "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/before-INSTALL/scripts/hook.py", line 28, in hook
import params
File "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/before-INSTALL/scripts/params.py", line 110, in <module>
hbase_user_dirs = format("/home/{hbase_user},/tmp/{hbase_user},/usr/bin/{hbase_user},/var/log/{hbase_user},{hbase_tmp_dir}")
File "/usr/lib/ambari-agent/lib/resource_management/libraries/functions/format.py", line 95, in format
return ConfigurationFormatter().format(format_string, args, **result)
File "/usr/lib/ambari-agent/lib/resource_management/libraries/functions/format.py", line 59, in format
result_protected = self.vformat(format_string, args, all_params)
File "/usr/lib64/python2.7/string.py", line 549, in vformat
result = self._vformat(format_string, args, kwargs, used_args, 2)
File "/usr/lib64/python2.7/string.py", line 582, in _vformat
result.append(self.format_field(obj, format_spec))
File "/usr/lib64/python2.7/string.py", line 599, in format_field
return format(value, format_spec)
File "/usr/lib/ambari-agent/lib/resource_management/libraries/script/config_dictionary.py", line 73, in __getattr__
raise Fail("Configuration parameter '" + self.name + "' was not found in configurations dictionary!")
resource_management.core.exceptions.Fail: Configuration parameter 'hbase-env' was not found in configurations dictionary!
... View more
10-13-2020
02:13 PM
we have ambari cluster and HDP version 2.6.5
we want to add the service OOZIE on specific nodes in our cluster
for example - we have 10 workers machines in the HDP cluster
so we want to add he service OOZIE on each worker machine
from documentation we found the following API that really add the service to ambari
curl --user admin:admin -H "X-Requested-By: ambari" -i -X POST -d '{"ServiceInfo":{"service_name":"OOZIE"}}' http://localhost:8080/api/v1/clusters/HDP/services
but we not succeeded to improve the API in order to install the service OOZIE on each worker node
the final target is to add the service to amabri while each worker node in amabri should have the service OOZIE
any ideas how to continue from this stage?
... View more
Labels:
10-12-2020
02:47 AM
little question why just not stop the service - HDFS on each new data node and set it to maintenance mode ?
... View more
10-09-2020
12:59 AM
you said that not need to run it , but the post that I mentioned say to run it , so what is right
... View more