Member since
02-08-2016
793
Posts
669
Kudos Received
85
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2583 | 06-30-2017 05:30 PM | |
3338 | 06-30-2017 02:57 PM | |
2788 | 05-30-2017 07:00 AM | |
3239 | 01-20-2017 10:18 AM | |
6624 | 01-11-2017 02:11 PM |
07-22-2020
05:45 AM
The given solution is certainly not true unfortunately. In HDFS a given block if it is open for write, then consumes 128MB that is true, but as soon as the file is closed, the last block of the file is counted just by the length of the file. So if you have a 1KB file, that consumes 3KB disk space considering replication factor 3, and if you have a 129MB file that consumes 387MB disk space again with replication factor 3. The phenomenon that can be seen in the output was most likely caused by other non-DFS disk usage, that made the available disk space for HDFS less, and had nothing to do with the file sizes. Just to demonstrate this with a 1KB test file: # hdfs dfs -df -h Filesystem Size Used Available Use% hdfs://<nn>:8020 27.1 T 120 K 27.1 T 0% # fallocate -l 1024 test.txt # hdfs dfs -put test.txt /tmp # hdfs dfs -df -h Filesystem Size Used Available Use% hdfs://<nn>:8020 27.1 T 123.0 K 27.1 T 0% I hope this helps to clarify and correct this answer.
... View more
07-01-2020
09:43 AM
Great article! I faced the following error while trying adding data to ldap (Step 13.) # ldapadd -x -W -D "cn=Manager,dc=example,dc=com" -f /root/ldap/base.ldif Enter LDAP Password: adding new entry "dc=example,dc=com" ldap_add: Invalid syntax (21) additional info: objectClass: value #1 invalid per syntax After some research, found that we need to add the cosine and nis LDAP schemas before running the preceding command. # ldapadd -Y EXTERNAL -H ldapi:/// -f /etc/openldap/schema/cosine.ldif # ldapadd -Y EXTERNAL -H ldapi:/// -f /etc/openldap/schema/nis.ldif # ldapadd -Y EXTERNAL -H ldapi:/// -f /etc/openldap/schema/inetorgperson.ldif
... View more
04-13-2020
01:57 PM
Can this be set in "Accumulo Service Advanced Configuration Snippet (Safety Valve) for accumulo-site.xml"
... View more
04-02-2020
01:47 AM
you shoul either use ssh (id_rsa) or install ambari-agent on all hosts after ambari agent is installed (change the /etc/ambari-agent/conf/ambari-agent.ini ) and set the hostname of ambari-server machine for all machines then change the /etc/hosts file ( ip_of_machine1 name_of_machine1 ip_of_machine2 name_of_machine2....) then ==> ambari-agent restart for all machines trusted and tested
... View more
09-20-2019
06:28 AM
I found the Ranger KMS Admin Guide for HDP 2.4.0, hopefully this is what you are looking for.
... View more
06-28-2017
04:16 PM
1 Kudo
SYMPTOM: While performing HDP upgrade [HDP-2.3.0 upgrade to HDP-2.5.3] there was error for which we had to start ambari server. But ambari server startup failed with below error - ERROR: Caused by: Exception [EclipseLink-116] (Eclipse Persistence Services - 2.6.2.v20151217-774c696): org.eclipse.persistence.exceptions.DescriptorException
Exception Description: No conversion value provided for the value [UPGRADED] in field [host_version.state].
Mapping: org.eclipse.persistence.mappings.DirectToFieldMapping[state-->host_version.state]
Descriptor: RelationalDescriptor(org.apache.ambari.server.orm.entities.HostVersionEntity --> [DatabaseTable(host_version)])
at org.eclipse.persistence.exceptions.DescriptorException.noFieldValueConversionToAttributeValueProvided(DescriptorException.java:1066)
at org.eclipse.persistence.mappings.converters.ObjectTypeConverter.convertDataValueToObjectValue(ObjectTypeConverter.java:226)
at org.eclipse.persistence.mappings.converters.EnumTypeConverter.convertDataValueToObjectValue(EnumTypeConverter.java:141)
at org.eclipse.persistence.mappings.foundation.AbstractDirectMapping.getObjectValue(AbstractDirectMapping.java:616)
at org.eclipse.persistence.mappings.foundation.AbstractDirectMapping.valueFromRow(AbstractDirectMapping.java:1220)
at org.eclipse.persistence.mappings.DatabaseMapping.readFromRowIntoObject(DatabaseMapping.java:1539)
at org.eclipse.persistence.internal.descriptors.ObjectBuilder.buildAttributesIntoObject(ObjectBuilder.java:462)
at org.eclipse.persistence.internal.descriptors.ObjectBuilder.buildObject(ObjectBuilder.java:1005)
at org.eclipse.persistence.internal.descriptors.ObjectBuilder.buildWorkingCopyCloneNormally(ObjectBuilder.java:899)
at org.eclipse.persistence.internal.descriptors.ObjectBuilder.buildObjectInUnitOfWork(ObjectBuilder.java:852)
at org.eclipse.persistence.internal.descriptors.ObjectBuilder.buildObject(ObjectBuilder.java:735)
at org.eclipse.persistence.internal.descriptors.ObjectBuilder.buildObject(ObjectBuilder.java:689)
at org.eclipse.persistence.queries.ObjectLevelReadQuery.buildObject(ObjectLevelReadQuery.java:805)
at org.eclipse.persistence.queries.ReadAllQuery.registerResultInUnitOfWork(ReadAllQuery.java:962)
at org.eclipse.persistence.queries.ReadAllQuery.executeObjectLevelReadQuery(ReadAllQuery.java:573)
at org.eclipse.persistence.queries.ObjectLevelReadQuery.executeDatabaseQuery(ObjectLevelReadQuery.java:1175)
at org.eclipse.persistence.queries.DatabaseQuery.execute(DatabaseQuery.java:904)
at org.eclipse.persistence.queries.ObjectLevelReadQuery.execute(ObjectLevelReadQuery.java:1134)
ROOT CAUSE: There was manual entry in ambari db table "host_version" where the 'state' was modified to "UPGRADED" as shown below which cause the issue- ambari=> select * from host_version;
id | repo_version_id | host_id | state
----+-----------------+---------+---------
1 | 1 | 1 | CURRENT
2 | 1 | 3 | CURRENT
3 | 2 | 2 | UPGRADED
RESOLUTION: Steps followed for resolution - 1. Took ambari DB backup 2. Check the latest repo_version_id using below command - select (repo_version_id, version) from repo_version ;
row
---------------
(1, 2.3.0.1-15)
(2, 2.3.5.3-8)
3. Check for latest version and repo_version_id from above command and modified respective repo_version_id in "host_version" table pointing the status to CURRENT and made rest of the entries to INSTALLED. So the output should look like below - ambari=> select * from host_version;
id | repo_version_id | host_id | state
----+-----------------+---------+---------
1 | 1 | 1 | INSTALLED
2 | 1 | 3 | INSTALLED
3 | 2 | 2 | CURRENT
4. Restart ambari-server.
... View more
Labels:
06-28-2017
03:50 PM
1 Kudo
SYMPTOM: During upgrade from HDP 2.4.3 to HDP2.6.0 the express upgrade failed on the step "Yarn NodeMangers / Restarting NodeManger." ERROR: 2017-06-22
15:02:22,791 INFO recovery.ZKRMStateStore
(ZKRMStateStore.java:runWithRetries(1227)) - Exception while executing a ZK
operation. org.apache.zookeeper.KeeperException$AuthFailedException:
KeeperErrorCode = AuthFailed for /rmstore at
org.apache.zookeeper.KeeperException.create(KeeperException.java:123) at
org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at
org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
ROOT CAUSE: The issue was with property Ambari-Yarn-Configs-"yarn-resourcemanager-zk-address" was set to localhost:2181 RESOLUTION: Modified address to "{zk-server-hostname}:2181" resolved the issue. After which node managers were able to start successfully.
... View more
Labels:
05-30-2017
07:33 AM
NameNode HA has been enabled successfully.
... View more