Member since
01-19-2017
3679
Posts
632
Kudos Received
372
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1562 | 06-04-2025 11:36 PM | |
| 2035 | 03-23-2025 05:23 AM | |
| 959 | 03-17-2025 10:18 AM | |
| 3645 | 03-05-2025 01:34 PM | |
| 2529 | 03-03-2025 01:09 PM |
02-09-2019
10:32 PM
1 Kudo
@christophe VALMIR Usually, after the restart give the process a minute or 2 for the processes to pick up. Please don't forget to vote a helpful answer as the issue is resolved,that way other HCC users can quickly find the solution when they encounter the same issue. HTH
... View more
02-08-2019
02:58 PM
The below steps describe how to change the Namenode log level while logged on as hdfs with the below steps, without the need to restart the namenode Get the current log level $ hadoop daemonlog -getlevel {namenode_host}:50070BlockStateChange Desired Output Connecting to http://{namenode_host}:50070/logLevel?log=BlockStateChange
SubmittedLogName:BlockStateChange
LogClass: org.apache.commons.logging.impl.Log4J
LoggerEffectiveLevel: INFO Change to DEBUG $ hadoop daemonlog -setlevel {namenode_host}:50070BlockStateChange DEBUG Desired Output Connecting to http://{namenode_host}:50070/logLevel?log=BlockStateChange&level=DEBUG
SubmittedLogName:BlockStateChange
LogClass: org.apache.commons.logging.impl.Log4J
LoggerSubmittedLevel: DEBUG
SettingLevel to DEBUG ...
EffectiveLevel: DEBUG Validate DEBUG mode $ hadoop daemonlog -getlevel {namenode_host}:50070BlockStateChange Desired Output Connecting to http://{namenode_host}:50070/logLevel?log=BlockStateChange
SubmittedLogName:BlockStateChange
LogClass: org.apache.commons.logging.impl.Log4J
LoggerEffectiveLevel: DEBUG You should be able to notice the logging level in namenode.log has been updated, without restarting the service. After finishing your diagnostics you can reset the logging level back to INFO Reset to INFO $ hadoop daemonlog -setlevel {namenode_host}:50070BlockStateChange INFO Desired Output Connecting to http://{namenode_host}:50070/logLevel?log=BlockStateChange&level=INFO
SubmittedLogName:BlockStateChange
LogClass: org.apache.commons.logging.impl.Log4J
LoggerSubmittedLevel: INFO
SettingLevel to INFO ...
EffectiveLevel: INFO Validate INFO $ hadoop daemonlog -getlevel {namenode_host}:50070BlockStateChange Output Connecting to http://{namenode_host}:50070/logLevel?log=BlockStateChange
SubmittedLogName:BlockStateChange
LogClass: org.apache.commons.logging.impl.Log4J
LoggerEffectiveLevel: INFO Happy hadooping !!!!
... View more
Labels:
02-08-2019
08:45 AM
@ram sriram If you found this answer addressed your question, please take a moment to log in and click the "accept" link on the answer. Can you tag me for "Can you please send me a document for Ambari installation on ubuntu thread" so I see the information you already received.
... View more
02-07-2019
11:02 PM
2 Kudos
@Pavel Orekhov You can change the log level while logged on as hdfs with the below steps, you don't need to restart the namenode Get the current log level $ hadoop daemonlog -getlevel {namenode_host}:50070 BlockStateChange Output Connecting to http://{namenode_host}:50070/logLevel?log=BlockStateChange
Submitted Log Name: BlockStateChange
Log Class: org.apache.commons.logging.impl.Log4JLogger
Effective Level: INFO Change to DEBUG $ hadoop daemonlog -setlevel {namenode_host}:50070 BlockStateChange DEBUG Output Connecting to http://{namenode_host}:50070/logLevel?log=BlockStateChange&level=DEBUG
Submitted Log Name: BlockStateChange
Log Class: org.apache.commons.logging.impl.Log4JLogger
Submitted Level: DEBUG
Setting Level to DEBUG ...
Effective Level: DEBUG Validate DEBUG mode $ hadoop daemonlog -getlevel {namenode_host}:50070 BlockStateChange Output Connecting to http://{namenode_host}:50070/logLevel?log=BlockStateChange
Submitted Log Name: BlockStateChange Log
Class: org.apache.commons.logging.impl.Log4JLogger
Effective Level: DEBUG You should be able to notice the logging level in namenode.log has been updated, without restarting the service. After finishing your diagnostics you can reset the logging level back to INFO Reset to INFO $ hadoop daemonlog -setlevel {namenode_host}:50070 BlockStateChange INFO Output Connecting to http://{namenode_host}:50070/logLevel?log=BlockStateChange&level=INFO
Submitted Log Name: BlockStateChange
Log Class: org.apache.commons.logging.impl.Log4JLogger
Submitted Level: INFO
Setting Level to INFO ...
Effective Level: INFO Validate INFO $ hadoop daemonlog -getlevel {namenode_host}:50070 BlockStateChange Output Connecting to http://{namenode_host}:50070/logLevel?log=BlockStateChange
Submitted Log Name: BlockStateChange
Log Class: org.apache.commons.logging.impl.Log4JLogger
Effective Level: INFO There you go If you found this answer addressed your question, please take a moment to log
in and click the "accept" link on the answer.
... View more
02-07-2019
06:36 PM
@Richard Wheeler if you left the sandbox idling then for sure it MUST be the logs generated in /var/logs/{component}/ the HDP components continually generate logs with the components statuses and on the sandbox, it's mount on / # du -a /var/log/ | sort -n -r | head -n 20 Sample output 3363560 /var/log/
1966344 /var/log/kafka
494300 /var/log/ambari-metrics-collector
267092 /var/log/hadoop
265560 /var/log/hadoop/hdfs
171528 /var/log/hadoop/hdfs/hadoop-hdfs-namenode-test.tarta.se.log
159432 /var/log/ambari-agent
98756 /var/log/ambari-infra-solr
81932 /var/log/ambari-metrics-collector/ambari-metrics-collector.log.3
81932 /var/log/ambari-metrics-collector/ambari-metrics-collector.log.2
81932 /var/log/ambari-metrics-collector/ambari-metrics-collector.log.1
81928 /var/log/ambari-metrics-collector/ambari-metrics-collector.log.4
81924 /var/log/ambari-metrics-collector/ambari-metrics-collector.log.5
69956 /var/log/oozie
49116 /var/log/hbase
40056 /var/log/ranger
39136 /var/log/ranger/admin
36420 /var/log/hive
36232 /var/log/hadoop-yarn
36176 /var/log/hbase/hbase-hbase-regionserver-test.tarta.se.log So you will need to delete the old files to regain some space. You can also run discretely a script in the cron !
... View more
02-07-2019
06:06 PM
1 Kudo
@Michael Bronson Apache Kafka uses Zookeeper to select a controller, Zookeeper tracks the status of Kafka cluster nodes and also plays a vital role for serving many other purposes, such as leader detection, configuration management, synchronization, detecting when a new node joins or leaves the cluster, etc.and maintain cluster membership by storing configuration, including the list of topics in the cluster. In order to remain part of the Kafka cluster, each broker has to send keep-alive to Zookeeper in regular intervals. This is something every Zookeeper client does by default. If the broker doesn't heartbeat Zookeeper every zookeeper.session.timeout.ms milliseconds (6000 by default), Zookeeper will assume the broker is dead. This will cause leader election for all partitions that had a leader on that broker. If this broker happened to be the controller, you will also see a new controller elected. In a Kafka cluster, service discovery helps the brokers find each other and know who’s in the cluster; and consensus helps the brokers elect a cluster controller, know what partitions exist, where they are, if they’re a leader of a partition or if they’re a follower and need to replicate, and so on. A controller is not too complex it is a normal broker that simply has an additional responsibility. That means it still leads partitions, has writes/reads going through it and replicates data. The most important part of that additional responsibility is keeping track of nodes in the cluster and appropriately handling nodes that leave, join or fail. This includes rebalancing partitions and assigning new partition leaders. There is always exactly one controller broker in a Kafka cluster. HTH
... View more
02-07-2019
03:42 PM
@Shraddha Singh Where machine is the FQDN and {rangerkms_password} is the rangerkms user password. The FQDN is the output of $hostname -f Re-run the below commands grant all privileges on rangerkms.* to 'rangerkms'@'machine' identified by '{rangerkms_password}';
grant all privileges on rangerkms.* to 'rangerkms'@'machine' with grant option; And let me know
... View more
02-07-2019
10:17 AM
@christophe VALMIR Any updates?
... View more
02-07-2019
08:05 AM
1 Kudo
@ram sriram The below command sets your replication factor to 1 for all new files you will create, with a potential data loss unless you are running HDP 3.x which has a new HDFS algorithm EC erasure coding $ hdfs dfs -setrep -w 1 -R / As responded above the changes only affect new files you will create. After changing the replication factor you won't see any hdfs size changes until the trash time interval which was set on 360 minutes configured by the hdfs trash interval has been reached fs.trash.interval Once the NameNode metadata has been updated, it is the DataNodes which would actually do the operation. There could be some delay, but space is definitely reclaimed. HTH
... View more
02-06-2019
01:03 PM
@Chris Jenkins My pleasure I made your day and welcome to Big data space, having to go all through all this will make you better technically you've now seen the different facets to resolving a problem. Happy Hadooping !
... View more