Member since
09-02-2016
523
Posts
89
Kudos Received
42
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2724 | 08-28-2018 02:00 AM | |
| 2697 | 07-31-2018 06:55 AM | |
| 5688 | 07-26-2018 03:02 AM | |
| 2988 | 07-19-2018 02:30 AM | |
| 6466 | 05-21-2018 03:42 AM |
03-02-2017
04:39 PM
@Akira191 1. Go to Cloudera Manager -> Spark -> Instance -> Identify the node where you have Spark server installed 2. Login to the above identified node using CLI and go to path "/opt/cloudera/parcels/CDH-<version>/lib/spark/bin" , it will list binaries for "spark-shell, pyspark, spark-submit, etc". It helps us to login to spark & submit jobs. if it has spark-sql, then you can run the command that you have mentioned. In your case, spark-sql binary should be missing, so you are getting this error. You need to talk to your admin
... View more
02-28-2017
09:10 PM
@matt123 Go to link http://ipaddress:8088 and check the Cluster Metrics for the RAM, Container, vcore usage Also Click on "active nodes" to see the same information by node Cloudera Manager -> HDFS -> Web UI -> Namenode UI -> See the complete metrics
... View more
02-28-2017
08:46 PM
@codenchips Go to CM -> Hosts -> Click on each hosts -> Health history (left down) -> share me the details
... View more
02-28-2017
07:11 AM
@codenchips Go to Cloudera Manager -> Hosts , check the Host status and understand what kind of issue it shows Also login as root in linux and run the below command service ntpd status service ntpd start service ntpd status restart the CM and try again
... View more
02-27-2017
09:24 AM
upon further analysis, i've noticed that "navigator policies" might help on this https://www.cloudera.com/documentation/enterprise/5-5-x/topics/navigator_policies.html It seems that I need to write search query, let me try to write one... In the mean time, it will be great if some share the query for the above scenario...
... View more
02-27-2017
08:38 AM
Hi Does Cloudera navigator has an option to identify Unused objects for a particular period (like more than 6 months, 1 year, etc)? The object can be HDFS files, Hive/Impala tables/Oozie, dataset, etc This is my requirement: Our non-prod environment has been used by multiple users for different reasons like dev, test, etc. Sometimes they use common user id & user space to create db, create/import tables, etc. After the task finished, they will move to the next task without cleaning the old DB, tables, files which become garbage after few days. It has been accumulated and become a big garbage now (with 3 replication). I want to identify the DB, tables, files which are not in use for more than 6 months (or) 1 year and delete them (with proper approval...) Is it possible with Navigator? is there any other option/ideas? Thanks Kumar
... View more
Labels:
- Labels:
-
Cloudera Navigator
02-26-2017
06:31 PM
1 Kudo
@wenjie Pls check the telnet response 1. Cloudera Manager -> Impala -> Instance -> Impala Catalog Server -> Get the hostname 2. Try the below command in Linux $ telnet <hostname> 25020
... View more
02-24-2017
10:48 AM
@Rashmi22 As an alternate, you can also use "hdfs version" or "hadoop version" command in CLI to get CDH version
... View more
02-23-2017
08:57 AM
@RakeshE You can use Access Control List (ACL) to protect your file in HDFS. Pls refer the below links https://community.cloudera.com/t5/Security-Apache-Sentry/Hadoop-Security-for-beginners/m-p/48576#M174 https://hortonworks.com/blog/hdfs-acls-fine-grained-permissions-hdfs-files-hadoop/
... View more
02-21-2017
02:16 PM
@MasterOfPuppets Very Hypothetical "one line" question. I don't think just adding few extra nodes will double the performance...Few of the additional parameters that you need to consider as 1. The way services are configured in the cluster is also very important. Ex: You have 3 nodes now, Consider 10 services are configured in 3 nodes. After 3 more nodes are added, you need to properly distribute the services to the new nodes as well On Existing Cluster - without adding new nodes: 1. If possible, Add RAM to existing nodes 2. Identify which particular services required better performance like hive, impala, etc. You can tune the environment configuration for those services. Ex: Increase Java heap size, etc 3. Prioritize the jobs etc
... View more