Member since
04-22-2014
1218
Posts
341
Kudos Received
157
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 26248 | 03-03-2020 08:12 AM | |
| 16395 | 02-28-2020 10:43 AM | |
| 4716 | 12-16-2019 12:59 PM | |
| 4472 | 11-12-2019 03:28 PM | |
| 6657 | 11-01-2019 09:01 AM |
01-01-2019
06:49 PM
Hi, tried this and it still looks for auto-tls setting. I note that this auto-tls feature cant be turn off as after saving new setting in CM security section, and restarting cm server. It will still revert to original setting which has auto-tls enabled. As such, ive decided to use CDH5 & CM5 instead. Thanks for assistance.
... View more
12-19-2018
03:59 PM
Hi Ben, But if we cannot find that file, what should we do? Why these files are missing? Thanks, Mo
... View more
12-19-2018
02:23 PM
@bgooley Thanks a bunch! This is good info. I do see the below now which means /usr/lib/jvm is good for openJDK. Note: Cloudera strongly recommends installing Oracle JDK at /usr/java/<jdk-version> and OpenJDK at /usr/lib/jvm (or /usr/lib64/jvm on SLES 12), which allows Cloudera Manager to auto-detect and use the correct JDK version. Unfortunately in the CDH 5.16 install guide it doesnt clarify that for openJDK /usr/lib/jvm is good path but makes a blanket statement that The JDK must be installed at /usr/java/jdk-version. Hopefully they will update the doc in future. https://www.cloudera.com/documentation/enterprise/5-16-x/topics/cdh_ig_jdk_installation.html .
... View more
12-15-2018
11:57 PM
So the problem was with Snapshots. I had configured snapshots a long time ago on the /user/hive/warehouse directory, and they were still being generated. I was finding the space using the commands hadoop fs -du -h /user/hive hadoop fs -du -h /user/hive/warehouse Snapshot directories can be found using command: hdfs lsSnapshottabledir hadoop fs -delteSnapshot <path without .snapshot> <snapshotname>
... View more
11-18-2018
05:38 AM
Thank you @bgooley Thats what i did and it worked. thank you and appreciate your response.
... View more
11-16-2018
09:04 AM
@bgooley @Tomas79 I ended up loging in a user when the EMR is launching in the bootstrap action.I did this via curl commands.This will avoid any user being given super user status. For anyone needing guidance on the workaround you could follow the below steps. 1) Curl command to get the cookie.txt(it has the session id and csrf token) file. 2) Curl command to login(You have to grep from the cookie.txt file the session id and csrftoken) If anyone has a better idea please let me know
... View more
11-15-2018
05:59 AM
Hi , I want to install cdh5.15.1 hadoop client on my Ubuntu 16.04 host. I have added the repo url in /etc/apt/sources.list.d/cloudera.list file. How can I specify a particular hadoop-client version in apt-get command. I thought apt-get install hadoop-client=< version > should work. But I am unable to figure out the version number. Here is the link for repo http://archive.cloudera.com/cdh5/ubuntu/xenial/amd64/cdh/pool/contrib/h/hadoop/ Thanks 🙂
... View more
11-14-2018
05:39 PM
2 Kudos
@orak, OpenLDAP is just fine for hadoop LDAP purposes. Active Directory is part of many existing IT infrastructures, so it is often used due to the way it does combine LDAP and Kerberos (along with other things). Users in your Kerberos KDC and LDAP server do not necessarily need to originate in the same object. Any true relationship between the two where the KDC principal exists in an end user object that is used for authentication would exist due to some sort of integration at the KDC / LDAP server level. This is not necessary for hadoop services to work. In general, there are 3 needs if you are going to secure your cluster with Kerberos: - Kerberos - means of mapping users to groups (usually OS shell-based, but can be LDAP based) - OS users as which services will run and end user OS users for YARN containers (running MR jobs) If I kinit as bgooley@EXAMPLE.COM and then attempt to perform a listing on a directory that is read for user/group and owned by someone else, then the NameNode must be able to determine if the user is a member of the group who has permission to list files. The principal would be trimmed to a "short name" by trimming off the realm to arrive at bgooley. The user bgooley's group membership would then be determined (shell group mapping or ldap group mapping) . See the following for details: https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/GroupsMapping.html This mapping is used by several services so it is part of core hadoop. Then, you have the OS users that must exist at the OS level so that various processes can start as those users and files be owned. Also YARN containers will store information in the OS file system as the user running the job. This means that users who run jobs need to exist on all nodes in the cluster. Some of these topics are covered in a bit more detail here: https://www.cloudera.com/documentation/enterprise/latest/topics/sg_auth_overview.html That's a lot to process, so I'll stop there and wait to see if you have any questions.
... View more
11-09-2018
09:39 AM
@bgooley Thanks for the explanation and help. Intially I have tried with self signed and now I have received signed certificates. Fixed the issue after creating a trustore.
... View more