Member since
05-29-2017
43
Posts
7
Kudos Received
6
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3873 | 10-02-2017 08:40 AM | |
2016 | 10-02-2017 08:13 AM | |
3444 | 09-20-2017 09:24 AM | |
10265 | 09-18-2017 06:35 AM | |
18081 | 09-06-2017 04:52 AM |
08-05-2018
02:35 AM
Hello, How are you able to resolve the firewall issue, can you tell me the steps you followed to make the port reachable.Can you please elaborate. Thanks in advance.
... View more
02-22-2018
11:51 AM
1 Kudo
@resolution, What version of Cloudera Manager are you using? In recent releases, it should not be possible to have this sort of thing happen unless your database is indeed without any information. Go to the following directory: /etc/cloudera-scm-server Next, look at the db.properties file. Make sure it is configured properly with your correct database host, port, etc. If not, you can also look at other db.properties files in the directory to see if any of them may contain the correct information. If you are unsure, feel free to share the information here (redcacting your password if necessary)
... View more
10-29-2017
05:22 AM
There is a way to deploy R Server 9.1 on a Cloudera cluster using parcels and run it as a service from Cloudera Manager. I have tested recently and the following guide from Microsoft explains the required steps: https://docs.microsoft.com/en-us/machine-learning-server/install/r-server-install-cloudera I have recently performend an installation or R Server and R Studio on Cloudera Manager & CDH 5.13, following the guides. The instructions are quite good, there are some gotchas though: 1) You can download the parcel from the sources listed under https://docs.microsoft.com/en-us/machine-learning-server/install/r-server-install-cloudera-generate-parcel If you just want to test the setup the easiest way is to download R Studio from Visual Studio Dev Essentials. Just make sure you download the correct file for Hadoop (not for Linux): Microsoft R Server 9.1.0 for Hadoop (x64) - (English) No additional details are available for this file. Released: 4/19/2017 SHA1: F6A03F858BB6733C8B074458CF18C102B64DD3A8 File name: en_microsoft_r_server_910_for_hadoop_x64_10323951.tar.gz 2) You can unpack the files and run the generate_mrs_parcel.sh script to generate the parcel, parchel SHA and the CSD file: MRS-9.1.0-[..].parcel (file name depends on select OS version) MRS-9.1.0-[..].parcel.sha MRS-9.1.0-CONFIG.JAR 3) You may also want to download Microsoft R Open for Microsoft R Server. Fortunately, R Open is also available as a Cloudera parcel: https://mran.revolutionanalytics.com/download/mro-for-mrs/ Select the following: R Open for Microsoft R Server – Hadoop Cloudera Manager Parcel sha256: 317f6b455411303363d6694e7b9c055bffef010d5681e52054be31e32b57b6f9 The archive contains both a parcel and parcel sha file. Both should go under /opt/cloudera/parcel-repo/ as well I was installing on CentOS 6.8 and I ran into an issue where the file names for the generated parcel were not correct (MRO-3.3.2-_dist_.parcel). If you have something like this, make sure to rename the parcel and parcel sha files according to your RHEL/CentOS version: MRO-3.3.2-el6.parcel MRO-3.3.2-el6.parcel.sha Preferably do this in the folder where the files were generated and only move to /opt/cloudera/parcel-repo/ AFTER you have renamed them. (Otherwise Cloudera Manager may not find the new parcels and an additional restart of the cloudera-scm-server service on the CM host may be needed.) Conclusion If you have done everything correctly, clicking Check for New Parcels in Cloudera Manager (top right corner -> Parcels page) will refresh the list of available parcels for MRO and MRS should show up. If they do not show, check the exact filename (should represent the OS version), their location (/opt/cloudera/parcel-repo/), owner (cloudera-scm: cloudera-scm) and permissions (644). You may also need to restart the cloudera-scm-server service. Installing R Studio Server As mentioned earlier, there is no parcel available for R Studio Server, this needs to be managed manually. There is a short guide on how to get it up and running on Cloudera https://support.rstudio.com/hc/en-us/articles/235841328-Using-RStudio-Server-with-Microsoft-R-Server-for-Cloudera I also needed some additional steps, which are very well documented in a Microsoft blog post: https://blogs.msdn.microsoft.com/microsoftrservertigerteam/2017/04/30/using-rstudio-server-with-microsoft-r-server-parcel-for-cloudera/
... View more
10-03-2017
04:00 PM
I would like to submit my dissertation in CD-ROM or pendrive , so how can I export cloudera data into cd rom or pendrive ??
... View more
10-03-2017
08:57 AM
Hello prashanthyadav, I'm not sure which Cloudera software would you like to install on top of Windows Server, but in case of Cloud, it's best to run Linux VMs directly. As a Cloud VM is already running in a virtualized container, another layer of virtualization on top of that is usually not supported. If you're interested in Cloudera Enterprise on Microsoft Azure, you may find useful our documentation about reference architecture of Cloudera Enterprise for Azure Zsolt
... View more
10-02-2017
10:55 AM
Dear zherczeg Thank you so much for your kind help! I believe your method will work perfectly. I will implement it tomorrow and report to you the result within this week due to that the admin team might want to observe several days. Your sincerely, Chao-Kang Chen
... View more
10-02-2017
08:40 AM
Hello Awisawe, In order for Hue to work over HTTPS, it needs a TLS certificate and a corresponding private key. These are configured via ssl_certificate, ssl_private_key, and ssl_password variables. We've documentation about setting up Hue as TLS/SSL server for Cloudera Enterprise, which you may be able to adapt to your distribution. Best regards, Zsolt
... View more
09-21-2017
07:41 AM
Thanks for your reply. I understand the issue you mentioned about HBase not knowing or caring about the Avro content. What I'm trying to learn is how to create Avro objects without having to use the Avro write/read functions. Basically I want to: * Load content into variables (var1 = schema + data) * HBase put table, var1, var2, etc. Thanks again
... View more
09-20-2017
02:28 PM
Hi venkatachetkuri, In addition, the "LDAP User Filter" needs to be in the form of: (sAMAccountName={0}) . Substitue the "sAMAccountName" part with the user attribute that you use for usernames. "cn" or common name is usually the first name + last name for individual users and typically is not used for usernames.
... View more
09-20-2017
08:30 AM
Hi Rabk, I think you're asking more information about the step 8. in our enabling TLS documentation, which recommends turning on authentication if you set up TLS. In a CDH cluster, authentication is usually provided by kerberos, whereas encryption on the wire is provided by TLS. By turning on " Enable Kerberos Authentication for HTTP Web-Consoles " the webservices will require SPNEGO authentication, with a trusted kerberos principal. We've a documentation on Enable Kerberos Authentication for HTTP Web-Consoles as a part of our kerberos setup guide, and on setting up browsers for SPNEGO authentication. You can also find information about that authentication in the Apache documentation. Do you have any specific questions about that? Zsolt
... View more
09-19-2017
06:11 AM
Hello Zsolt, thanks for the reply. The problem was, that I don't have the permissions to install python packages like pydoop. I was not sure if there is a native way, but I will ask the sysadmin to install some packages.
... View more
09-18-2017
05:55 AM
Hello asherma5, You can configure Hue with an external load-balancer, that way it'll keep the browser on the load-balancer and does not redirect it to itself: http://gethue.com/automatic-high-availability-with-hue-and-cloudera-manager/ This is only partial solution, as it'll not solve your issue with moving Hue to a subfolder. Hue is based on django, which handles the routing for it. I'm not aware of a way to reliably relocate Django apps to a subfolder without editing the source-code. Zsolt
... View more
09-08-2017
01:01 AM
Hi Nadi1998, The download website http://go.cloudera.com/ced has also been updated, and it's working for me now. Can you re-try it?
... View more
09-06-2017
08:37 AM
Hi SumeetRaheja, Wildcard certificates are only working from CDH 5.10, which version do you plan to install? A valid certificate is required for all nodes. In case of a wildcard certificate, it may be used on multiple nodes where it matches the hostname. If all of the DNs FQDN is matching the wildcard certificate, and your cluster version is at least 5.10.0, than I don't see any problems with your setup. Zsolt
... View more
09-06-2017
07:45 AM
Hi Vamp, Unfortunately at the current version only Cloudera Navigator has a configuration to log events to syslog. You may find the Cloudera Manager API as a useful alternative, where all the Cloudera Manager Events are exposed: http://cloudera.github.io/cm_api/apidocs/v17/path__events.html
... View more
08-31-2017
07:37 AM
Hi, Hue is connecting directly to the mysql server that this configuration provides. Which machine do you use to login mysql? Is it the same node where Hue is installed?
... View more
08-28-2017
08:21 AM
1 Kudo
Hello, At the moment, only Amazon S3 or another Cloudera cluster is supported as a BDR replication target, but we have support for Microsoft Azure Data Lake Store as HDFS backend for Cloudera clusters. For Hive table replication databases or whole tables can be selected. It is not possible to set up replication to only a single partition of a hive table.
... View more
08-28-2017
07:37 AM
Hello, Receiving "Results Expired" error in Hue for Impala queries, is usually a symptom of having a misconfigured load-balancer in front of Impala. Impala has the assumption, that the query results will be downloaded from the same impalad where the query was issued. A load-balancer can break that assumption, if it sendis sequential connections of the same client to different impalad daemons. If Hue connects two a different impalads to run the query and download the results, the "Results Expired" error will appear. To avoid the above issue, it's required that the proxy routes subsequent connections of the same client to the same impalad. This proxy behavior is usually called sticky sessions. Our documentation has detailed information about impala load-balancer setup, includeing the above information as well: https://www.cloudera.com/documentation/enterprise/latest/topics/impala_proxy.html#proxy_overview
... View more
06-07-2017
07:41 AM
thanks, it was the http:// in the proxy server config
... View more
06-07-2017
12:38 AM
Hi autumnSea! Cloudera Manager use SSH to install the agent on new hosts. This error message indicates an SSH timeout. Are you able to SSH to the new hosts? Could you check on the new hosts logs that Cloudera Manager authenticates successfully via ssh or not? If everything seems OK, please share more information about your environment, such as your Cloudera Manager version, OS type, maybe SSH configuration of the new hosts if possible.
... View more