Member since
02-27-2020
157
Posts
38
Kudos Received
43
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
190 | 05-20-2022 09:46 AM | |
106 | 05-17-2022 08:42 PM | |
169 | 05-06-2022 06:50 AM | |
184 | 04-18-2022 07:53 AM | |
145 | 04-12-2022 11:17 AM |
05-20-2022
09:46 AM
Looks like there are different types of AWS Load Balancers, and the one that can handle TCP sessions are either Network Load Balancer (NLB) or Classic Load Balancer. Which one are you using? I believe stickiness also works for NLB, but do validate with your AWS team. The session timeout is two-fold: LB session timeout. These are settings inside of the load balancer, nothing to do with Impala itself. The recommendation is to set this timeout to 12 hours, 6 hours at a minimum for both client and server timeouts. Hue session timeout (idle_session_timeout). This is how long Hue will keep the connection to Impala alive. Default may be 15 minutes. If no queries run in this time amount of time, and then you run a new query, you'll get this error "Results have expired". Hue will need to start a new session. Also if your query takes longer than idle_session_timeout, then you'll definitely need to increase the timeout setting. 1 hour timeout may be appropriate here. Regards, Alex
... View more
05-20-2022
08:33 AM
That's right, you'll need to look into how to setup AWS LB to have sticky sessions, which is not enabled by default. Also, please look into tuning the session timeout to a high value (long enough for Impala to finish the query and send the results back to Hue) Reference: https://docs.aws.amazon.com/elasticloadbalancing/latest/application/sticky-sessions.html Kind regards, Alex Akulov
... View more
05-19-2022
09:34 PM
1 Kudo
Hi @yagoaparecidoti , If you have Impala daemons that are load-balanced (LB), then there is a configuration that needs to be applied on the LB, related to source IP persistance. There is a Knowledge Base article about this: https://community.cloudera.com/t5/Customer/Load-balancing-between-Hue-and-Impala/ta-p/73364 If there is no LB at play here, then has this issue been ongoing or just started to pop up recently? What changed on the cluster recently? Regards, Alex
... View more
05-17-2022
08:42 PM
Hi Andrea, There is not much different from how you would install R and SparkR on any other host. The steps to install would be the same. How that R runtime integrates with CDP DataHub would need to be configured after the installation. Instead of SparkR, I would actually recommend you use sparklyr, as it's more compatible with Spark. Regards, Alex
... View more
05-10-2022
12:45 PM
1 Kudo
This likely needs attention from Cloudera support. Please open a case through your mycloudera portal. Regards, Alex
... View more
05-10-2022
12:05 PM
Hi @mmk , please provide what version of CDP you are using, on-prem or public cloud. Also see this documentation page for working with S3: https://docs.cloudera.com/cdp-private-cloud-base/7.1.7/cloud-data-access/topics/cr-cda-configuring-access-to-s3.html Kind regards, Alex
... View more
05-06-2022
06:50 AM
Ok, @aval, if you are after the CDP Base requirements that are needed to install CDP Data Services (formerly experiences), then check the list at the bottom of this page: https://docs.cloudera.com/cdp-private-cloud-data-services/1.3.4/installation-ecs/topics/cdppvc-installation-base-software-requirements.html For all dependencies between various services, you can consult this table: https://docs.cloudera.com/cdp-private-cloud-base/7.1.7/installation/topics/cdpdc-service-dependencies.html Hope this helps, Alex
... View more
05-05-2022
10:10 PM
CDP 7.x uses parcels through Cloudera Manager. RPMs are no longer provided. Regards, Alex
... View more
05-05-2022
10:03 PM
Hi @aval , These are all great questions. You mentioned that these are licensed installations, so you may want to work with the account team from Cloudera to answer these questions. There is a lot of detail that would not be tractable to cover in a forum post. If you are not sure who to talk to, open a support case with Cloudera and ask to be connected with the account team for your enterprise. Best wishes, Alex
... View more
04-19-2022
01:03 PM
Happy to hear, @Data1701 . Accept reply as Solution if the answer helped.
... View more
04-18-2022
08:01 AM
Hi @Jaguare , What you likely want to use is not sqoop, but hdfs' native distcp command. There are guides on the internet on how to do this. A quick google found this: https://docs.microsoft.com/en-us/azure/hdinsight/hadoop/apache-hadoop-on-premises-migration-best-practices-data-migration Hope this helps. Regards, Alex
... View more
04-18-2022
07:53 AM
1 Kudo
Hello @Data1701 , The metrics you are looking for would be under the Hosts view in Cloudera Manager. On the left-hand panel in Cloudera Manager click on Hosts > All Hosts. This will list all of the hosts in the cluster, along with their physical memory and cores for each. Note that, by default, not all of the necessary columns may be visible in the table. To add columns of interest (e.g. cores), in the top right corner next to "Columns:" click Selected dropdown menu and choose the columns you need to see. Hope this helps. Regards, Alex
... View more
04-12-2022
03:24 PM
@THR-Mario , if you want to use Grafana as a general-purpose data visualization tool, then unfortunately that's not its purpose. Grafana is meant for time-series visualization, and in the context of CDP is used for metrics monitoring. In this task Grafana is great! If you want to use Grafana to query Hive, Impala, or other large datasets, you will have a bad time. Sorry, that's not what you want to hear, but that's the reality. Regards, Alex
... View more
04-12-2022
11:17 AM
1 Kudo
Hi @Ragavend, Happy to hear that you are exploring CDP Private Cloud and taking on the learning of the platform in a lab environment. To answer your questions: 1. There are a few steps that are needed to access CDP Private Cloud CLI. Instructions are here: https://docs.cloudera.com/management-console/1.3.3/private-cloud-cli/topics/mc-private-cloud-cli-client-setup.html. Note that you will need to allow external connections to your AWS EC2 instances in order to be able to issue commands from your laptop to the CDP cluster. This is also assuming you are talking about CDP CLI. If you are talking about AWS CLI (different tool entirely), then please see the many AWS tutorials available. 2. In order to run pyspark, the user who is executing the the job needs to be able to create a log directory on hdfs. So, instead of running your command as root (i.e. centos) try running it as your CDP admin user. Hope this helps. Regards, Alex
... View more
04-06-2022
01:55 PM
2 Kudos
Hi @THR-Mario , The built in Grafana dashboards that come with CDP expose operational metrics that are relevant to cluster administration. If you are looking to use visualization tools to view the data in, for example, Hive tables, you'll need to look at Cloudera Data Visualization which is a component of Data Warehousing in public cloud or CML/CDSW on-premise. Alternatively, CDP Hive is compatible with popular 3rd party data visualization tools. Hope this helps, Alex
... View more
04-04-2022
09:35 AM
Hi @modususer , This may indicate a problem with your license file (e.g. corruption or expired dates). The easiest check is to open the file and check the license expiry date. It will be in plain text. If that is good, there still may be a problem with the hash, if it got corrupted. Does the same exact license file work in any other environment in your organization? Also, is this a trial or production installation of CDP Base you are trying to do? Regards, Alex
... View more
03-21-2022
02:02 PM
1 Kudo
If you are talking about the little icons that show up next to the table names in the table list on the left-hand side of your Hue editor, then yes, "table" icon is a proper physical table, "eye" icon is a view defined on top of some statement. Hope this helps, Alex
... View more
03-21-2022
07:53 AM
2 Kudos
Hi @rahuledavalath , Generally, having internet access on the nodes where you are installing CDP Private Cloud simply saves you a few manual steps. However, it is absolutely acceptable to have nodes without internet and do an installation on "air-gapped" environment (e.g. see here: https://docs.cloudera.com/cdp-private-cloud-experiences/1.3.3/installation/topics/cdppvc-installation-airgap.html) As for license activation, there is no need for internet access either. The license you get from Cloudera is a text file. As long as you upload it to Cloudera Manager in the appropriate installation step, there is no internet validation that needs to happen. Everything is self-contained on your hosts. Hope this helps, Alex
... View more
02-16-2022
09:59 PM
Hi @banshidhar_saho , Unfortunately, there is no configuration in CM that will let you change the header of the alert email or the format of the alert email. However, there is a customization you can do by developing a custom script that (a) takes in a JSON string of alerts; and (b) builds and email object and sends it off to your SMTP server. It takes some effort to write the script, but gives you all the flexibility. See more details here: https://docs.cloudera.com/cdp-private-cloud-base/7.1.3/monitoring-and-diagnostics/topics/cm-alerts-script.html If this is helpful, please accept this reply as a solution. Regards, Alex
... View more
02-15-2022
09:17 AM
Hi @Juanes, The command you shared should work, given that kudu-tserver process is actually running on the host. Can you list all processes and grep for "kudu" on that host? For more information see this documentation page on starting kudu on the host. If this helps, please accept as a solution. Regards, Alex
... View more
02-13-2022
03:37 PM
1 Kudo
When you go to this download page and login, just before you can download the file you will be presented with a click-through license agreement. The text of the agreement can be copied and studied separately. Hope this helps, Alex
... View more
02-13-2022
03:32 PM
1 Kudo
Hi @Melon , Sounds like you are wanting to send an email alert (e.g. when a CDSW job finishes) via your corporate SFTP server. Is that correct? For configuring CDSW SMTP connection you will need to have at least CDSW Admin role and go to Admin > Settings (tab). Here under SMTP you will need to provide the following information: SMTP host address SMTP port SMTP username SMTP password Optionally, you may check "Use TLS" if your SMTP server requires this. What you are talking about in terms of VPN (Citrix Gateway) is something you may want to check with your IT organization. You should be able to upload and download project files in CDSW via the browser, when you are connected through VPN. Hope this help, Alex
... View more
02-10-2022
04:42 PM
2 Kudos
HiveAccessControlException suggests you are accessing this s3 location through a SQL engine (Hive or Impala perhaps). Check in Ranger, under Hadoop SQL, if the policies are set properly there to access the table you are looking at. Also, is this a RAZ-enabled environment, by any chance? If it is, please see here for RAZ setup specific to Hive table access: https://docs.cloudera.com/management-console/cloud/fine-grained-access-control-aws/topics/raz-aws-create-ranger-policies.html Hope this helps, Alex
... View more
01-19-2022
11:28 AM
Hi @Chhavi , the discussion of HIPPA certification is an intricate one. I would recommend to talk to your Cloudera account representative as there are many details to go through. Thank you, Alex
... View more
01-13-2022
08:21 AM
@Chhavi , if you are asking about "Apache NiFi", then the answer is Appache NiFi is not and never will be HIPPA certified or certified for anything similar. This is because it's an Apache project and nobody owns certification for it as such. With that said, it is likely possible to configure NiFi in a HIPPA compliant way (e.g. repository encryption). You'll need to work with your NiFi vendor for that and Cloudera is happy to help. Regards, Alex Akulov
... View more
01-13-2022
07:36 AM
Hi @grlzz , CDSW is typically installed/runs on edge nodes, so CM going down does not necessarily preclude you from accessing CDSW. However, if certs expired across the cluster then all services may be inaccessible. Regarding your question for command line access to CDSW. There is a CLI you can use for that: https://docs.cloudera.com/cdsw/1.9.2/cli-reference/topics/cdsw-cli-reference.html The first thing you'll need to do to use the CLI is to generate some access keys from CDSW UI, so if you can't access CDSW UI, then you may be stuck until that is resolved. Have you opened a support case for this? Kind regards, Alex Akulov
... View more
10-21-2021
11:00 AM
Hi Alexis, for CDP Experiences (aka Data Services) there is integration available. Please see the link below: https://docs.cloudera.com/cdp-private-cloud-experiences/1.1/installation/topics/cdppvc-installation-external-vault.html For CDP Base, I don't believe we support external vault. Hope this helps, Alex
... View more
08-10-2021
12:00 PM
Could you provide the output of the DESCRIBE command for your table in Kudu.
... View more
07-13-2021
10:42 AM
Hi @Anyy , Your SQL statement seems to be malformed. You have "from" keyword duplicated. Please fix and try running again. If that helps, please accept as solution. Regards, Alex
... View more
07-08-2021
03:03 PM
1 Kudo
Strange.... As a test can you replace your hard-coded toppic name with something like ${tableName} or just try a different hardcoded string and see if that gets you further along. Not saying that's a solution, but trying to eliminate causes. Also, you mentioned you've implemented other flows already. Could it be that the topic name BUBB-MX_ALL_SOFT is already taken in your kafka cluster and thats why goldengate handler can't create the topic (though I would expect it to just write to the existing topic, instead of throwing an error). I also found this on Oracle site that may have your answer. You'll need an Oracle account to look if there is a solution though. https://support.oracle.com/knowledge/Middleware/2512462_1.html Could this also be an encoding or trailing character issue in the props file between Unix and AIX? Overall this sounds like a GoldenGate error that may be better answered by Oracle community. Regards, Alex
... View more