Member since
09-29-2015
5226
Posts
22
Kudos Received
34
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1396 | 07-13-2022 07:05 AM | |
3595 | 08-11-2021 05:29 AM | |
2330 | 07-07-2021 01:04 AM | |
1576 | 07-06-2021 03:24 AM | |
3552 | 06-07-2021 11:12 PM |
08-27-2020
02:05 AM
Hello @Mohsenhs , thank you for showing interest in the CCA159. Based on the description of the exam: "CCA159 is a hands-on, practical exam using Cloudera technologies. Each user is given their own CDH6 (currently 6.1.1) cluster pre-loaded with Impala, HiveServer1 and HiveServer2." You can download the required Cloudera product following the instructions from the documentation: "A 60-day trial can be enabled to provide access to the full set of Cloudera Enterprise Cloudera Enterprise features." Please let us know if it answers your inquiry! Thank you: Ferenc
... View more
08-26-2020
05:03 PM
Sentry-HDFS authorization is focused on Hive warehouse data - that is, any data that is part of a table in Hive or Impala. For HDFS-only control, you should looke at HDFS ACLs or Extended ACLs. See this doc.
... View more
08-26-2020
01:22 AM
Hello @KSKR , thank you for raising the question on "how to fetch the CPU utilization for a Spark job programmatically". One way to do this is via the Spark REST API. You should consider if you need the "live data" or you are looking for analysis once the application finished running. While the application is running, you can consider to connect to the driver and fetch the live data. Once the application finished running, you can consider parse the JSON files (the event log files) for the CPU time or use the Spark REST API and let the Spark History Server serve you with the data. What is your exact requirement? What would you like to achieve? Thank you: Ferenc
... View more
08-25-2020
07:24 AM
Hello @xinfengz , thank you for your interest in the installation of CM6.3.3. Based on the documentation under "Managing Licences", you have the below options: When you install Cloudera Manager, you can select among the following editions: Cloudera Express (no license required), a 60-day Cloudera Enterprise Cloudera Enterprise trial license, or Cloudera Enterprise (which requires a license). To obtain a Cloudera Enterprise license, fill in this form or call 866-843-7207. Please let us know if it addresses your inquiry. Thank you: Ferenc
... View more
08-25-2020
05:52 AM
Hello @Love-Nifi and @vchhipa , Thank you for posting your inquiry about timeouts. Without the full log, I can provide only some "if you see this, do that" kind of instructions. If you see an ERROR message with: org.apache.nifi.controller.UninheritableFlowException: Failed to connect node to cluster because local flow is different than cluster flow, then follow the below is the steps to resolve the issue: 1. Go to NIFi UI > Global Menu > Cluster 2. Check which host is the coordinator and login to that host on the shell. 3. Go to flow.xml.gz file location. [default location is /var/lib/nifi/conf/] 4. Copy flow.xml.gz on the disconnected node and replace the original flow.xml.gz with copied flow.xml.gz file. 5. Check permissions and ownership of newly copied flow.xml.gz file and then restart Nifi on the disconnected node only. If you are suspecting purely timeout issues, please attempt to tweak the below values in nifi.properties and restart the service: - nifi.cluster.node.protocol.threads=50 (Default 10) - nifi.cluster.node.connection.timeout=30 sec (Default 5 sec) - nifi.cluster.node.read.timeout=30 sec (Default 5 sec) Please find below a set of configurations that worth tuning on larger clusters based on https://nifi.apache.org/docs/nifi-docs/html/administration-guide.html The below are some example values for larger clusters (you need to tune it based on your unique setup): nifi.cluster.node.protocol.threads=70 nifi.cluster.node.protocol.max.threads=100 nifi.zookeeper.session.timeout=30 sec nifi.zookeeper.connect.timeout=30 sec nifi.cluster.node.connection.timeout=60 sec nifi.cluster.node.read.timeout=60 sec nifi.ui.autorefresh.interval=900 sec nifi.cluster.protocol.heartbeat.interval=20 sec nifi.components.status.repository.buffer.size=300 nifi.components.status.snapshot.frequency=5 mins nifi.cluster.node.protocol.max.threads=120 nifi.cluster.node.protocol.threads=80 nifi.cluster.node.read.timeout=90 sec nifi.cluster.node.connection.timeout=90 sec nifi.cluster.node.read.timeout=90 sec Please check if you notice any certificate related exception, like: WARN [Clustering Tasks Thread-2] o.apache.nifi.controller.FlowController Failed to send heartbeat due to: org.apache.nifi.cluster.protocol.ProtocolException: Failed marshalling 'HEARTBEAT' protocol message due to: javax.net.ssl.SSLHandshakeException: Received fatal alert: bad_certificate In this case, create a new keystore and truststore and add client auth in the keystore. Best regards: Ferenc
... View more
08-25-2020
01:33 AM
Hello @Bender , Thank you for your answer that is very clear. Even HDP last version, which is 3.1.5, uses Hive 3.1.0 (not 3.1.2). And Hive 3.1.2 should be a minor release to fix bugs from 3.1.0 version. If I really want to "force" upgrade from 3.1.0 to 3.1.2, do you have any advice to proceed ? or any upgrade procedure ? Thank you
... View more
08-06-2020
11:52 PM
Hello @ameya , thank you for your question about how to upgrade your cluster from HDP 2.6.5 to CDP 7.1.1. This quick start guide details the steps to follow. If you’re a Cloudera Subscription Support customer, we can connect you with your Account team to explore a possible Services engagement for this request. Let us know if you’re interested in this path, we’ll private message you to collect more information. Please let us know if you need any further assistance! Best regards: Ferenc
... View more
08-06-2020
01:45 AM
1 Kudo
Hello @emeric , the "kinit: KDC has no support for encryption type while getting initial credentials" usually occurs after configuring encryption types that do not match the ones present in the tgt principal (such as krbtgt/CLOUDERA@CLOUDERA) in the KDC. This can also happen while starting a service where the enctypes in the krbtgt principal do not match those used in service keytab. From an earlier Community post: Please compare the Kerberos server and client configurations and reconfigure krb5.conf on all your nodes to explicitly use the supported encryption type. The documentation describes: "Kerberos client OS-specific packages must be installed on all cluster hosts and client hosts that will authenticate using Kerberos." Wondering if some missing packages might be the issue? Kind regards: Ferenc
... View more
08-05-2020
11:02 PM
Thank you @Bender
... View more
08-05-2020
06:09 AM
2 Kudos
Hi SAMSAL, I believe we're hitting a similar issue here to this stackoverflow thread. As in, the third party utility that creates the csv files adds this non-breaking space character to the file that displays as a red dot. You mentioned that it's not trivial to change the input files but maybe this is a minor adjustment that the vendor can make.
... View more