Member since
06-26-2015
515
Posts
137
Kudos Received
114
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2022 | 09-20-2022 03:33 PM | |
5602 | 09-19-2022 04:47 PM | |
3037 | 09-11-2022 05:01 PM | |
3354 | 09-06-2022 02:23 PM | |
5297 | 09-06-2022 04:30 AM |
02-07-2022
07:57 PM
@an_dutra My guess is that it's a misconfiguration on your cluster. I just tested this on my Kafka cluster and once the certificate expires, if I try to connect to the cluster with a Kafka client I get the following exception: Caused by: sun.security.validator.ValidatorException: PKIX path validation failed: ... Caused by: java.security.cert.CertPathValidatorException: validity check failed ... Caused by: java.security.cert.CertificateExpiredException: NotAfter: Tue Feb 08 03:45:00 UTC 2022 The Kafka brokers will continue to run, though. However, if they are stopped and I try to start them again, they will fail to start with the same exception as the one above.
... View more
02-07-2022
02:36 PM
What @DigitalPlumber said. Also make sure that Match Requirement is set to "content must contain match", otherwise it won't work.
... View more
02-06-2022
04:18 PM
You can try the following Jolt spec in a JoltTransformRecord processor to change the name of one or more columns: Jolt Transformation DSL: Chain Jolt Specification: [ { "operation": "shift", "spec": { "*": "&", "surname": "lastname" } }, { "operation": "remove", "spec": { "surname": "" } } ]
... View more
02-02-2022
06:35 PM
That's an odd thing to do 🙂 If you need to use another port for some reason, it would be better to change the ports on all hosts consistently, using the "TCP Port" or the "TLS/SSL Port" properties in Cloudera Manager, whether are you connecting without TLS or with it, respectively. It is possible to configure ports on a host-by-host basis, but it makes it harder to maintain and client configuration becomes a little more cumbersome. To change the port for a particular host, go to Kafka > Instances > Click on the broker your want the change the port for > Configuration > Continue Editing Role Instance. Then enter the following in the "Kafka Broker Advanced Configuration Snippet (Safety Valve) for kafka.properties" property: port=9096 listeners=PLAINTEXT://:9096 The PLAINTEXT value will depend on your cluster config: PLAINTEXT: No Kerberos, No TLS SSL: No Kerberos, Using TLS SASL_PLAINTEXT: Using Kerberos, No TLS SASL_SSL: Using Kerberos, Using TLS After that restart the Broker instance that was reconfigured.
... View more
02-01-2022
08:45 PM
Some more examples here: https://github.com/asdaraujo/cdp-examples
... View more
08-17-2019
04:05 PM
1 Kudo
This was reported as a bug, and has already been fixed in CM 6.3.0, 6.2.1 as part of OPSAPS-49111
... View more
08-17-2019
07:46 AM
Hi @araujo You are correct. Just added the jars from "CSD & Manifest Files" from https://docs.hortonworks.com/HDPDocuments/CFM/CFM-1.0.0/release-notes/content/download-locations.html Now I can install Nifi. Thanks.
... View more
08-15-2019
12:23 PM
OK - so it seems that the yum repo for cloudera manager was defined / pinned to 6.2 in the cluster definition - so of course Director did what it was being asked to do. Updated to point to 6.3. repository: "https://archive.cloudera.com/cm6/6.3/redhat7/yum/" repositoryKeyUrl: "https://archive.cloudera.com/cm6/6.3/redhat7/yum/RPM-GPG-KEY-cloudera"
... View more
08-14-2019
06:46 PM
1 Kudo
Hi Andre, Your solution is right. But my situation was little different. Below are the checks and fix I did with cloudera support helping me in the process: 1. From Hive-server2 logs we found that one of the Hiveserver2 instance is not talking to zookeeper quorum(only in case of querying Hbase data) 2. Installed Hbase-gateway services on all the Hue instances and Hiveserver2 instances. 3. restart Hbase services and Deploy client configuration. 4. Restart the Hiveserver2 instance which had the problem of trying to connect to localhost:2181 as zookeeper quorum Then tried to submit the query from beeline and Hue . All worked as expected this time.
... View more
08-14-2019
05:27 PM
For number 2, ANY changes outside of Impala, you will need INVALIDATE METADATA, or if new data added, then REFRESH will do. Work is underway to improve it: https://issues.apache.org/jira/browse/IMPALA-3124 Cheers Eric
... View more
- « Previous
- Next »