Member since
02-28-2022
147
Posts
13
Kudos Received
14
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
151 | 10-18-2024 12:29 PM | |
2929 | 09-05-2024 09:06 AM | |
2457 | 08-08-2024 05:10 AM | |
1919 | 05-16-2024 05:33 AM | |
650 | 04-22-2024 10:24 AM |
01-23-2024
10:50 AM
hi cloudera, I need to use Spark on a host that is not part of the Cloudera cluster to run Spark jobs on the Cloudera cluster. Is it possible to use it this way? If yes, how to configure? what I've already tried: 1. Download "https://www.apache.org/dyn/closer.lua/spark/spark-3.3.4/spark-3.3.4-bin-hadoop3.tgz" 2. Copy the "conf" files from the Cloudera cluster and send them to the new Spark directory 3. exported the variables "HADOOP_CONF_DIR" and "SPARK_CONF_DIR" and "SPARK_HOME" using the new spark directory "spark-3.3.4-bin-hadoop3" with the files 4. When trying to run spark-shell as an example, nothing happens, it hangs as shown below: Setting default log level to "WARN". To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel). Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 3.3.4 /_/ Using Scala version 2.13.8 (Java HotSpot(TM) 64-Bit Server VM, Java 11.0.16.1) Type in expressions to have them evaluated. Type :help for more information. note: the cluster has kerberos, so before running spark-shell, kinit was run
... View more
Labels:
01-16-2024
04:53 AM
1 Kudo
hi @PrathapKumar thanks for the answer! 😉 but, I had already seen this KB from fixed CVE 😊
... View more
12-19-2023
12:07 PM
hi @vaishaakb , 1. Impala-shell queries work without any problems 2. Nifi version 2.2.6.4 on CDP Public Cloud 3. the problem is in the impala connection controller on the nifi to the impala which is on another datahub 4. idle_session_timeout = no value idle_query_timeout = no value fe_service_threads = 64 idle_client_poll_time_s = this parameter does not exist in Impala 5. the environment is Kerberos and TLS enabled
... View more
11-23-2023
07:00 AM
1 Kudo
hi community cloudera, we are verifying a strange behavior of an Impala connection controller on nifi, when the processor starts the connection, it displays the error below: [Cloudera][ImpalaJDBCDriver](500593) Communication link failure. Failed to connect to server. Reason: java.net.SocketException: Connection or outbound has closed communication between hosts is ok, ports communicating without problem. when we start and stop the processor the connection occurs without problem, this shows that the error is intermittent. has anyone experienced this type of problem?
... View more
Labels:
- Labels:
-
Apache Impala
-
Apache NiFi
-
Cloudera Manager
08-10-2023
09:35 AM
i managed to fix the problem i changed the VERSION file of another disk that is OK and datanode was able to recognize the disk again regards
... View more
08-10-2023
07:53 AM
hello cloudera community, I have a problem in a datanode only in hdfs this datanode has 10 disks and in the hdfs settings (dfs.datanode.data.dir) is configured the 10 disks but for some time this datanode is not recognizing one of the disks configured only was already done the restart of the datanode and even so did not back to recognize looking at the datanode logs, it is not really recognizing the disk, only the 9 disks these disks are also being used by other services in the cluster and these other services are able to recognize the 10 disks without any problem, only hdfs is having this problem could you help us to solve this problem?
... View more
Labels:
- Labels:
-
Cloudera Manager
-
HDFS
07-06-2023
07:53 AM
hello cloudera community, do I need to insert a new datanode in the hdfs and with that I need to know how much size will be used from this new datanode after performing a rebalance on the hdfs? for example, I have the following scenario: datanode 1 has 36tb total to use, 30tb is being used, containing about 2.5 million blocks datanode 2 has 36tb total to use, 30tb is being used, containing about 2.5 million blocks datanode 3 has 36tb total to use, 30tb is being used, containing about 2.5 million blocks datanode 4 has 36tb total to use, 30tb is being used, containing about 2.5 million blocks datanode 5 has 36tb total to use, 30tb is being used, containing about 2.5 million blocks datanode 6 has 36tb total to use, 30tb is being used, containing about 2.5 million blocks i will insert the "datanode 7" which has 36tb total, after performing the rebalance in hdfs, how much will this new datanode 7 receive data?
... View more
Labels:
- Labels:
-
Cloudera Manager
-
HDFS
06-23-2023
05:52 AM
hello cloudera community, today the cluster already has thousands of blocks in hdfs, more than 23 million blocks. after configuring the rack in the cluster, the hdfs will recognize the racks and will start moving the blocks to the racks to increase the availability of the blocks or will i have to rebalance the hdfs?
... View more
Labels:
- Labels:
-
Cloudera Manager
-
HDFS
06-22-2023
07:44 AM
hi @rki_ / @cjervis , I forgot to ask, today the cluster already has thousands of blocks in hdfs, more than 23 million blocks. after configuring the rack in the cluster, the hdfs will recognize the racks and will start moving the blocks to the racks to increase the availability of the blocks or will i have to rebalance the hdfs?
... View more
06-22-2023
07:00 AM
hi @rki_ excellent, thanks a lot for the help. hugs. have a nice day!
... View more