Member since
03-01-2016
609
Posts
12
Kudos Received
7
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1600 | 02-20-2024 10:42 PM | |
1948 | 10-26-2023 05:40 PM | |
1266 | 06-13-2023 07:55 PM | |
2053 | 04-28-2019 12:21 AM | |
1376 | 04-28-2019 12:12 AM |
03-03-2024
11:32 PM
@yanseoi, Did the response assist in resolving your query? If it did, kindly mark the relevant reply as the solution, as it will aid others in locating the answer more easily in the future.
... View more
01-02-2024
03:09 AM
To save a DataFrame as a text file in PySpark, you need to convert it to an RDD first, or use DataFrame writer functions. Using DataFrame writer: df.write.format("text").save("path_to_output_directory") Converting to RDD and then using saveAsTextFile rdd = df.rdd.map(lambda row: str(row)) rdd.saveAsTextFile("path_to_output_directory")
... View more
10-30-2023
10:40 PM
@AK- Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future. Thanks.
... View more
06-14-2023
02:01 PM
Hi Yuexin, you have been very helpful. Unfortunately, if I wanted to use "Dynamic Queue Scheduling" in CDP 717 at the moment, I would not have the guarantee to solve any problems via Cloudera support. In fact, it is not recommended to use it in production. Thank you very much
... View more
03-29-2021
12:36 AM
@ishika as this is an older post, you would have a better chance of receiving a resolution by starting a new thread. This will also be an opportunity to provide details specific to your environment that could aid others in assisting you with a more accurate answer to your question. You can link this thread as a reference in your new post.
... View more
01-05-2020
06:33 AM
Hi, This parameter spark.executor.memory (or) spark.yarn.executor.memoryOverhead can be set in Spark submit command or you can set it Advanced configurations. Thanks AKR
... View more
09-04-2019
10:31 AM
Hi, Check for the total no of applications in the Application history path, if the total no of files is more try to increase the heap size and look whether it works. Alternatively look for the spark history server logs too for any errors. Thanks AKR
... View more
05-28-2019
10:30 AM
Hi Sara, I run vulnerability scans and our scanner picking up Spark 1.6 banner from the following path- for CVE-2018-8024 vulnerability, you did mention this vulnerability doesn't affect SPark 1.6 but didn't give detail reasons. This is where Qualys picksup the banner- /opt/cloudera/parcels/CDH-5.15.1-1.cdh5.15.1.p0.4/lib/spark/conf/spark-env.sh: line 75: /usr/appl/cloudera/java/jdk1.8.0_162: is a directory Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.6.0 /_/ Also we have 2 versions of SPARK running- do you really need version 1.6 to run v 2.3.0 Can you please help, advise.
... View more
05-07-2019
09:58 PM
Depends on what you mean by 'storage locations'. If you mean "can other apps use HDFS?" then the answer is yes, as HDFS is an independent system unrelated to YARN and has its own access and control mechanisms not governed by a YARN scheduler. If you mean "can other apps use the scratch space on NM nodes" then the answer is no, as only local containers get to use that. If you're looking to strictly split both storage and compute, as opposed to just some form of compute, then it may be better to divide up the cluster entirely.
... View more
05-07-2019
02:26 AM
2 Kudos
Hi MKay, As mentioned in my previous posts the Anaconda parcel for CDH comes only with Python 2.7 and I could find a free way to get a parcel with Python 3+. We ended up manually installing the different Python versions we needed by keeping different virtual envs for different Python versions. We executed the following procedure to install python 3.5: yum install python-pip curl "https://bootstrap.pypa.io/get-pip.py" -o "get-pip.py" python get-pip.py pip install virtualenv yum install -y https://centos7.iuscommunity.org/ius-release.rpm yum install -y python35u python35u-libs python35u-devel python35u-pip mkdir -p /opt/venv35 cd /opt/venv35 virtualenv venv35 -p python3.5 source venv35/bin/activate Best, Eyal
... View more