Member since
06-02-2020
331
Posts
67
Kudos Received
49
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2806 | 07-11-2024 01:55 AM | |
| 7876 | 07-09-2024 11:18 PM | |
| 6579 | 07-09-2024 04:26 AM | |
| 5914 | 07-09-2024 03:38 AM | |
| 5624 | 06-05-2024 02:03 AM |
10-11-2022
02:13 AM
Hi @Ploeplse Still, if you are facing the issue, could you share the requested information (i.e code and impala table creation script)
... View more
09-30-2022
02:01 AM
Hi @imule Add the following parameter to your spark-submit --conf spark.yarn.appMasterEnv.PYSPARK_PYTHON=<python3_path>
--conf spark.yarn.appMasterEnv.PYSPARK_DRIVER_PYTHON=<python3_path> Note: 1. Ensure python3_path exists in all nodes. 2. Ensure required modules are installed in each node.
... View more
09-21-2022
10:22 PM
Hi @Boron Could you please set the spark-home environment variable like below before creating spark-session. import os
os.environ['SPARK_HOME'] = '/usr/hdp/current/spark-client' Reference: https://stackoverflow.com/questions/55569985/pyspark-could-not-find-valid-spark-home https://stackoverflow.com/questions/40087188/cant-find-spark-submit-when-typing-spark-shell
... View more
09-16-2022
05:19 AM
Hi @poorva Please check the application logs for failed applications from Resource Manager UI, there exception message is printed. Fix the exception and resubmit the job.
... View more
09-14-2022
12:54 AM
Hi @Ploeplse Could you please share reproducible sample code and impala tab creation script?
... View more
08-31-2022
10:53 PM
Hi @Yosieam Please avoid calling read_file_log.collect() method. It will bring whole data to the driver and the driver needs to have more memory to hold that much data. Please check the modified code: move_to_rdd = sc.textFile("datalog2.log").map(lambda row : row.split("time=")).filter(lambda x : x != "")
ReSymbol = move_to_rdd.map(lambda x : re.sub(r'\t', ' ', x)).map(lambda x : re.sub(r'\n', ' ', x)).map(lambda x : re.sub(r' +', ' ', x))
... View more
08-31-2022
10:48 PM
Hi @mmk I think you have shared the following information. 7 nodes with each having 250 gb memory and vcpu = 32 per each node spark-defaults.conf spark.executor.memory = 100g spark.executor.memoryOverhead = 49g spark.driver.memoryOverhead=200g spark.driver.memory = 500g You have maximum of 250 gb for node and you have specified driver memory is (500gb and 200gb overhead). How it possible to driver to get 700gb? Generally you should not exceed the driver/executor memory beyond yarn physical memory. Coming to the actual problem, please avoid the show() to print 8000000 records. If you need to get the print the all values, then implement a logic to 1000 records at once and next 1000 records for another iteration. https://stackoverflow.com/questions/29227949/how-to-implement-spark-sql-pagination-query
... View more
08-31-2022
09:45 PM
Hi @mmk By default, Hive will load all SerDe under the hive/lib location. So you are able to do the create/insert/select operations. In order to read the Hive table created with Custom or external SerDe we need to provide to spark, so spark internally use those libraries and it will load the Hive table data. If you are not provided the serde you can see the following exception: org.apache.hadoop.hive.serde2.SerDeException Please add the following library to the spark-submit command: json-serde-<version>.jar
... View more
08-31-2022
09:36 PM
Hi @suri789 I think you haven't shared the full code, sample data and expected output to provide a solution. Please share the code proper format.
... View more
08-31-2022
09:33 PM
Hi @AZIMKBC Please try to run the SparkPi example and see if is there any error in the logs. https://rangareddy.github.io/SparkPiExample/ If still issue is not resolved and you are a Cloudera customer please raise a case we will work on internally.
... View more