Created on 07-17-2017 11:49 AM - edited 08-18-2019 01:59 AM
Hi. I have a problem with Spark 2 interpreter in Zeppelin. I configured interpreter like this:
When I run query like this:
%spark2.sql select var1, count(*) as counter from database.table_1 group by var1 order by counter desc
Spark job runs only 3 containers and job takes 13 minutes.
Does anyone know why Spark interpreter takes only 4.9 % of queue? How I should configure the interpreter to increase this factor?
Created 07-17-2017 09:24 PM
@Mateusz Grabowski, You should enable Dynamic Resource Allocation in Spark to automatically increase/decrease executors of an app as per resource availability.
You can choose to enable DRA in either Spark or Zeppelin .
1) Enable DRA for Spark2 as below.
2) Enable DRA via Livy Interpreter. Run all spark notebooks via livy interpreters.
https://zeppelin.apache.org/docs/0.6.1/interpreter/livy.html
Created 07-17-2017 09:24 PM
@Mateusz Grabowski, You should enable Dynamic Resource Allocation in Spark to automatically increase/decrease executors of an app as per resource availability.
You can choose to enable DRA in either Spark or Zeppelin .
1) Enable DRA for Spark2 as below.
2) Enable DRA via Livy Interpreter. Run all spark notebooks via livy interpreters.
https://zeppelin.apache.org/docs/0.6.1/interpreter/livy.html
Created 07-18-2017 10:01 AM
It works! Thank you 🙂