Support Questions

Find answers, ask questions, and share your expertise
Announcements
Celebrating as our community reaches 100,000 members! Thank you!

How to run spark sql in parallel?

avatar
Explorer

We are doing spark programming in java language. The ‘DataFrame’ has been stored in temporary table and we are running multiple queries from this temporary table inside loop. The quires are running in sequential order. We need to run in parallel from temporary table. Please find code snippet below.

 

Thanks in advance for your cooperation.

 

HiveContext hiveContext = new HiveContext(sparkContext);

              String mainQueryHql = getFileContent(mainQueryFilePath);

              DataFrame df = hiveContext.sql(mainQueryHql).persist(StorageLevel.MEMORY_AND_DISK_SER());

              df.show();

              System.out.println("Total Records in Main Query " + df.count());

             

              df.registerTempTable(tempTable);

              ArrayList<DataFrame> dataFrameList = new ArrayList();

             

              DataFrame dfSubQuery = null;

              String subQuires = getFileContent(subQueryFilePath);

             

              String[] alQuires = subQuires.split(";");

             

              for(int i=0; i<alQuires.length; i++) {

                     System.out.println("Query no " + i +" is : " + alQuires[i]);

                     logger.debug("Query no " + i +" is : " + alQuires[i]);

                     dfSubQuery = hiveContext.sql(alQuires[i]);                   

                     dfSubQuery.show();

                     dataFrameList.add(dfSubQuery);

              }

1 ACCEPTED SOLUTION

avatar
Expert Contributor

There is nothing native within Spark to handle running queries in parallel.  Instead you can take a look at Java concurrency and in particular Futures[1] which will allow you to start queries in parallel and check status later.

 

1.  https://docs.oracle.com/javase/7/docs/api/java/util/concurrent/Future.html

 

 

View solution in original post

6 REPLIES 6

avatar
Expert Contributor

There is nothing native within Spark to handle running queries in parallel.  Instead you can take a look at Java concurrency and in particular Futures[1] which will allow you to start queries in parallel and check status later.

 

1.  https://docs.oracle.com/javase/7/docs/api/java/util/concurrent/Future.html

 

 

avatar
Explorer
Thanks Hubbarja, Will check and respond very shortly.

avatar
New Contributor

Hi Kamalakanta,

 

I am also in need of a solution for this problem.

Have you checked with Java Concurrency and Futures ? or else do u got any other solution?

Please share the solution.

I am also in need of executing the sparksql in parallel(similar to your program).

 

 

Thanks in advance

avatar
Explorer

Please use Future interface in your code. Request to explorebelow mentioned object or interfaces.

java.util.concurrent.ExecutionException;
java.util.concurrent.ExecutorService;
java.util.concurrent.Executors;
java.util.concurrent.Future;

 

You can get more information from below mentioned URLs.

https://docs.oracle.com/javase/7/docs/api/java/util/concurrent/Future.html

 

 

avatar
Explorer

avatar
New Contributor

It will be of great help to me.
Thanks a lot Kamalakanta 🙂

 

I have some more doubts regarding running spark SQL queries in parallel.