Support Questions

Find answers, ask questions, and share your expertise
Announcements
We’ve updated our product names and community labels - click here for full details

Reading 1000 tables from RDBMS via pyspark and create parquet tables in hive

avatar
Contributor

Team,

 

I have 1000 tables in my source RDBMS and I would like to get them migrated to hive using pyspark

 

I read through documentation and found that below two commands would help. Is there a way I can loop these two commands 1000 times if I have all the list of tables in a python array?

 

arr = ("table1","table2")

for x in arr:

            df = spark.read.format("jdbc").blah.blah 

 

            data.write.saveAsTable.blah.blah

 

If someone has a working solution for this could you please share. I tried but it is not throwing any error but at same time not writing anything.

 

Thanks

Meher

 

1 ACCEPTED SOLUTION

avatar
Contributor

I'm able to get this working. Will close this post. 

 

Thanks,

Meher

View solution in original post

2 REPLIES 2

avatar
Contributor

I'm able to get this working. Will close this post. 

 

Thanks,

Meher

avatar
Community Manager

@Meher I am happy to see that you resolved your issue. Would you mind sharing how you solved it in case someone else encounters the same situation?


Keep the questions coming,

Cy Jervis | Senior Manager, Knowledge Programs

if (helpful) { mark_as_solution(); } | if (appreciated) { give_kudos(); }