- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
Reading 1000 tables from RDBMS via pyspark and create parquet tables in hive
- Labels:
-
Apache Spark
Created on ‎01-07-2019 08:21 PM - edited ‎09-16-2022 07:02 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Team,
I have 1000 tables in my source RDBMS and I would like to get them migrated to hive using pyspark
I read through documentation and found that below two commands would help. Is there a way I can loop these two commands 1000 times if I have all the list of tables in a python array?
arr = ("table1","table2")
for x in arr:
df = spark.read.format("jdbc").blah.blah
data.write.saveAsTable.blah.blah
If someone has a working solution for this could you please share. I tried but it is not throwing any error but at same time not writing anything.
Thanks
Meher
Created ‎01-08-2019 09:11 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Created ‎01-08-2019 09:11 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
I'm able to get this working. Will close this post.
Thanks,
Meher
Created ‎01-09-2019 05:57 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
@Meher I am happy to see that you resolved your issue. Would you mind sharing how you solved it in case someone else encounters the same situation?
Cy Jervis, Manager, Community Program
Was your question answered? Make sure to mark the answer as the accepted solution.
If you find a reply useful, say thanks by clicking on the thumbs up button.
