Member since
08-25-2020
2
Posts
0
Kudos Received
0
Solutions
10-06-2020
12:51 PM
So, we recently moved to using spark as HIVE engine instead of mr and we are seeing some significant improvements in certain queries which needs intermediate tables/storage to process. Can someone provide a complete or optimum list of configuration that we should be using that will not cause memory issues but still be able to get the best out of using spark as engine? Nodes : 30 Cores: 16 Memory: 112GB /node Hadoop 2.6.0-cdh5.13.0 Hive 1.1.0-cdh5.13.0
... View more
Labels: