Options
- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
Solved
Go to solution
Optimum HIVE parameters for using hive.execution.engine=spark ?
Labels:
New Contributor
Created ‎10-06-2020 12:51 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
So, we recently moved to using spark as HIVE engine instead of mr and we are seeing some significant improvements in certain queries which needs intermediate tables/storage to process. Can someone provide a complete or optimum list of configuration that we should be using that will not cause memory issues but still be able to get the best out of using spark as engine?
Nodes : 30
Cores: 16
Memory: 112GB /node
Hadoop 2.6.0-cdh5.13.0
Hive 1.1.0-cdh5.13.0
1 ACCEPTED SOLUTION
Expert Contributor
Created ‎10-12-2020 01:20 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
There is no such optimal configurations. Fine tuning will be done based upon your load and how HOS reacts on the workload submitted on it. Again this will vary from customer to customer
1 REPLY 1
Expert Contributor
Created ‎10-12-2020 01:20 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
There is no such optimal configurations. Fine tuning will be done based upon your load and how HOS reacts on the workload submitted on it. Again this will vary from customer to customer
