on 09-08-2016 09:52 AM
Top 5 mistakes when writing Spark Applications
In the world of distributed computing, Spark has simplified development and open the doors for many to start writing distributed programs. Folks with little to none distributed coding experience can now start writing just a couple lines of code that will get 100s or 1000s of machines, immediately, working on creating business value. However, even through Spark code is easy to write and read, that doesn’t mean that users don’t run into issues of long running, slow performing jobs or out of memory errors. Thankfully most of the issues with using Spark have nothing to do with Spark but the approach we take when using it. This session will go over the top 5 things that we’ve seen in the field that prevent people from getting the most out of their Spark clusters. When some of these issues are addressed, it is not uncommon to see the same job running 10x or 100x faster with the same clusters, the same data, just a different approach.
Mark Grover is a software engineer working on Apache Spark at Cloudera. He is a co-author of the O'Reilly book Hadoop Application Architectures and also wrote a section in Programming Hive book. Mark is also a committer on Apache Bigtop and a committer and PMC member on Apache Sentry. He has contributed to a number of open source projects including Apache Hadoop, Apache Hive, Apache Sqoop and Apache Flume projects. Mark is sought after speaker on topics related to Big Data at various national and international conferences. He occasionally blogs on topics related to technology on his blog.
• 6:30 Networking, Food & Drinks
• 7:00 Announcements and featured talk
• 8:30 Adjourn to pub