The basic principles and concepts behind Spark, as a framework for distributed processing.
How do we process massive distributed data sets in a cluster? With high-level APIs! We have two major alternative APIs available at your fingertips: statically typed and dynamically typed.
The last module focused on transformations, whereas this one focuses on the data side: formats, optimizations, management, etc.
Learn how to take the most of the spark optimizations for free.
Learn the best ways to optimize and organize your Spark code to make it more robust and performant.
Your email address will not be published. Required fields are marked *
Save my name, email, and website in this browser for the next time I comment.
Δ