Key Components of Spark

Now let's see how Spark actually works, how it thinks about processing data and how we understand scale. We'll also look at the various libraries available - modern Spark is a series of libraries built upon each other to make the engine as flexible and adaptable as any system we've seen before.

In this video we look at the cluster architecture, what the core parts of Spark are and touch on RDDs (Resilient Distributed Datasets). They're a fundamental building block of Spark, but honestly, the engine is so good these days that we barely need to think about them!

Complete and Continue