In this video, we use PySpark to analyze data with Resilient Distributed Datasets (RDD).  RDDs are the foundation of Spark. You learn what RDDs are, what Lazy Evaluation is and why it matters, and how to use Transformations and Actions.  Everything is demonstrated using a Databricks notebook.

#apache-spark #bigdata 

Use PySpark to analyze Data with RDDs
2.95 GEEK