Apache Spark: Apply & Evaluate Big Data Workflows
This course introduces beginners to the foundational and intermediate concepts of distributed data processing using Apache Spark, one of the most powerful engines for large-scale analytics. Through two progressively structured modules, learners will identify Spark’s architecture, describe its core components, and demonstrate key programming constructs such as Resilient Distributed Datasets (RDDs).
In Module 1, learners will recognize the principles behind Spark’s distributed computing model and illustrate basic RDD transformations. In Module 2, they will apply advanced transformation logic, i…
Watch on Coursera ↗
(saves to browser)
DeepCamp AI