1. Home
  2. Data Science
  3. Big Data
  4. Course detail

Apache Spark with Scala — Hands On with Big Data!

  • Job DurationUdemy
  • Job Duration9 hours worth of material
  • Job DurationPaid Course

Project detail


Apache Spark tutorial with 20+ hands-on examples of analyzing large data sets, on your desktop or on Hadoop with Scala!

What you’ll learn:

  • Develop distributed code using the Scala programming language
  • Transform structured data using SparkSQL, DataSets, and DataFrames
  • Frame big data analysis problems as Apache Spark scripts
  • Optimize Spark jobs through partitioning, caching, and other techniques
  • Build, deploy, and run Spark scripts on Hadoop clusters
  • Process continual streams of data with Spark Streaming
  • Traverse and analyze graph structures using GraphX
  • Analyze massive data set with Machine Learning on Spark

New! Completely updated and re-recorded for Spark 3, IntelliJ, Structured Streaming, and a stronger focus on the DataSet API.

“Big data» analysis is a hot and highly valuable skill – and this course will teach you the hottest technology in big data:Apache Spark. Employers includingAmazon,EBay,NASA JPL, andYahooall use Spark to quickly extract meaning from massive data sets across a fault-tolerantHadoopcluster. You’ll learn those same techniques, using your own Windows system right at home. It’s easier than you might think, and you’ll belearning from an ex-engineer and senior manager from Amazon andIMDb.

Spark works best when using the Scala programming language, and this course includes a crash-course inScala to get you up to speed quickly. For those more familiar with Python however, a Python version of this class is also available:»Taming Big Data withApacheSpark and Python -Hands On».

Learn and master the art of framing data analysis problems as Spark problems through over 20hands-on examples, and then scale them up to run on cloud computing services in this course.

  • Learn the concepts of Spark’s Resilient Distributed Datasets, DataFrames, and Datasets.

  • Get a crash course in the Scala programming language

  • Develop and run Spark jobs quickly using Scala, IntelliJ, and SBT

  • Translate complex analysis problems into iterative or multi-stage Spark scripts

  • Scale up to larger data sets using Amazon’sElastic MapReduceservice

  • Understand howHadoop YARNdistributes Spark across computing clusters

  • Practice using other Spark technologies, like Spark SQL, DataFrames, DataSets,Spark Streaming, Machine Learning, and GraphX

By the end of this course, you’ll be running code that analyzes gigabytes worth of information – in the cloud – in a matter of minutes.

We’ll have some fun along the way. You’ll get warmed up with some simple examples of using Spark to analyze movie ratings data and text in a book. Once you’ve got the basics under your belt, we’ll move to some more complex and interesting tasks. We’ll use a million movie ratings to find movies that are similar to each other, and you might even discover some new movies you might like in the process! We’ll analyze a social graph of superheroes, and learn who the most “popular» superhero is – and develop a system to find “degrees of separation» between superheroes. Are all Marvel superheroes within a few degrees of being connected to SpiderMan? You’ll find the answer.

This course is very hands-on; you’ll spend most of your time following along with the instructor as we write, analyze, and run real code together – both on your own system, and in the cloud using Amazon’s Elastic MapReduce service. over 8 hours of videocontent is included, withover 20real examplesof increasing complexity you can build, run and study yourself. Move through them at your own pace, on your own schedule. The course wraps up with an overview of other Spark-based technologies, including Spark SQL, Spark Streaming, and GraphX.

Enroll now, and enjoy the course!

«I studied Spark for the first time using Frank’s course «Apache Spark 2 with Scala — Hands On with Big Data!». It was a great starting point for me, gaining knowledge in Scala and most importantly practical examples of Spark applications. It gave me an understanding of all the relevant Spark core concepts, RDDs, Dataframes & Datasets, Spark Streaming, AWS EMR. Within a few months of completion, I used the knowledge gained from the course to propose in my current company to work primarily on Spark applications. Since then I have continued to work with Spark. I would highly recommend any of Franks courses as he simplifies concepts well and his teaching manner is easy to follow and continue with! « — JoeyFaherty

Industry Categories

Languages required