Cloud Hadoop: Scaling Apache Spark
With Lynn Langit
Liked by 425 users
Duration: 3h 13m
Skill level: Beginner + Intermediate
Released: 4/1/2020
Course details
Apache Hadoop and Spark make it possible to generate genuine business insights from big data. The Amazon cloud is natural home for this powerful toolset, providing a variety of services for running large-scale data-processing workflows. Learn to implement your own Apache Hadoop and Spark workflows on AWS in this course with big data architect Lynn Langit. Explore deployment options for production-scaled jobs using virtual machines with EC2, managed Spark clusters with EMR, or containers with EKS. Learn how to configure and manage Hadoop clusters and Spark jobs with Databricks, and use Python or the programming language of your choice to import data and execute jobs. Plus, learn how to use Spark libraries for machine learning, genomics, and streaming. Each lesson helps you understand which deployment option is best for your workload.
Skills you’ll gain
Earn a sharable certificate
Share what you’ve learned, and be a standout professional in your desired industry with a certificate showcasing your knowledge gained from the course.
LinkedIn Learning
Certificate of Completion
-
Showcase on your LinkedIn profile under “Licenses and Certificate” section
-
Download or print out as PDF to share with others
-
Share as image online to demonstrate your skill
Meet the instructor
Learner reviews
-
Alireza Sharifikia
Alireza Sharifikia
Senior Group Product Manager at Corpy & Co.
-
Stephen Davison
Stephen Davison
Bioinformatics Engineer of Product Features and Systems
-
Motshedisi Rose Korir
Motshedisi Rose Korir
Motshedisi Rose Korir
Contents
What’s included
- Practice while you learn 1 exercise file
- Learn on the go Access on tablet and phone