Share on LinkedIn Feed Twitter Facebook

Data Lake Modernization on Google Cloud: Migrate Workflows

Data Lake Modernization on Google Cloud: Migrate Workflows

magic_button Cloud Dataproc Data Lake Apache Spark
These skills were generated by A.I. Do you agree this course teaches these skills?
7 hours 15 minutes Intermediate universal_currency_alt 5 Credits

Welcome to Migrate Workflows, where we discuss how to migrate Spark and Hadoop tasks and workflows to Google Cloud.

Complete this activity and earn a badge! Boost your cloud career by showing the world the skills you’ve developed.

Badge for Data Lake Modernization on Google Cloud: Migrate Workflows
info
Course Info
Objectives
  • Describe options on migrating Spark to Google Cloud
  • Migrate spark Jobs directly to to Dataproc (“Lift and Shift”)
  • Optimize Spark Jobs to run on Google Cloud
  • Refactor Spark Jobs to use native Google Cloud services
  • Convert a Spark job to a serverless implementation using Cloud Functions
Prerequisites
Required: Have completed the Data Engineering on Google Cloud training, Be a Google Cloud Certified Professional Data Engineer or have equivalent expertise in Data Engineering, Have access to Cloud Connect – Partners. Recommended: Experience building data processing pipelines, Experience with Apache Beam and Apache Hadoop Java or Python programming expertise. Organizational requirements: The Cloud Partner organization must have implemented at least one Data Warehouse solution previously on any Data Warehouse platform.
Audience
Data Warehouse Deployment Engineers, Data Warehouse Consultants, Data Warehouse Architects. Technical Project Leads, Technical Project Managers, Data / Business Analyst.
Available languages
English
Preview