$140K - 180K a year
Design, build, and maintain scalable data processing systems on Google Cloud, managing the entire data lifecycle to enable data-driven decisions.
8+ years data engineering experience, Google Cloud Platform certification, strong SQL and Python skills, hands-on experience with multiple Google Cloud data services, data migration expertise, Terraform and CI/CD familiarity.
Google Cloud Data Migration Data Migration Team Lead Job Summary: Seeking a Google Cloud data engineer to design, build, and maintain scalable and efficient data processing systems on the Google Cloud platform. This engineer will be responsible for the entire data lifecycle, from ingestion and storage to processing, transformation, and analysis. Their work will enable client organizations to make data-driven decisions by providing clean, high-quality data to business intelligence tools, AI systems, analysts and data scientists. Required Skills & Qualifications: Google Cloud Platform Certified professional 8+ years of data engineering experience developing large data pipelines in very complex environments Very Strong SQL skills and ability to build very complex transformation data pipelines using custom ETL framework in Google BigQuery environment Very strong understanding of data migration methods and tooling, with hands-on experience in at least three (3) data migrations to Google Cloud Google Cloud Platform: Hands-on experience with key Google Cloud Platform data services is essential, including: o BigQuery: For data warehousing and analytics. o Cloud Dataflow: For building and managing data pipelines. o Cloud Storage: For storing large volumes of data. o Cloud Composer: For orchestrating workflows. o Cloud Pub/Sub: For real-time messaging and event ingestion. o DataProc: For running Apache Spark and other open-source frameworks. Programming languages: Strong proficiency in programming languages, most commonly Python, is mandatory. Experience with Java or Scala is also preferred. SQL expertise: Advanced SQL skills for data analysis, transformation, and optimization within BigQuery and other databases. ETL/ELT: Deep knowledge of Extract, Transform, Load (ETL) and Extract, Load, Transform (ELT) processes. Infrastructure as Code (IaC): Experience with tools like Terraform for deploying and managing cloud infrastructure. CI/CD: Familiarity with continuous integration and continuous deployment (CI/CD) pipelines using tools such as GitHub Actions or Jenkins. Data modeling: Understanding of data modeling, data warehousing, and data lake concepts
This job posting was last updated on 10/11/2025