Google Cloud Platform Data Engineer (Python)

Overview

On Site
Depends on Experience
Contract - Independent
Contract - W2
Contract - 9 Month(s)
No Travel Required

Skills

Apache Airflow
Cloud Computing
Cloud Storage
Collaboration
Continuous Delivery
Continuous Integration
Data Engineering
Data Flow
Data Governance
Data Integrity
Data Lake
Data Modeling
Data Processing
Data Quality
Data Validation
Data Warehouse
Docker
Documentation
ELT
Extract
Transform
Load
GitHub
Good Clinical Practice
Google Cloud
Google Cloud Platform
Jenkins
Kubernetes
Management
Orchestration
Python
SQL
Scripting
Terraform
Workflow
Writing

Job Details

Job Title: Data Engineer Google Cloud Platform & Python Development
Location: Houston, TX
Key Responsibilities
  • Design, develop, and maintain reliable and scalable data pipelines on Google Cloud Platform using tools like Dataflow, BigQuery, Pub/Sub, and Cloud Composer.
  • Write efficient and reusable Python scripts and modules for ETL/ELT workflows and data transformations.
  • Collaborate with data scientists, analysts, and other engineers to integrate data from various sources, ensure data quality, and optimize performance.
  • Build and manage data lake and data warehouse solutions leveraging BigQuery and Cloud Storage.
  • Automate data validation and monitoring workflows for data integrity and reliability.
  • Implement CI/CD pipelines for data engineering workflows using tools like Cloud Build, GitHub Actions, or Jenkins.
  • Monitor and optimize job performance, cost efficiency, and error handling across Google Cloud Platform services.
  • Maintain proper documentation of data flows, schemas, and transformation logic.

Requirements
Technical Skills
  • Strong proficiency in Python, with experience in writing scalable, modular, and testable code.
  • Solid experience with Google Cloud Platform especially BigQuery, Cloud Functions, Cloud Storage, Dataflow, Pub/Sub, Cloud Composer (Airflow).
  • Experience with SQL and building optimized queries for large-scale data processing.
  • Hands-on experience with data orchestration tools like Apache Airflow (Composer preferred).
  • Knowledge of data modeling, data warehousing concepts, and data governance best practices.
  • Familiarity with Docker, Terraform, or Kubernetes is a plus.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.