Build and maintain robust ETL pipelines using Python, Dataflow, and Apache Beam
Design scalable data processing workflows using Cloud Composer (Airflow DAGs)
Work with structured and semi-structured data in BigQuery, optimizing for performance and cost
Implement data transformation and enrichment processes to support reporting and analytics
Develop distributed data processing solutions using DataProc (Spark/Hadoop)
Collaborate with data scientists, analysts, and engineers to integrate data solutions across teams
Monitor data pipeline performance and troubleshoot issues proactively
Maintain data pipeline documentation and follow best practices for data governance
Requirements:
Proficiency in Python for scripting and data manipulation
Strong knowledge of SQL for querying and transforming large datasets
Hands-on experience with Google Cloud Platform data services:
BigQuery
Cloud Dataflow
Cloud Composer (Airflow DAGs)
DataProc (Spark/Hadoop)
Solid understanding of ETL concepts, data modeling, and workflow orchestration
Experience working in agile teams and delivering production-ready data solutions
Preferred Qualifications:
GCP certification (e.g., Professional Data Engineer)
Familiarity with CI/CD practices for data engineering
Experience with streaming data and real-time pipeline architectures
Understanding of data security, privacy, and compliance in cloud environments
Employement Category:
Employement Type: Full timeIndustry: IT Services & ConsultingRole Category: DBA / DatawarehousingFunctional Area: Not SpecifiedRole/Responsibilies: Data Engineer