Primary Skill: GCP with Pyspark Secondary Skill: Big data , Spark SQL , Hadoop, Hive. 9-12 years of Data Engineering experience. Experience building Data pipelines and ELT/ETL. Experience with SQL and NoSQL modern data stores. Experience in implementing Data Pipelines leveraging Google Cloud products such as Cloud BigQuery, GCS, Cloud DataFlow, Cloud Pub/Sub, Cloud BigTable. Implementation/migration of ODS data from on-prem to GCP (using Google BigQuery, DataFlow, DataProc etc.) Act as a subject matter expert in data engineering and GCP data technologies. Work with Data teams to design and implement modern, scalable data solutions using a range of new and emerging technologies from the Google Cloud Platform. Experience working with Google Data Products (BigQuery, Dataflow, Dataproc, Dataprep, PubSub). Work with Agile and DevOps techniques and implementation approaches in the delivery. Build Data pipelines in setting up new Data warehouse environments and data flows. Be required to build and deliver Data solutions using GCP products and offerings. Responsible for extract, transform and load (ETL) processes and the creation of applications that can connect to remote APIs. Including bringing ODS data into environments such as BigQuery on Google Cloud Management Platform. Experience extracting data from a variety of sources, and a desire to expand those skills (working knowledge in SQL and Spark is mandatory) Excellent Data Analysis skills. Must be comfortable with querying and analyzing large amount of data on Hadoop HDFS using Hive and Spark. GCP certifications preferred. Strong work ethic and ability to handle high pressure situations with minimum direction. Team player, with proven ability to work under pressure and within tight deadlines. Ability to multi-task and prioritize work load effectively. Strong verbal and written communication skills. Skills: python,python.,python,
Employement Category:
Employement Type: Full timeIndustry: IT Services & ConsultingRole Category: Application Programming / MaintenanceFunctional Area: Not SpecifiedRole/Responsibilies: Big Data( GCP, Pyspark ) Shailja Job in