Role Overview: We are seeking a highly skilled and motivated Data Engineer with expertise in Python, PySpark , Databricks, and SQL. The ideal candidate will have experience working with large datasets, optimizing performance, and leveraging cloud-based data platforms. If you are passionate about building scalable data solutions and enjoy problem-solving, weu2019d love to hear from you.
Experience: 5-8 Years
Location: Pune||Mumbai
Key Responsibilities:
Develop and maintain scalable data pipelines using Python and PySpark .
Work on Databricks to build data workflows and manage big data transformations.
Write complex SQL queries for data analysis and transformation.
Optimize performance for large datasets across distributed computing environments.
Collaborate with cross-functional teams to understand data requirements and implement robust solutions.
Debug and troubleshoot data pipeline and processing issues.
Leverage cloud services (preferably Azure or GCP) to support data storage, processing, and orchestration.
Required Skills:
Proficiency in Python, PySpark , Databricks, and SQL.
Hands-on experience in performance tuning for large datasets.
Strong debugging and analytical problem-solving skills.
Basic knowledge of cloud platforms, preferably Azure or Google Cloud Platform (GCP).
Good to Have:
Experience with BigQuery .
Familiarity with Delta Lake or data warehouse technologies.
Job Classification
Industry: IT Services & ConsultingFunctional Area / Department: Engineering - Software & QARole Category: Software DevelopmentRole: Software Development - OtherEmployement Type: Full time