Your browser does not support javascript! Please enable it, otherwise web will not work for you.

Data Engineer - II @ ClearDemand

Home > Software Development

 Data Engineer - II

Job Description

  • Building on the foundation of the SDE-I role, the SDE-II position at Bungee Tech takes on a greater level of responsibility and leadership. You'll play a crucial role in driving the evolution and efficiency of our data collection and analytics platform, capable of handling terabyte-scale data and billions of data points.

Key Responsibilities:

  • Lead the design, development, and optimization of large-scale data pipelines and infrastructures using technologies like Apache Airflow, Spark, Kafka, and more.
  • Architect and implement distributed data processing solutions to handle terabyte-scale datasets and billions of records efficiently across multi-region cloud infrastructure (AWS, GCP, DO).
  • Develop and maintain real-time data processing solutions for high-volume data collection operations using technologies like Spark Streaming and Kafka.
  • Optimize data storage strategies using technologies such as Amazon S3, HDFS, and Parquet/Avro file formats for efficient querying and cost management.
  • Build and maintain high-quality ETL pipelines, ensuring robust data collection and transformation processes with a focus on scalability and fault tolerance.
  • Collaborate with data analysts, researchers, and cross-functional teams to define and maintain data quality metrics, implement robust data validation, and enforce security best practices.
  • Mentor junior engineers (SDE-I) and foster a collaborative, growth-oriented environment.
  • Participate in technical discussions, contributing to architectural decisions, and proactively identifying improvements for scalability, performance, and cost-efficiency.
  • Ensure application performance monitoring (APM) is in place, utilizing tools like Datadog, New Relic, or similar to proactively monitor and optimize system performance, detect bottlenecks, and ensure system health.
  • Implement effective data partitioning strategies and indexing for performance optimization in distributed databases such as DynamoDB, Cassandra, or HBase.
  • Stay current with advancements in data engineering, orchestration tools, and emerging cloud technologies, continually enhancing the platforms capabilities.

Required Skills:

  • 4+ years of hands-on experience with Apache Airflow and other orchestration tools for managing large-scale workflows and data pipelines.
  • Expertise in AWS technologies, Athena, AWS Glue, DynamoDB, Apache Spark, PySpark, SQL, and NoSQL databases.
  • Experience in designing and managing distributed data processing systems that scale to terabyte and billion-scale datasets using cloud platforms like AWS, GCP, or Digital Ocean.
  • Proficiency in web crawling frameworks, including Node.js, HTTP protocols, Puppeteer, Playwright, and Chromium for large-scale data extraction.
  • Experience with monitoring and observability tools such as Grafana, Prometheus, Elasticsearch, and familiarity with monitoring and optimizing resource utilization in distributed systems.
  • Strong understanding of infrastructure as code using Terraform, automated CI/CD pipelines with Jenkins, and event-driven architecture with Kafka.
  • Experience with data lake architectures and optimizing storage using formats such as Parquet, Avro, or ORC.
  • Strong background in optimizing query performance and data processing frameworks (Spark, Flink, or Hadoop) for efficient data processing at scale.
  • Knowledge of containerization (Docker, Kubernetes) and orchestration for distributed system deployments.
  • Deep experience in designing resilient data systems with a focus on fault tolerance, data replication, and disaster recovery strategies in distributed environments.
  • Strong data engineering skills, including ETL pipeline development, stream processing, and distributed systems.
  • Excellent problem-solving abilities, with a collaborative mindset and strong communication skills.

Why Join Us:
At Bungee Tech, youll be at the forefront of innovation in the data engineering space, working with cutting-edge technologies and a talented team. If you're passionate about building scalable systems, handling large-scale distributed data, and solving complex data challenges, wed love to have you on board.

Job Classification

Industry: Software Product
Functional Area / Department: Engineering - Software & QA
Role Category: Software Development
Role: Data Engineer
Employement Type: Full time

Contact Details:

Company: ClearDemand
Location(s): Chennai

+ View Contactajax loader


Keyskills:   Airflow Big Data Advance Sql AWS

 Fraud Alert to job seekers!

₹ Not Disclosed

Similar positions

Senior Delivery Consultant - Application Architect

  • Amazon
  • 10 - 15 years
  • Hyderabad
  • 14 hours ago
₹ Not Disclosed

System Development Engineer

  • Amazon
  • 0 - 5 years
  • Hyderabad
  • 2 days ago
₹ Not Disclosed

Software Engineer

  • Orange Business
  • 3 - 8 years
  • Noida, Gurugram
  • 2 days ago
₹ Not Disclosed

File Transfer- IBM Sterling Connect Direct Professional

  • Capgemini
  • 4 - 7 years
  • Bengaluru
  • 2 days ago
₹ Not Disclosed

ClearDemand

Company Description: About Clear Demand:  Clear Demand is the leader in AI-driven price and promotion optimization for retailers. Our platform transforms pricing from a challenge to a competitive advantage, helping retailers make smarter, data-backed decisions across the entire pricing lifecy...