Role & responsibilities
Preferred candidate profile
Bachelor's and/or masters degree in computer science or equivalent experience.
Deep understanding of Star and Snowflake dimensional modelling.
Strong knowledge of Data Management principles
Good understanding of Databricks Data & AI platform and Databricks Delta Lake Architecture
Should have hands-on experience in SQL, Python and Spark (PySpark)
Candidate must have experience in AWS/ Azure stack
Desirable to have ETL with batch and streaming (Kinesis).
Experience in building ETL / data warehouse transformation processes
Experience with Apache Kafka for use with streaming data / event-based data
Experience with other Open-Source big data products Hadoop (incl. Hive, Pig, Impala)
Experience with Open Source non-relational / NoSQL data repositories (incl. MongoDB, Cassandra, Neo4J)
Experience working with structured and unstructured data including imaging & geospatial data. Experience working in a Dev/Ops environment with tools such as Terraform, CircleCI, GIT.
Proficiency in RDBMS, complex SQL, PL/SQL, Unix Shell Scripting, performance tuning and troubleshoot
Databricks Certified Data Engineer Associate/Professional Certification (Desirable).
Comfortable working in a dynamic, fast-paced, innovative environment with several ongoing concurrent projects
Should have experience working in Agile methodology
Strong verbal and written communication skills.
Strong analytical and problem-solving skills with a high attention to detail
Keyskills: Azure Data Factory Datafactory Databricks Engineer Data Bricks