Collect, store, process and analyze large datasets to build and implement extract, transfer, load (ETL) processes
Develop reusable frameworks to reduce the development effort involved thereby ensuring cost savings for the projects.
Develop quality code with thought through performance optimizations in place right at the development stage.
Appetite to learn new technologies and be ready to work on new cutting-edge cloud technologies.
Work with team spread across the globe in driving the delivery of projects and recommend development and performance improvements
Expertise in Python and experience writing Azure functions using Python/Node.js
Experience using Event Hub for data integrations.
Hands on expertise in: Implementing analytical data stores in Azure platform using ADLS/Azure data factory /Data Bricks and Cosmos DB (mongo/graph API)
Experience ingesting using Azure data factory, Complex ETL using Data Bricks.
Eagerness to learn new technologies on the fly and ship to production
8+ years of solutions development and delivery experience with 5+ years of recent experience in data engineering.
Proficiency and extensive Experience with Spark & Scala/Python and performance tuning
Hive database management and Performance tuning - Partitioning / Bucketing.
Strong SQL knowledge and data analysis skills for data anomaly detection and data quality assurance.
Very good problem solver and excellent communication skills - both written and verbal