Key skills: Python, spark, pyspark, sql, ETL informatica / Abinitio
- 4+ years of development experience in using Spark to build applications through Python and PySpark
- 3+ years’ hands-on experience developing optimized, complex SQL queries and writing PLSQL code across large volumes of data in both relational and multi-dimensional data sources such as Teradata, Hive, Impala, Oracle, Teradata
- Experience in deploying and developing application using Azure
- Experience working with disparate data-sets in multiple formats like JSON, Avro, text files, Kafka queues, and log data and storage like blob/ADLS GEN2
- 2+ years of strong ETL experience on either Informatica, Ab-Initio, Talend, DataStage, Syncsort
- Enthusiastic to work on disparate datasets in multiple formats like JSON, Avro, text files, Kafka queues, and Knowledge of software design and programming principles.
- Experience working in Scrum Agile framework and using DevOps to deploy and manage code.
- Good communication and team-working skills.