Excellent experience in using Spark to build applications through Python and PySpark
Excellent Proficiency in SQL coding, Strong database fundamentals including SQL, performance and schema design
Excellent experience developing optimized, complex SQL queries and writing PLSQL code across large volumes of data in both relational and multi-dimensional data sources such as Teradata, Hive, Impala, Oracle, Teradata
ETL experience on either DataStage, Informatica, Ab-Initio, Talend etc
Experience in deploying and developing application using Azure
Experience working with disparate data-sets in multiple formats like JSON, Avro, text files, Kafka queues, and log data and storage like blob/ADLS GEN2
Experience working in Scrum Agile framework and using DevOps to deploy and manage code.