AWS Cloud Computing with knowledge of service like AWS Lambda, EC2, S3, EMR, Redshift, OpenSearch, Glue, Step Functions
Must have Hands on Experience on Spark with Python/Scala
Must have experience in design and build production data pipelines from ingestion to consumption within a big data architecture, using Java, Python, Scala.
Implemented complex projects dealing with the considerable data size (TB/ PB) and with high complexity in the production environment.
HANDS-ON experience on Hadoop tools/technologies like Map Reduce, Hive, HDFS.
Demonstrated strength in data modelling, ETL development
Experience in designing and implementing an enterprise data lake.
Experience with Iaac, CI/CD, Agile methodologies is an added advantage
Knowledge of AWS security services such (IAM, Shield, WAF) is an added plus