5+ years of experience working with enterprise data platforms, building and managing data lakes and using big data technologies
3+ years of experience working with AWS platform. Experience with solutioning on AWS infrastructure using services like AWS IAM, S3, Lambda, Lake formation, Redshift
2+ years of experience with using Python for data engineering and developing high performance and scalable data pipelines
Prior experience with traditional ETL tools like Talend Open Studio, Informatica, Pentaho or something similar is strongly preferred
Experience with automating and orchestrating jobs on a big data platform using Airflow, Jenkins or something similar
Good understanding and experience working with various products in the big data ecosystem like Spark, EMR, Hive, NoSQL databases like DynamoDB, Cassandra
Prior experience with working in a SQL server based environment and using SSIS, SSRS, TSQL is preferred.
Has to be a team player and open to working with newer technologies as well as supporting legacy systems