Must Have:
- Advanced in one or more programming language(s), such as Java, Python
- Hands-on practical experience delivering data pipelines
- Proficient in all aspects of the Software Development Life Cycle
- Advanced understanding of agile methodologies such as CI/CD, Applicant Resiliency, and Security
- Demonstrated proficiency and experience with cloud-native distributed systems
- Ability to develop reports, dashboards, and processes to continuously monitor data quality and integrity
- Working knowledge of bitbucket and JIRA
Preferred qualifications, capabilities, and skills:
- Hands-on experience building data pipelines on AWS using Lambda, SQS, SNS, Athena, Glue, EMR
- Strong experience with distributed computing frameworks such as Apache Spark, specifically PySpark
- Strong hands-on experience building event driven architecture using Kafka
- Experience writing Splunk or Cloudwatch queries, DataDog metrics