The candidate will also be required to have a deep familiarity with the following technologies and its application to business solutions: Erwin, Hadoop, Hive, Impala, SQL, Linux, Azure.
Candidates experienced with Supply Chain will be given preference.
Experience with Scrum and an Agile Development environment is required.
Strong working knowledge of tools like Erwin and Metadata Mapping techniques are required.
5+ years of experience in area of Data Modeling/Management, Automation, and Business Analytics
5+ years in direct experience with designing, modeling and deploying multi-dimensional systems
Experience with Linux, including command-line and Bash scripting, is preferred.
Experience with SQL, Hadoop, Hive and Impala is required.
Experience with table storage and optimizing performance on Data Lake
Experience with tools like JIRA, Confluence, etc . is required.
Hands-on experience with end-end solution delivery of Data & Analytics applications involving structured and un-structured data on Hadoop based platforms, such as Cloudera and AWS EMR, is essential.
Proven ability to work with project managers and stakeholders to help refine and capture software requirements in an iterative fashion. Accurately forecast engineering effort.
A commitment and passion for continuous integration, unit tests, and using automation to ensure stable deliverables.
Extreme attention to detail and commitment to quality.
Self-motivated, driven to achieve committed milestones.
Strong collaboration skills, ability to mentor and be mentored.
Ability to work seamlessly as part of a multi-site, multicultural development team.
Desire to work in a fast paced, growing, deadline driven, startup environment.
Experience documenting best practices and enforcing strong governance in a team is required.
Experience with data lineage, data profiler and metadata management tools and processes is preferred.
Experience integrating Data Lakes with AWS Redshift/Azure/Teradata would be strongly considered.