Citizen
Full Time
Direct Hire
UG :- - Not Required
PG :- - Not Required
No of position :- ( 1 )
Post :- 27th Aug 2022
As a member of the data-pipeline team building the analytics pipeline you will, work in the areas of orchestration, data modeling, data pipelines and data-lakes, APIs, storage, distribution, distributed computation, consumption and infrastructure management.
Collaborate with product management and other engineering teams to understand and define the analytics requirements for our customers and design and build features, data source integrations, data platform data-pipelines.
Build internal applications to administer, monitor and troubleshoot the data pipelines data-integrations.
Collaborate with cloud-infrastructure team on infrastructure automation, cloud engineering and security design.
Implement technically best practices for optimized data flow and data quality.
Production deployment design and technical support of our large scale data-pipelines.
Create user documentation to maintain both development and operations continuity.
Work together with your agile team to improve process and delivery through collaborative problem-solving.
Cultivate and enhance a culture built around standard methodologies for CI/CD, alerting, and monitoring
What you will bring?
At least 4+ years experience. Computer science graduate from a good college.
Proficient in Python with good knowledge of dealing with data analysis python libraries - Pandas, Numpy
Experience in Apache PySpark Distributed computing platform.
Experience / Familiarity with Database Modeling, data management eco system.
Experience with dev ops tools like Git, Maven, Gradle, Jenkins, Docker etc.
Excellent written and verbal communication skills.
Passion for learning and implementing new technologies.
Ability to operate under fast-paced environment
Must have Skills:
Python, SQL,
Good To have Skills:
Java, AWS Glue Kafka, data-pipeline design patterns, RedShift, Cloud Data Platform, MongoDB or other NoSQl databases, Docker, REST APIs.