Required Skills

Pyspark SQL ETL Python Kafka Aws Glue Numpy NoSQL Docker Pandas Cloud API MongoDB Spark Restful Webservices AWS

Work Authorization

  • Citizen

Preferred Employment

  • Full Time

Employment Type

  • Direct Hire

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 27th Aug 2022

JOB DETAIL

As a member of the data-pipeline team building the analytics pipeline you will, work in the areas of orchestration, data modeling, data pipelines and data-lakes, APIs, storage, distribution, distributed computation, consumption and infrastructure management.

Collaborate with product management and other engineering teams to understand and define the analytics requirements for our customers and design and build features, data source integrations, data platform data-pipelines.

Build internal applications to administer, monitor and troubleshoot the data pipelines data-integrations.

Collaborate with cloud-infrastructure team on infrastructure automation, cloud engineering and security design.

Implement technically best practices for optimized data flow and data quality.

Production deployment design and technical support of our large scale data-pipelines.

Create user documentation to maintain both development and operations continuity.

Work together with your agile team to improve process and delivery through collaborative problem-solving.

Cultivate and enhance a culture built around standard methodologies for CI/CD, alerting, and monitoring

What you will bring?

At least 4+ years experience. Computer science graduate from a good college.

Proficient in Python with good knowledge of dealing with data analysis python libraries - Pandas, Numpy

Experience in Apache PySpark Distributed computing platform.

Experience / Familiarity with Database Modeling, data management eco system.

Experience with dev ops tools like Git, Maven, Gradle, Jenkins, Docker etc.

Excellent written and verbal communication skills.

Passion for learning and implementing new technologies.

Ability to operate under fast-paced environment

Must have Skills:

Python, SQL,

Good To have Skills:

Java, AWS Glue Kafka, data-pipeline design patterns, RedShift, Cloud Data Platform, MongoDB or other NoSQl databases, Docker, REST APIs.

Company Information