Required Skills

Data Engineer

Work Authorization

  • US Citizen

  • Green Card

  • H1B Work Permit

Preferred Employment

  • Corp-Corp

  • W2-Permanent

  • W2-Contract

  • Contract to Hire

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 27th Mar 2023

JOB DETAIL

Key Responsibilities:

Translates business requirements and specifications into logical program designs, modules, stable application systems, and data solutions with occasional guidance from senior colleagues; partners with Product Team to understand business needs and functional specifications

Develops, configures, or modifies integrated business and/or enterprise application solutions within various computing environments by designing and coding component-based applications using various programming languages

Conducts the implementation and maintenance of complex business and enterprise data solutions to ensure successful deployment of released applications

Supports systems integration testing (SIT) and user acceptance testing (UAT), provides insight into defining test plans, and ensures quality software deployment

Participates in the end-to-end product lifecycle by applying and sharing an in-depth understanding of company and industry methodologies, policies, standards, and controls

Understands Computer Science and/or Computer Engineering fundamentals; knows software architecture and readily applies this to Data or Platform solutions

Automates and simplifies team development, test, and operations processes; develops conceptual, logical, and physical architectures consisting of one or more viewpoints (business, application, data, and infrastructure) required for business solution delivery

Solves difficult technical problems; solutions are testable, maintainable, and efficient

 

Data Engineering Qualifications:

2 years of experience in Hadoop or any Cloud Bigdata components

Expertise in Java/Scala/Python, SQL, Scripting, Teradata, Hadoop (Sqoop, Hive, Pig, Map Reduce), Spark (Spark Streaming, MLib), Kafka or equivalent Cloud Bigdata components

 

Preferred Qualifications:

Degree in Computer Science, Engineering, Mathematics, Data Science, Analytics, Information System, or related quantitative field.

3 years hand-on experience in building robust, reliable, and scalable machine learning pipelines from ground up as well as transitioning from MVP to production (ingestion, scheduling, security, notifications, validation, backups, optimizations).

Experience with CI/CD pipelines such as Concourse, Jenkins.

Experience with distributed systems (Spark, Hive, HDFS, Hadoop, HBase, Druid, Cloudera, Kafka) along with Airflow to orchestrate data pipeline.

Experience with process optimizations and fine-tuning existing applications in distributed computing environment.

Write quality code in Python/Pyspark followed by unit tests and documentation.

Experience in configuring/building data quality frameworks - PyDeequ, great expectations or any other tools to increase reliability.

Define, execute, and operate monitoring and alerting steps over critical SLA's - (Prometheus/Nagios, Grafana).

Out-of-the box thinker and deep diver to provide efficient solutions to problems

Agile development skills and experience.

Data Science Exposure.

Company Information