Required Skills

ETL PythonData Engineer

Work Authorization

  • Us Citizen

  • Green Card

  • EAD (OPT/CPT/GC/H4)

  • H1B Work Permit

Preferred Employment

  • Corp-Corp

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 16th Oct 2021

JOB DETAIL

thefocus is on ETL processes andPython. Likely need to be from an AWS/GCPenvironment, 5 years of experience.

MustHaves:

-         Advanced Python skills

-         Mid-to-Advanced SQL Skills

-         Previous experience with “bigdata”

-         At least some exposures toExtendable Analytics Data Architecture (DataLakes or Warehouses).

Notes:

Developedandimplemented a data pipeline that has become an integral part ofsuccessfullydelivering seeding rate prescriptions to customers. 

o  Developed and maintained custom python packages fordataprocessing and ETL. 

o  Built ETL pipelines to query data from different sources including databases(psycopg2), APIs (Requests), anddatawares using python, SQL,Apache Spark and storing them in various locationincluding Postgres Database, AWSS3, Redshift(using boto3) and HIVE dataware. 

o  Builtand maintained custom ETL workflowsorchestrations in Dominoand Airflow 

?    Managed thedesign anddelivery of grower reports generated from insights derived from analysisofdata from field experiments. 

?    Assembled digital data layers for different sources and created geospatial exploratory data analysis for datascientistsusing Apache Spark. 

?    Lead contractors on complex efforts, managed sprints and provided technical support on strategy and implementation. 

•Collaboratedwithdata scientists and other key stakeholders to understand dataneedsand clearly define their requirements to deliver datapipelinesolutions at scale across teams. 

Minimum Qualifications:

?       Bachelorof Science (B.S.) from a 4-year collegeor university in Computer Science orequivalent experience

?       2years of experience withRelational databases or Cloud datawarehousing or analytics systems suchasMySQL/PostgreSQL/Redshift/BigQuery/Snowflake/Elasticsearch.

?       2 years of experience in designingand developing data pipelines.

?       2 year experience writing SQLqueries and code using Python/Node.js/Java.

?       Experience with cloud servicessuch as AWS/GCP/Azure.

?       Excellent verbal and writtencommunications skills along with the abilityto present technical data andapproaches to both technical and nontechnicalaudiences

Preferred Qualifications:

?       3years of experience working with relationaland unstructured databases andenterprise data warehouses, such as work withMySQL, PostgreSQL, SQL Server,Oracle and MongoDB, Elasticsearch

?       Workingexperience in a Cloud environment (AWSor GCP) with a subset of the followingtools or their equivalent - RDS, S3,EC2, Lambda, Kinesis, EMR, BigQuery, GCS,and  BigTable

?       3 years of experience with withdata architecture, data modeling,schema design and/or software development

?       Experiencebuilding open source data pipelinesystems (e.g. Kafka, Apache AirFlow, ApacheBeam and Apache Spark)

Company Information