Required Skills

ETL Data stage Python SQL Snowflake/DBT RDBMS

Work Authorization

  • Us Citizen

  • Green Card

  • EAD (OPT/CPT/GC/H4)

  • H1B Work Permit

Preferred Employment

  • Corp-Corp

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 17th Dec 2020

JOB DETAIL

ETL Data stage, Python, SQL, Snowflake/DBT preferred but we are okay to train, if they are ready to go for certification in 3 months on Snowflake.

 

  • 5+ years of strong (8.5 out of 10) ETL experience on either DataStage, Informatica, Ab-Initio, Talend etc. (Datastage is preferred)
  • Experience in designing solutions for multiple large data warehouses with a good understanding of cluster and parallel architecture as well as high-scale or distributed RDBMS
  • Excellent Proficiency in SQL coding, Strong database fundamentals including SQL, performance and schema design. (8.5 out of 10)
  • Should have hands-on experience in Python scripting.  (7 out of 10)
  • Ability to interpret/write custom shell scripts. (7 out of 10)
  • Experience with AWS platform and Snowflake CDW platform big plus. DBT knowledge will be a big plus too.
  • Experience with Git.
  • To be able to work in a fast-paced agile development environment.
  • Design, develop & implement ETL/ELT processes for Cloud Datawarehouse platform Snowflake
  • Ability to understand legacy DataStage ETL components and rewrite them into SQLs and DBT (Data Build Tool) components
  • Work closely with other data engineering teams to ensure alignment of methodologies and best practice

 

Key skills: Python, spark, pyspark, sql, ETL informatica / Abinitio

•             Bachelor's Degree or master’s degree in Computer Science, Mathematics, Statistics.

•             4+ years of development experience in using Spark to build applications through Python and PySpark

•             3+ years’ hands-on experience developing optimized, complex SQL queries and writing PLSQL code across large volumes of data in both relational and multi-dimensional data sources such as Teradata, Hive, Impala, Oracle, Teradata

•             Experience in deploying and developing application using Azure

•             Experience working with disparate data-sets in multiple formats like JSON, Avro, text files, Kafka queues, and log data and storage like blob/ADLS GEN2

•             2+ years of strong ETL experience on either Informatica, Ab-Initio, Talend, DataStage, Syncsort

•             Enthusiastic to work on disparate datasets in multiple formats like JSON, Avro, text files, Kafka queues, and Knowledge of software design and programming principles.

•             Experience working in Scrum Agile framework and using DevOps to deploy and manage code.

•             Good communication and team-working skills.

Company Information