Required Skills

Data Engineer

Work Authorization

  • Us Citizen

  • Green Card

  • EAD (OPT/CPT/GC/H4)

  • H1B Work Permit

Preferred Employment

  • Corp-Corp

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 8th Oct 2021

JOB DETAIL

1. Databricks
2. Pyspark
3. ETL

Detailed Job Description:
· Must have Azure Cloud experience and understand Azure components like ADF, Azure SQL and Azure DataBricks.
· Very Strong DataBricks, Spark, Pyspark, DataBricks SQL with Azure
· Must have Strong ETL and ELT experience
· Must have strong Python, DataBricks SQL skills beyond just calling Spark Api, must be fluent in Python programming language
· Must have Relational Database knowledge for optimum loading of Data from on premise system and data lake
· Must have experience with Data Lake and Data Bases
· Must have knowledge of OOP and functional programming to create reusable framework for ETL
· Must understand encryption and security required for PII, Financial and other sensitive data.
· Must understand Delta Lake and other big data file formats
· Good to have exposure to DevOps and CICD skills in Big Data Space
· Good to have Airflow or AppWorx.
· Good to have exposure to Manufacturing Domain
· Good to have SQL as well as No SQL databases
· Good to have AD security experience

· Work closely with Architect and Business units in understanding technical requirement and implement independently reusable code.
· Develop ETL Framework for template driven ETL
· Develop Databricks code that can call Scala and other required libraries
· Work with offshore and on shore teams and mentor team members on ETL and do KT on framework and design
· Implement transformations and aggregations as requirement
· Work in Agile manner and resolve ambiguous requirements and communicate effectively with his peers

Minimum years of experience*:>10 years

Top 3 responsibilities you would expect the Subcon to shoulder and execute*:
1. Work closely with Architect and Business units in understanding technical requirement and implement independently reusable code
2. Develop ETL Framework for template driven ETLDevelop Databricks code that can call Scala and other required libraries
3. Develop Databricks code that can call Python and other required libraries
 

Company Information