Required Skills

Data Engineer

Work Authorization

  • US Citizen

  • Green Card

  • EAD (OPT/CPT/GC/H4)

  • H1B Work Permit

Preferred Employment

  • Corp-Corp

  • W2-Permanent

  • W2-Contract

  • Contract to Hire

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 2nd May 2023

JOB DETAIL

Role: Data engineer: 

 

Fundamental Components: 

 

Develops large scale data structures, pipelines and efficient ETL (extract/load/transform) workflows to organize, collect and standardize data that helps generate insights and addresses reporting needs. 

 

Collaborates with other data teams to transform data and integrate algorithms and models into automated processes. 

 

Uses knowledge in Hadoop architecture, HDFS commands and experience designing & optimizing queries to build data pipelines. 

 

Builds data marts and data models to support Data Science and other internal customers. 

 

Analyzes current information technology environments to identify and assess critical capabilities and recommend solutions. 

 

Experiments with available tools and advises on new tools in order to determine optimal solution given the requirements dictated by the model/use cases 

 

BACKGROUND/EXPERIENCE desired: 

 

5 or more years of progressively complex related experience. 

 

Has strong knowledge of large scale search applications and building high volume data pipelines. 

 

Experience building data transformation and processing solutions. 

 

Knowledge in Hadoop architecture, HDFS commands and experience designing & optimizing queries against data in the HDFS environment. 

 

Ability to understand complex systems and solve challenging analytical problems. 

 

Experience and proficiency with cloud platforms and their architectures, preferably GCP. 

 

Experience with Google Cloud services like, Dataproc, BigQuery, etc. 

 

Ability to leverage multiple tools and programming languages to analyze and manipulate data sets from disparate data sources. 

 

Experience with Apache NiFi ,Rest API and Apache Spark required. 

 

Strong collaboration and communication skills within and across teams. 

 

Must possess a well-recognized GCP Architect certificate or GCP Developer certificate 

 

Strong problem solving skills and critical thinking ability.   

 

Experience working with a DevOps model.  

 

Vast knowledge of GCP Services. 

 

SKILL SET desired: 

 

Data Warehouse 

 

GCP: Dataproc, Cloud Composer, Dataflow, BigQuery, CloudSQL.  

 

Spark 

 

Shell Scripting 

 

Unix 

 

Hadoop Concepts (Sqoop, YARN, MapReduce, Hive ,etc.) 

 

Python, Scala and Java 

 

Infoworks 

Company Information