US Citizen
Green Card
EAD (OPT/CPT/GC/H4)
H1B Work Permit
Corp-Corp
W2-Permanent
W2-Contract
Contract to Hire
Consulting/Contract
UG :- - Not Required
PG :- - Not Required
No of position :- ( 1 )
Post :- 2nd May 2023
Role: Data engineer:
Fundamental Components:
Develops large scale data structures, pipelines and efficient ETL (extract/load/transform) workflows to organize, collect and standardize data that helps generate insights and addresses reporting needs.
Collaborates with other data teams to transform data and integrate algorithms and models into automated processes.
Uses knowledge in Hadoop architecture, HDFS commands and experience designing & optimizing queries to build data pipelines.
Builds data marts and data models to support Data Science and other internal customers.
Analyzes current information technology environments to identify and assess critical capabilities and recommend solutions.
Experiments with available tools and advises on new tools in order to determine optimal solution given the requirements dictated by the model/use cases
BACKGROUND/EXPERIENCE desired:
5 or more years of progressively complex related experience.
Has strong knowledge of large scale search applications and building high volume data pipelines.
Experience building data transformation and processing solutions.
Knowledge in Hadoop architecture, HDFS commands and experience designing & optimizing queries against data in the HDFS environment.
Ability to understand complex systems and solve challenging analytical problems.
Experience and proficiency with cloud platforms and their architectures, preferably GCP.
Experience with Google Cloud services like, Dataproc, BigQuery, etc.
Ability to leverage multiple tools and programming languages to analyze and manipulate data sets from disparate data sources.
Experience with Apache NiFi ,Rest API and Apache Spark required.
Strong collaboration and communication skills within and across teams.
Must possess a well-recognized GCP Architect certificate or GCP Developer certificate
Strong problem solving skills and critical thinking ability.
Experience working with a DevOps model.
Vast knowledge of GCP Services.
SKILL SET desired:
Data Warehouse
GCP: Dataproc, Cloud Composer, Dataflow, BigQuery, CloudSQL.
Spark
Shell Scripting
Unix
Hadoop Concepts (Sqoop, YARN, MapReduce, Hive ,etc.)
Python, Scala and Java
Infoworks