Required Skills

Core Java Python

Work Authorization

  • US Citizen

  • Green Card

  • EAD (OPT/CPT/GC/H4)

  • H1B Work Permit

Preferred Employment

  • Corp-Corp

  • W2-Permanent

  • W2-Contract

  • Contract to Hire

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 13th Oct 2023

JOB DETAIL

Roles and Responsibilities:

Develop and maintain data pipelines using Big Data processes.

Focus on ingesting, storing, processing, and analysing large datasets.

Work closely with data science team to implement data analytics pipelines.

Help define data governance policies and support data-versioning processes.

Maintain security and data privacy, working closely with data protection officer.

Analyse vast number of data stores to uncover insights.

Enhancing data collection procedures to include all relevant information for developing analytic systems.

Processing, cleansing, and validating the integrity of data to be used for analysis.

Analysing large amounts of information to find patterns and solutions.

Developing prediction systems and machine learning algorithms

Presenting results in a clear manner.

Propose solutions and strategies to tackle business challenges.

Collaborate with Business and IT teams.

 

Required skills and qualifications:

Bache Degree in Engineering or Computer Science or equivalent OR master’s in computer applications or equivalent.  

10+ years of software development experience and leading teams of engineers and scrum teams

5+ years of hands-on experience of working with Informatica, Oracle

Solid Data warehousing concepts   Knowledge of Financial reporting ecosystem will be a plus.

5+ years of experience within Data Engineering/ Data Warehousing using Big Data technologies will be a addon.  

Expert on Distributed ecosystem 

Hands-on experience with programming using Core Java or Python/Scala 

Expert on Hadoop and Spark Architecture and its working principle 

Hands-on experience on writing and understanding complex SQL(Hive/PySpark-dataframes),

optimizing joins while processing huge amount of data

Experience in UNIX shell scripting   Ability to design and develop optimized Data pipelines for batch and real time data processing.

Should have experience in analysis, design, development, testing, and implementation of system applications.

Demonstrated ability to develop and document technical and functional specifications and analyse software and system processing flows.

 

Company Information