Required Skills

Hadoop applications Big Data ecosystem. queries script ETLs automated processes Data Integration EDW tools. Hadoop Impala Oracle Terabytes of data.

Work Authorization

  • Us Citizen

  • Green Card

  • EAD (OPT/CPT/GC/H4)

Preferred Employment

  • Corp-Corp

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 13th Nov 2020

JOB DETAIL

Job Title :Hadoop Developer

Location : Bellevue,WA

Duration :Contract

Mandatory:
Responsible for Development, implementation & maintenance of various Hadoop applications / data feeds in Big Data ecosystem.

Responsible for integrating & providing development support including code check-in, library updates and management, and deployment control for updated or modified jobs.

Troubleshoot any failures and ensure the job completes. Modifying and troubleshooting queries, script and ETLs for the supported feeds inline with the runbook.

Maintain the parser and automated processes. Perform daily check to verify the integrity. 

Review system and application logs, and verifying completion of scheduled jobs & optimize the performance.

Manage the ingestion and transformation of data feeds into production cluster.

Coordinate major code level issues and changes with development and Quality Assurance teams.

Monitor & Manage Scheduler.

Re responsible for Shell Scripting programming, Java, EDW platforms with Knowledge around any Data Integration and/or EDW tools.

Perform daily system monitoring (alarms/KPI’s), verifying the integrity and availability of all hardware, server resources, systems and key processes, reviewing system and application logs, and verifying completion of scheduled jobs such as backups.

Manage & optimize Daily Aggregation jobs. Troubleshoot in case of any issues / delays in order to ensure on time delivery.

Any major issues with the Code, Modifications / change in Source data type warrants rewriting the code.
Monitor & Manage Scheduler.


Skills :

Experience in developing and supporting Java, Python, Storm, Spark and Map Reduce, Tez applications in a distributed platform

Experience in working MPP systems like Hadoop, Impala, Teradata and Oracle and analyzing Hundreds of Terabytes of data.

Experience in implementing with real-time data ingestion frameworks with Spark, Kafka and Flume

Should have complete hands-on experience on Cassandra & Scala.

Experience in building and implementing data pipelines with Hadoop ecosystem tools to capture near real-time/batch processing pipelines.

Hands on experience in working with Hortonworks and Cloudera Enterprise Hadoop distributions.

Experience in processing and moving large structured and/or unstructured and semi-structured data into data lake or out of data Lake.

Working experience in Jenkins, CI/CD pipeline is an added advantage.


Krishna Kanth

IDC Technologies

Ph: 408-385-2661

Email: krishna.k@idctechnologies.com 

Company Information