Required Skills

Python AWS

Work Authorization

  • US Citizen

  • Green Card

Preferred Employment

  • Corp-Corp

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 30th Sep 2022

JOB DETAIL


Must have:  Python / Lambda

                   AWS

                   Kafka and Kafka streaming 

 

Responsibilities: Works closely with Analytics and Marketing users to understand informational needs and business challenges, document those requirements, and translate into solutions. Partners with the work stream leads to ensure overall cost, delivery, and quality targets are met. Reviews functional and business specifications from the team and defines detailed ETL technical specifications. 

Utilizes the ETL application to analyze data, necessary to resolve issues uncovered and related to source data, ensuring technical specifications are thorough. Defines, develops, documents and maintains Informatica ETL mappings and scripts. 

Creates and deploys Informatica work flows in Linux environments. Defines, develops, documents and maintains procedural and SQL programs. Assists in the development, execution, and documentation of system and integration test plans. 

Performs ETL / Informatica tuning and SQL tuning. Optimizes Informatica code performance and processing methods. Performs design validation, reconciliation and error handling in load and extract processes (initial and incremental). Support Production batch environment. 

Qualifications: Knowledge & Experience: 3+ years of Informatica Power Center software experience is Strong SQL Knowledge

Strong understanding of using ETL tools to integrate systems. Knowledge of System Development Life Cycle (SDLC). 3+ years working with DB2 or equivalent RDBMS is preferred. Informatica Power Center 9.6 experience preferred. Java development experience preferred. Shell scripting experience preferred. Autosys experience preferred. Personal Attributes: Strong analytical and problem solving skills. Strong quality assurance and accuracy focus. Ability to effectively prioritize and execute tasks in time-sensitive situations. Excellent organizational skills with the ability to give direction and offer process education to others. Excellent verbal and written communication skills. Experience working in a Global, team-oriented, collaborative environment.

 

Project Details: 2 to 5+ years of experience within the field of data engineering or related technical work including business intelligence, analytics
Experience designing and building scalable and robust data pipelines to enable data-driven decisions for the business
Very good understanding of Data warehousing concepts and approaches
Hands-on experience building complex business logics and ETL workflows using Informatica IICS and PowerCenter.
Good Proficient in SQL, PL/SQL and preferably experience in Snowflake
Good Experience in one of the scripting languages: Python or Unix Scripting
Experience in data cleansing, data validation and data wrangling
Hands-on experience in AWS cloud and AWS native technologies such as Glue, Lambda, Kinesis, Lake Formation, S3, Redshift
Experience using Spark EMR, RDS, EC2, Athena, API capabilities, CloudWatch, CloudTrail is a plus
Experience with Business Intelligence tools like Tableau, Cognos, ThoughtSpot.

Company Information