Required Skills

ETL

Work Authorization

  • US Citizen

  • Green Card

  • EAD (OPT/CPT/GC/H4)

  • H1B Work Permit

Preferred Employment

  • Corp-Corp

  • W2-Permanent

  • W2-Contract

  • Contract to Hire

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 17th Jan 2024

JOB DETAIL

Responsibilities: 

  • Hands-on building of ETL pipelines using our internal framework written in Java and Python 
  • Hands-on solutioning of real time REST APIs or other solutions for streaming data from Graph 
  • Modify existing application code or interfaces or build new application components from detailed requirements. 
  • Analysis of requirements, support of the design, development of the code, testing, debugging, deployment, and maintenance of those programs and interfaces. Documentation of the work is essential 
  • Participation in most aspects of programming and application development, including file design, update, storage, and retrieval 
  • Enhance processes to resolve operational problems and add new functions taking into consideration schedule, resource constraints, process complexity, dependencies, assumptions and application structure 
  • Ability to maintain the developed solution on an on-going basis is essential 
  • Ability to follow the existing development methodology and coding standards, and ensure compliance with the internal and external regulatory requirements 
  • Develop and implement databases, data collection systems, data analytics and other strategies that optimize statistical efficiency and quality 
  • Acquire data from primary or secondary data sources and maintain databases/data systems 
  • Work with management to prioritize business and information needs 
  • Locate and define new process improvement opportunities 
  • Document design and data flow for existing and new applications being built.  
  • Co-ordinate with multiple different teams QA, Operations and other development team within the organization.  
  • Testing methods, including unit and integration testing (PyTest, PyUnit) 
  • Ability to integrate with large teams, demonstrating strong verbal and written communication skills 
  • Utilization of software configuration management tools 
  • Code deployment and code versioning tools 
  • Excellent Communication Skills 

Qualifications: 

  • Bachelor’s degree preferably with Computer Science background. 
  • Lead team of 4 members and more
  • At least8+ years of experience implementing complex ETL pipelines preferably with Spark toolset. 
  • At least 8+ years of experience with Python particularly within the data space 
  • Technical expertise regarding data models, database design development, data mining and segmentation techniques 
  • Good experience writing complex SQL and ETL processes 
  • Excellent coding and design skills, particularly either in Scala or Python. 
  • Strong practical working experience with Unix scripting in at least one of Python, Perl, Shell (either bash or zsh). 
  • Experience in AWS technologies such as EC2, Redshift, Cloud formation, EMR, AWS S3, AWS Analytics required. 
  • Experience designing and implementing data pipelines in a onprem/cloud environment is required. 
  • Experience building/implementing data pipelines using Databricks/On prem or similar cloud database. 
  • Expert level knowledge of using SQL to write complex, highly optimized queries across large volumes of data. 
  • Hands-on object-oriented programming experience using Python is required. 
  • Professional work experience building real-time data streams using Spark and Experience in Spark. 
  • Knowledge or experience in architectural best practices in building data lakes 
  • Develop and work with APIs 
  • Develop and maintain scalable data pipelines and build out new API integrations to support continuing increases in data volume and complexity. 
  • Collaborate with analytics and business teams to improve data models that feed business intelligence tools, increase data accessibility, and foster data-driven decision making across the organization. 
  • Implement processes and systems to monitor data quality, to ensure production data accuracy, and ensure key stakeholder and business process access. 
  • Write unit/integration tests, contribute to engineering wiki, and documents. 
  • Perform data analysis required to troubleshoot data related issues and assist in the resolution of data issues. 
  • Experience developing data integrations and data quality framework based on established requirements. 
  • Experience with CI/CD processes and tools (e.g., concourse, Jenkins). 
  • Experience with test driven development writing unit tests, test coverage using PyTest, PyUnit, pytest-cov libraries. 
  • Experience working in an Agile environment. 
  • Good understanding & usage of algorithms and data structures 
  • Good Experience building reusable frameworks. 
  • Experience working in an Agile Team environment. 
  • AWS certification is preferable: AWS Developer/Architect/DevOps/Big Data 
  • Excellent communication skills both verbal and written 

 

Company Information