Required Skills

AWS PySpark Python

Work Authorization

  • US Citizen

  • Green Card

  • EAD (OPT/CPT/GC/H4)

  • H1B Work Permit

Preferred Employment

  • Corp-Corp

  • W2-Permanent

  • W2-Contract

  • Contract to Hire

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 1st Mar 2024

JOB DETAIL

Responsibilities / Qualifications:

  • Candidate must have 10+ years of IT working experience with at least 5 years of experience in the AWS Cloud environment preferred
  • Experience and knowledge of AWS Cloud Architectures, Databases, Data models, and Data integration patterns.
  • Conduct data profiling activities, discover data quality challenges, document the impacts on business and make recommendations for data quality improvements
  • Expertise in defining new architectures and ability to drive project from architecture standpoint with AWS Cloud platform.
  • Experience with Data ingestion & transformation framework development and enhancements
  • Define and design Data pipeline architecture for data ingestion processes using AWS native services.
  • Design and optimize AWS Data platform using AWS services such as Redshift, RDS, S3, Glue ETL, Glue Data Catalog, EMR, PySpark, Python, Lake formation, etc
  • Experience with building data analytical platform using Databricks (data pipelines), Starburst (semantic layer) on AWS cloud environment
  • Collaborate with different teams to ensure quality and compliance to the Enterprise Data Architecture by participating in data analysis/design activities and conducting appropriate technical data design reviews at various stages during the development life cycle. This includes providing data modelling expertise with both relational and dimensional modelling techniques.
  • Guide, educate, and mentor the Data Architecture Strategy directives, principles, & standards to individuals who play data-related roles (e.g., data analysts, ETL developers, report developers & business analysts).
  • Establish and maintain data standards, policies, and architectures.
  • Provide direction, guidance, and oversight for data quality controls.
  • Capture and maintain metadata, creating business rules for the use of data
  • Design and optimize data governance framework including the management of data, operating model, data policies and standards.
  • Participate in client design workshops and provide trade-offs and recommendations towards building solutions.
  • Collaborating with teams to resolve technical requirements, feasibility, and expectations.
  • Mentor Data engineers in coding best practices and problem solving.
  • Experience working with source code management tools such as GitHub
  • Experience working with DevOps, Jenkins or any CI/CD Pipelines using AWS Services
  • Working experience with Agile Methodology

 

Company Information