US Citizen
Green Card
Corp-Corp
W2-Permanent
W2-Contract
Contract to Hire
Consulting/Contract
UG :- - Not Required
PG :- - Not Required
No of position :- ( 1 )
Post :- 19th Dec 2025
· 10+ years of experience in Data Engineering or a related role.
· Proficiency in AWS – specifically S3, Glue, Athena, and Lambda
· Experience with Hadoop ecosystem (think BIG data), including HDFS, MapReduce, and Hive.
· Hands-on experience with Apache Hudi for real-time data management.
· Strong programming skills in Python and PySpark.
· Familiarity with SQL and NoSQL databases.
· Knowledge of data governance and data security best practices.
· Experience with workflow scheduling tools like AWS Step Functions or Control M.
Day-to-day:
· Design, build, and manage ETL data pipelines using AWS services such as S3, Glue, Athena, and Lambda.
· Implement and manage real-time data streaming and batch processing solutions using Apache Hudi, PySpark, and Hadoop.
· Leverage AWS Athena for complex SQL queries over large datasets.
· Use Python and PySpark to perform data transformation and data cleansing.
· Ensure data quality and integrity by implementing proper data governance strategies.
· Collaborate with cross-functional teams to meet business objectives.
· Monitor performance and advise any necessary infrastructure changes.
· Develop technical documentation including data dictionaries, metadata, and pipeline architecture.
· Troubleshoot data issues and provide ongoing operational support.
Summary:
They will be responsible for providing AWS Cloud solutions using Hadoop.