Required Skills

Big Data Developer

Work Authorization

  • US Citizen

  • Green Card

Preferred Employment

  • Corp-Corp

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 16th Jul 2022


  • Hands-on experience in Scala/Big Data technologies (Spark/Flink, Hive, Oozie, Hbase/Cassandra/MongoDB, Redis, YARN, Kafka)
  • Experience with Java/Scala programming language
  • Experience with shell scripting
  • Experience with ETL job design using Nifi/Streamsets/Talend/Pentaho/etc.
  • Working experience on multiple Hadoop/Spark based projects which have been deployed on Production.
  • Knowledge of Lambda/Kappa architectures
  • Experience on performance tuning of Hive and Spark jobs
  • Basic experience in data modelling for Hive and NoSQL (partitioning, bucketing, row key design, etc.)
  • Experience in debugging/troubleshooting Hadoop and Spark jobs
  • Maven build
  • Deployment of Hadoop/Spark jobs on production

Good to have:

  • Experience with migration of data from datawarehouse to Hadoop based Data Lake
  • Knowledge of Datawarehousing concepts such as Facts, Dimensions, SCD, Star schema, etc.
  • Experience or at least good conceptual knowledge of Docker, Kubernetes, Microservices.
  • Experience in working on a Hadoop/Spark cluster on a Cloud platform (Google/Amazon/IBM cloud)

Company Information