-
UG :- - Not Required
-
PG :- - Not Required
-
No of position :- ( 1 )
-
Post :- 16th Jul 2022
- Hands-on experience in Scala/Big Data technologies (Spark/Flink, Hive, Oozie, Hbase/Cassandra/MongoDB, Redis, YARN, Kafka)
- Experience with Java/Scala programming language
- Experience with shell scripting
- Experience with ETL job design using Nifi/Streamsets/Talend/Pentaho/etc.
- Working experience on multiple Hadoop/Spark based projects which have been deployed on Production.
- Knowledge of Lambda/Kappa architectures
- Experience on performance tuning of Hive and Spark jobs
- Basic experience in data modelling for Hive and NoSQL (partitioning, bucketing, row key design, etc.)
- Experience in debugging/troubleshooting Hadoop and Spark jobs
- Maven build
- Deployment of Hadoop/Spark jobs on production
Good to have:
- Experience with migration of data from datawarehouse to Hadoop based Data Lake
- Knowledge of Datawarehousing concepts such as Facts, Dimensions, SCD, Star schema, etc.
- Experience or at least good conceptual knowledge of Docker, Kubernetes, Microservices.
- Experience in working on a Hadoop/Spark cluster on a Cloud platform (Google/Amazon/IBM cloud)