experience of Responsible for the integration of large, structured and unstructured data volumes into the existing platforms
Build out scalable and reliable ETL pipelines and processes to ingest data from a large number and variety of data sources
Experience with modern big data technologies like Hadoop, MapReduce, Kafka , Hive, Spark Streaming, Apache Nifi, etc. are must to have.
Experienced with ETL / ELT scripting and applications, preferably using PySpark
Experience working with Big Data ecosystems & Spark architecture and building data-intensive applications and pipelines.
Strong Experience with programming languages in Spark, Python, SQL & Unix shell script
Highest Qualification:
Regular/Part Time:
Total years of Exp:
Relevant Years of Exp:
Current Organization:
Current Location
Prefer Location:
Current CTC:
Expected CTC:
Notice Period:
Holding offer from other Organization