At least 8 years of experience on designing and developing production data pipelines for data ingestion or transformation using tools from Hadoop stack (HDFS, Hive, Spark (PySpark/Scala), HBase, Kafka, NiFi, Oozie, Apache Beam, Apache Airflow etc.).
At least 4 years of experience in the following Big Data Platforms: Cloudera / Horton Works / Snowflake / AWS EMR / RedShift / AWS Glue.
At least 4 years of developing applications with Monitoring, Build Tools, Version Control, Unit Test, TDD, Change Management to support DevOps
At least 2 years of experience with SQL and Shell Scripting experience
Experience troubleshooting JVM-related issues.
Familiarity with machine learning implementation using Spark ML / Tensorflow.
Experience in data visualization tools like Cognos, Arcadia, Tableau.
Experience in data warehousing modeling techniques.