Years of experience in the following Big Data and distributed programming models and technologies and techniques such as Hadoop Spark Map Reduce Sqoop Hive Hadoop Distributed File System HDFS Distributed Indexing and Databases HBase Hive Cassandra Vertica Serialization Formats JSON Avro Parquet Pyspark Spark Scala Hive Tuning Bucketing Partitioning
Knowledge of database structures theories principles and practices both SQL and NoSQL.
Experience on data lakes datahub implementation.
Experience and knowledge working with relational databases RDBMS
Experience and knowledge working in Kafka Spark streaming Sqoop Oozie Airflow ControlM
Experience with Shell Scripting
Bachelors degree in Information Technology Computer Science Engineering or related field or equivalent combination of education and work experience
Well versed with Information and application security including LDAP certificates public key encryption SSH access credentials etc. or more years of experience working in Agile LeanKanban or Scaled Agile organization
Knowledge or experience in Jira Confluence and Bitbucket
Experience applying TDD BDD and Static Code analysis to improve quality and reliability of delivery
Ability to independently perform all duties from Analysis to deployment to postproduction defect fixing
Must have experience working on Production support Preferred
Experience in building Microservices using Java Python SPARK OCP RESTful APIs
Experience in one or more of the following Amazon Web Services AWS Cloud services EC EMR ECS DockerOpenShift KubernetesAmazon EKS S SNS SQS Cloud Formation Cloud watch Lambda