Required Skills

Hadoop Administration AWS hadoop platform admin

Work Authorization

  • Citizen

Preferred Employment

  • Full Time

Employment Type

  • Direct Hire

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 17th Aug 2022

JOB DETAIL

https://jobs.lever.co/pythian/ae48d343-5aa4-44ff-9a9c-db2af320a194

Why you?

Are you an India-based Infrastructure Engineer? Do you thrive on solving tough problems under pressure? Are you motivated by fast-paced environments with continuous learning opportunities? Do you enjoy collaborating with a team of peers who push you to constantly up your game?

At Pythian, we are building a Site Reliability Engineering team that is focused on Streaming and Big Data service operations and open source, cloud-enabled infrastructure architecture. We need motivated and talented individuals on our teams, and we want you!

You'll act as a technology leader and advisor for our clients, as well as a mentor for other team members. Projects would include things such as Kafka and Hadoop deployment, upgrade, disaster planning, system and ecosystem tuning, infrastructure architecture, performance analysis, deployment automation, and intelligent monitoring.

You will work with amazing clients from small, high-velocity startups to large enterprises with complex, hybrid infrastructures and large data processing requirements.

What will you be doing?

  • Deploy, operate, maintain, secure and administer solutions that contribute to the operational efficiency, availability, performance and visibility of our customers infrastructure and Big Data platform and related services, such as core Hadoop and streaming services such as Kafka and RabbitMQ.
  • Gather information and provide performance and root cause analytics and remediation planning for faults, errors, configuration warnings and bottlenecks within our customers infrastructure, applications and Big Data ecosystems.
  • Deliver well-constructed, explanatory technical documentation for architectures that we develop, and plan service integration, deployment automation and configuration management to business requirements within the infrastructure and Big Data ecosystem.
  • Understand distributed Java container applications, their tuning, monitoring and management; such as logging configuration, garbage collection and heap size tuning, JMX metric collection and general parameter-based Java tuning.
  • Observe and provide feedback on the current state of the clients infrastructure, and identify opportunities to improve resiliency, reduce the occurrence of incidents and automate repetitive administrative and operational tasks.
  • Contribute heavily to the development of deployment automation artifacts, such as images, recipes, playbooks, templates, configuration scripts and other open source tooling.
  • Be conversant on cloud architecture, service integrations, and operational visibility on common cloud (Google, AWS, Azure) platforms. Understanding of ecosystem deployment options and how to automate them via API calls is a huge asset.

What do we need from you?

  • Understand the end-to-end operations of complex Hadoop-based ecosystems and handle / configure core technologies such as HDFS, MapReduce, Spark, YARN, HBase, and ZooKeeper
  • Understand the dependencies and interactions between these core components, alternative configurations (i.e. MRv2 vs Spark, scheduling in YARN), availability characteristics and service recovery scenarios.
  • Identify workflow and job pipeline characteristics and tune the ecosystem to support high performance and scalability, from the infrastructure platform through to the application layers in the ecosystem.
  • Understand security tools and approaches available to configure different use cases based on clients needs, including being able to manage tools such as kerberos, AD, LDAP, the encryption at REST Hadoop services options and PKI concepts.
  • Understand end to end operations, deployment, troubleshooting, and tuning of streaming technologies such as Kafka and RabbitMQ for a wide variety of applications in a wide variety of environments, including on-prem and cloud.
  • Understand and enable metric collection at all layers of a complex infrastructure, ensuring good visibility for engineering and troubleshooting tasks, and ensure end to end monitoring of critical ecosystem components and workflows.
  • Understand the Hadoop toolset, how to manage and copy data between and within a Hadoop cluster, integrate with other ecosystems (for instance, cloud storage), configure replication and plan backups and resiliency strategies for data on the cluster.
  • Deep understanding of the Kafka ecosystem and the ability to troubleshoot and tune brokers, partition distribution, and topics.
  • Comprehensive systems hardware and network troubleshooting experience in physical, virtual and cloud platform environments, including the operation and administration of virtual and cloud infrastructure provider frameworks. Experience with at last one cloud provider (GCP, AWS, Azure) is required.
  • Experience with the design, development and deployment of at least one major configuration management framework (i.e. Puppet, Ansible, Chef) and one major infrastructure automation framework (i.e. Terraform, Spinnaker, CloudFormation). Knowledge of DevOps tools, processes, and culture (i.e. Git, continuous integration, test-driven development, Scrum).
  • A strong desire to learn and the ability to pick up new technologies and ecosystem components quickly, and establish their relevance, architecture and integration with existing systems.

What do you get in return?

  • Competitive total rewards package
  • Flexible work environment: Why commute? Work remotely from your home, theres no daily travel requirement to the office!
  • Outstanding people: Collaborate with the industrys top minds.
  • Substantial training allowance: Hone your skills or learn new ones; participate in professional development days, attend conferences, become certified, whatever you like!
  • Amazing time off: Start with a minimum 3 paid time off, 7 sick days, and 2 professional development days!
  • Office Allowance: Purchase a device of your choosing and personalise your work environment!
  • Fun, fun, fun: Blog during work hours ; take a day off and volunteer for your favorite charity.

Company Information