US Citizen
Green Card
H1B Work Permit
Corp-Corp
W2-Permanent
W2-Contract
Contract to Hire
Consulting/Contract
UG :- - Not Required
PG :- - Not Required
No of position :- ( 1 )
Post :- 27th Mar 2023
Key Responsibilities:
Translates business requirements and specifications into logical program designs, modules, stable application systems, and data solutions with occasional guidance from senior colleagues; partners with Product Team to understand business needs and functional specifications
Develops, configures, or modifies integrated business and/or enterprise application solutions within various computing environments by designing and coding component-based applications using various programming languages
Conducts the implementation and maintenance of complex business and enterprise data solutions to ensure successful deployment of released applications
Supports systems integration testing (SIT) and user acceptance testing (UAT), provides insight into defining test plans, and ensures quality software deployment
Participates in the end-to-end product lifecycle by applying and sharing an in-depth understanding of company and industry methodologies, policies, standards, and controls
Understands Computer Science and/or Computer Engineering fundamentals; knows software architecture and readily applies this to Data or Platform solutions
Automates and simplifies team development, test, and operations processes; develops conceptual, logical, and physical architectures consisting of one or more viewpoints (business, application, data, and infrastructure) required for business solution delivery
Solves difficult technical problems; solutions are testable, maintainable, and efficient
Data Engineering Qualifications:
2 years of experience in Hadoop or any Cloud Bigdata components
Expertise in Java/Scala/Python, SQL, Scripting, Teradata, Hadoop (Sqoop, Hive, Pig, Map Reduce), Spark (Spark Streaming, MLib), Kafka or equivalent Cloud Bigdata components
Preferred Qualifications:
Degree in Computer Science, Engineering, Mathematics, Data Science, Analytics, Information System, or related quantitative field.
3 years hand-on experience in building robust, reliable, and scalable machine learning pipelines from ground up as well as transitioning from MVP to production (ingestion, scheduling, security, notifications, validation, backups, optimizations).
Experience with CI/CD pipelines such as Concourse, Jenkins.
Experience with distributed systems (Spark, Hive, HDFS, Hadoop, HBase, Druid, Cloudera, Kafka) along with Airflow to orchestrate data pipeline.
Experience with process optimizations and fine-tuning existing applications in distributed computing environment.
Write quality code in Python/Pyspark followed by unit tests and documentation.
Experience in configuring/building data quality frameworks - PyDeequ, great expectations or any other tools to increase reliability.
Define, execute, and operate monitoring and alerting steps over critical SLA's - (Prometheus/Nagios, Grafana).
Out-of-the box thinker and deep diver to provide efficient solutions to problems
Agile development skills and experience.
Data Science Exposure.