US Citizen
Green Card
EAD (OPT/CPT/GC/H4)
H1B Work Permit
Corp-Corp
W2-Permanent
W2-Contract
Contract to Hire
Consulting/Contract
UG :- - Not Required
PG :- - Not Required
No of position :- ( 1 )
Post :- 15th Mar 2024
Generative AI Data Engineer SME
6 Months
Remote
USC Only
Phone/Skype
Need candidates with all the mentions & highlighted skills
"Essential Requirements:
• 10+ years working with Enterprise-level Data Analytics, Big Data, Data Warehousing, Data Lakes, Data Lake houses, and Data Meshes.
• 10+ years working with data modeling tools.
• 5+ years building data pipelines for large customers.
• 10+ years working with data quality management tools, and data ETL/ELT tools.
• 5+ years working with data catalogs and data governance tools,
• 3+ years building data products and services with APIs.
• 3+ years working with data virtualization and federation tools.
• 2+ years of experience working in the field of Artificial Intelligence that leverages Big Data. This should be in a customer-facing services delivery role.
• 3+ years of experience in Big Data database design.
• A good understanding of LLMs including prompt engineering, Retrieval Augmented Generation (RAG), fine tuning and training.
• Strong knowledge of SQL, NoSQL Graph, and Vector databases. Experience with popular enterprise databases such as SQL Server, MySQL and Redis is a must.
• Experience with major data warehousing providers such as Teradata.
• Experience with data lake tools such as Databricks, Snowflake and Starburst.
• Proven experience building data pipelines and ETLs for both data transformation and multiple data source data extraction. Experience with automation of the deployment and execution of these pipelines.
• Experience with tools such as Apache Spark, Apache Hadoop, Informatica and similar data processing tools. Proficient knowledge of Python and SQL is a must.
• Proven experience with building test procedures, ensuring the quality, reliability, performance, and scalability of the data pipelines.
• Ability to develop applications that expose and use Restful APIs for data querying and ingestion.
• Understanding of typical AI tooling ecosystem including knowledge and experience of Kubernetes, MLOps and AIOps tools.
• Ability to gain customer trust, ability to plan, organize and drive customer workshops. Good verbal and written communication skills in English is a must.
• Ability to work efficiently in highly collaborative teams using Agile methodologies such as Scrum or Kanban. Ability to have extended pairing sessions with customers, enabling knowledge transfer in complex domains.
• Ability to influence and interact with confidence and credibility at all levels within the Dell Technologies companies and with our customers, partners, and vendors.
• Experience working on project teams within a defined methodology while adhering to margin, planning and SOW requirements.
• Ability to be onsite during customer workshops and enablement sessions.
Desirable Requirements:
• General awareness of Dell Technologies products.
• Knowledge of industry widespread AI Studios and AI Workbenches is a plus.
• Experience preparing data for machine learning and Large Language Model ingestion and training (e.g., through vector databases, indexes, reinforcement learning, and RAG).
• Experience building and using Information Retrieval (IR) frameworks to support LLM inferencing.
• Working knowledge of Linux is a plus.
• Experience using Lean and Iterative Deployment Methodologies.
• Working knowledge of cloud technologies is a plus.
• University Degree aligned to Data Engineering and/or Data Science is a plus.
• In possession of relevant industry certifications (e.g., Databricks Certified Data Engineer, Microsoft Certifications, etc.)"