Required Skills

Data Analytics Big Data Data Warehousing Data Lakes Data Engineer

Work Authorization

  • US Citizen

  • Green Card

  • EAD (OPT/CPT/GC/H4)

  • H1B Work Permit

Preferred Employment

  • Corp-Corp

  • W2-Permanent

  • W2-Contract

  • Contract to Hire

Employment Type

  • Consulting/Contract

education qualification

  • UG :- - Not Required

  • PG :- - Not Required

Other Information

  • No of position :- ( 1 )

  • Post :- 15th Mar 2024

JOB DETAIL

Generative AI Data Engineer SME

6 Months

Remote

USC Only

Phone/Skype

Need candidates with all the mentions & highlighted skills

"Essential Requirements:

• 10+ years working with Enterprise-level Data Analytics, Big Data, Data Warehousing, Data Lakes, Data Lake houses, and Data Meshes.
• 10+ years working with data modeling tools.
• 5+ years building data pipelines for large customers.
• 10+ years working with data quality management tools, and data ETL/ELT tools.
• 5+ years working with data catalogs and data governance tools,
• 3+ years building data products and services with APIs.
• 3+ years working with data virtualization and federation tools.
• 2+ years of experience working in the field of Artificial Intelligence that leverages Big Data. This should be in a customer-facing services delivery role. 
• 3+ years of experience in Big Data database design.
• A good understanding of LLMs including prompt engineering, Retrieval Augmented Generation (RAG), fine tuning and training. 
• Strong knowledge of SQL, NoSQL Graph, and Vector databases. Experience with popular enterprise databases such as SQL Server, MySQL and Redis is a must.
• Experience with major data warehousing providers such as Teradata.
• Experience with data lake tools such as Databricks, Snowflake and Starburst.
• Proven experience building data pipelines and ETLs for both data transformation and multiple data source data extraction. Experience with automation of the deployment and execution of these pipelines.
 Experience with tools such as Apache Spark, Apache Hadoop, Informatica and similar data processing tools. Proficient knowledge of Python and SQL is a must.
• Proven experience with building test procedures, ensuring the quality, reliability, performance, and scalability of the data pipelines.
• Ability to develop applications that expose and use Restful APIs for data querying and ingestion.
• Understanding of typical AI tooling ecosystem including knowledge and experience of Kubernetes, MLOps and AIOps tools. 
• Ability to gain customer trust, ability to plan, organize and drive customer workshops. Good verbal and written communication skills in English is a must.
• Ability to work efficiently in highly collaborative teams using Agile methodologies such as Scrum or Kanban. Ability to have extended pairing sessions with customers, enabling knowledge transfer in complex domains.
• Ability to influence and interact with confidence and credibility at all levels within the Dell Technologies companies and with our customers, partners, and vendors.
• Experience working on project teams within a defined methodology while adhering to margin, planning and SOW requirements.
• Ability to be onsite during customer workshops and enablement sessions.

Desirable Requirements:

• General awareness of Dell Technologies products.
• Knowledge of industry widespread AI Studios and AI Workbenches is a plus.
• Experience preparing data for machine learning and Large Language Model ingestion and training (e.g., through vector databases, indexes, reinforcement learning, and RAG).
• Experience building and using Information Retrieval (IR) frameworks to support LLM inferencing.
• Working knowledge of Linux is a plus.
• Experience using Lean and Iterative Deployment Methodologies.
• Working knowledge of cloud technologies is a plus. 
• University Degree aligned to Data Engineering and/or Data Science is a plus.
• In possession of relevant industry certifications (e.g., Databricks Certified Data Engineer, Microsoft Certifications, etc.)"

 

Company Information