Related skills
snowflake sql python hadoop kafkaπ Description
- Design, develop, and maintain scalable data pipelines for ingestion, transformation, and delivery.
- Build and optimize workflows for semantic representations of unstructured data for search.
- Architect analytics and dashboards for natural language queries and AI insights.
- Manage prompt engineering, orchestration, and model fine-tuning for conversational interfaces.
- Oversee vector stores and indexing for retrieval-augmented generation (RAG).
- Collaborate with data stakeholders to translate requirements into scalable ML/data solutions.
π― Requirements
- 8+ years as Data Engineer with 2+ years in MLOps.
- Excellent English communication skills.
- Python for data engineering tasks including transformation and large-scale processing.
- Deep understanding of vector databases and RAG architectures.
- Experience with AWS or Azure ML for managed LLM deployments.
- Spark, Hadoop, and Kafka for distributed processing.
π Benefits
- DataStax AstraDB, LangChain, LlamaIndex for LLM apps and RAG.
- Open-source LLM frameworks like Transformers and LLaMA-4.
- MLOps tooling and CI/CD pipelines for model versioning and deployments.
Meet JobCopilot: Your Personal AI Job Hunter
Automatically Apply to Data Jobs. Just set your
preferences and Job Copilot will do the rest β finding, filtering, and applying while you focus on what matters.
Help us maintain the quality of jobs posted on Empllo!
Is this position not a remote job?
Let us know!