Related skills
aws sql python scala data modelingπ Description
- Design and build batch and streaming data pipelines for ingestion and analytics.
- Own end-to-end delivery of complex data initiatives, including architecture and deployment.
- Develop and evolve the data platform to support large-scale processing with cloud-native tech.
- Automate data operations (validation, quality checks, alerts, backfills, recovery).
- Optimize cost, performance, and reliability of data workloads.
- Partner with Data Science, Product, and Engineering to translate requirements into solutions.
π― Requirements
- 5+ years of hands-on data engineering experience, building and maintaining production-grade data platforms and pipelines.
- Strong programming skills in Python or Scala for data processing, and SQL for data analytics.
- Deep experience with distributed data processing frameworks like Apache Spark, including performance tuning.
- Proven experience building data solutions using AWS services (EMR, Lambda, S3, etc).
- Strong understanding of data modeling and data warehousing concepts, including partitioning and schema design.
- Experience operating and supporting production pipelines, including monitoring, alerting, incident response.
π Benefits
- Equal opportunity employer; values diversity.
- Accommodation available during hiring process.
- Opportunity to work on identity verification tech at scale.
Meet JobCopilot: Your Personal AI Job Hunter
Automatically Apply to Data Jobs. Just set your
preferences and Job Copilot will do the rest β finding, filtering, and applying while you focus on what matters.
Help us maintain the quality of jobs posted on Empllo!
Is this position not a remote job?
Let us know!