Related skills
java python apache spark spring boot apache airflow๐ Description
- Engage with stakeholders and PMs to understand needs and design solutions.
- Collaborate with product managers, data scientists, and engineers to deliver solutions.
- Design, develop, and maintain data pipelines/ETL with Spark and Airflow.
- Optimize data storage and processing for reliability, scalability, performance.
- Develop and optimize queries/analytics with Pinot and Trino for large datasets.
- Monitor and troubleshoot data systems to minimize downtime.
๐ฏ Requirements
- 5+ years building large-scale distributed systems.
- 5+ years in BigData technologies.
- Advanced SQL skills with query optimization for large datasets.
- Experience with AWS, GCP, or Azure and Docker/Kubernetes.
- Distributed data processing, data warehousing, and real-time data pipelines.
- Experience with Python, Scala, or Java.
๐ Benefits
- Global access to mental health and financial wellness resources.
- Healthcare, life, disability, and retirement options (401(k)/pension).
- Paid time off for vacation and personal reasons.
- Hybrid schedule: in-office Mon-Thu, Fridays can be remote.
Meet JobCopilot: Your Personal AI Job Hunter
Automatically Apply to Data Jobs. Just set your
preferences and Job Copilot will do the rest โ finding, filtering, and applying while you focus on what matters.
Help us maintain the quality of jobs posted on Empllo!
Is this position not a remote job?
Let us know!