Related skills
redshift etl sql databricks data modelingπ Description
- Data Pipelines - Create and improve pipelines using Spark (PySpark, Spark SQL)
- Data Modeling - Design logical and physical schemas; enhance data models
- Cross-functional Collaboration - Interface with Product, Engineering, Data Science, Analytics/BI
- Build data expertise and own data quality across healthcare domains
- Leverage best practices to build next-gen data ecosystem for data collection and analysis
- Agile oriented development
π― Requirements
- BS degree in CS or related field, or equivalent experience
- 2+ yrs data engineering; one language (Scala or Python/PySpark) plus SQL
- 2+ yrs schema design, dimensional modeling, and large-scale data warehousing
- ETL design, implementation and maintenance
- Experience with Spark, Databricks, Presto, Hive, and Redshift; schedulers a plus
- BI experience: permissions, dashboards, data governance; strong communication
π Benefits
- Stock options: 70,000 options and benefits
- Health insurance, 401(k), and paid time off
- Flexible work arrangements
- Internal mobility and professional development
Meet JobCopilot: Your Personal AI Job Hunter
Automatically Apply to Data Jobs. Just set your
preferences and Job Copilot will do the rest β finding, filtering, and applying while you focus on what matters.
Help us maintain the quality of jobs posted on Empllo!
Is this position not a remote job?
Let us know!