Related skills
aws sql python databricks apache sparkπ Description
- Design, implement, and optimize big data pipelines in Databricks.
- Develop scalable ETL workflows to process large datasets.
- Leverage Apache Spark for distributed data processing and real-time analytics.
- Implement data governance, security policies, and compliance standards.
- Optimize data lakehouse architectures for performance and cost-efficiency.
- Collaborate with data scientists, analysts, and engineers to enable AI/ML workflows.
π― Requirements
- Bachelor's or Master's degree in Computer Science, Data Engineering, or a related field.
- 5+ years of hands-on experience with Databricks and Apache Spark.
- Proficiency in SQL, Python, or Scala for data processing and analysis.
- Experience with cloud platforms (AWS, Azure, or GCP) for data engineering.
- Strong knowledge of ETL frameworks, data lakes, and Delta Lake architecture.
- Experience with CI/CD tools and DevOps best practices.
π Benefits
- Hybrid flexibility: 3 days/wk in downtown Toronto; Calgary applicants welcome.
- Fully covered health, dental, and vision insurance from day one.
- Growth & Learning: continuous opportunities and technical direction.
- Inclusive culture and accommodation support.
Meet JobCopilot: Your Personal AI Job Hunter
Automatically Apply to Data Jobs. Just set your
preferences and Job Copilot will do the rest β finding, filtering, and applying while you focus on what matters.
Help us maintain the quality of jobs posted on Empllo!
Is this position not a remote job?
Let us know!