Added
28 minutes ago
Location
Type
Full time
Salary
Upgrade to Premium to se...
Related skills
distributed systems observability autoscaling gpu model servingπ Description
- Design and implement core systems and APIs powering Model Serving.
- Drive architecture decisions to optimize CPU/GPU performance and autoscaling.
- Contribute to components across serving infra: containers, routing, observability.
- Collaborate with product, platform, and research teams to build reliable systems.
- Lead technical initiatives to improve latency, availability, and cost.
- Establish best practices for code quality, testing, and readiness; mentor engineers.
π― Requirements
- 5+ years building and operating large-scale distributed systems.
- Experience in model serving, inference systems, or related infra (routing, autoscaling).
- Strong foundation in algorithms, data structures, and system design for low-latency serving.
- Proven ability to deliver technically complex, high-impact initiatives with measurable value.
- Experience architecting large-scale CPU/GPU inference systems.
- Strong communication and collaboration across teams in fast-moving environments.
π Benefits
- Pay range transparency is provided; see the pay zone mapping page.
- Benefits vary by region; visit the Databricks benefits site for details.
Meet JobCopilot: Your Personal AI Job Hunter
Automatically Apply to Engineering Jobs. Just set your
preferences and Job Copilot will do the rest β finding, filtering, and applying while you focus on what matters.
Help us maintain the quality of jobs posted on Empllo!
Is this position not a remote job?
Let us know!