Added
1 day ago
Location
Type
Full time
Salary
Salary not provided
Related skills
distributed systems traffic_routing gpu_utilization provider_economics๐ Description
- Build the inference platform powering Cursor's AI interactions.
- Own the full inference path to improve speed, reliability, and cost.
- Develop the inference gateway to abstract provider APIs for config-based onboarding.
- Design cross-provider failover to prevent user-visible degradation.
- Implement routing backpressure and admission control for traffic spikes.
- Collaborate with teams; participate in onsite discussions and projects.
๐ฏ Requirements
- Deep experience building high-throughput, low-latency distributed systems.
- Experience in inference serving, traffic routing, or real-time data pipelines.
- Ability to reason about cost/performance tradeoffs at scale (GPU utilization).
- Strong software engineering fundamentals; shipping production systems.
- Good calls balancing reliability, cost, latency, and user experience.
- Collaborative, proactive team player with cross-functional partners.
๐ Benefits
- Onsite work in our office with a collaborative culture.
- Flat organizational structure and a talented, passionate team.
- Opportunity to build production-grade AI infrastructure at scale.
- Competitive compensation package and strong growth potential.
Meet JobCopilot: Your Personal AI Job Hunter
Automatically Apply to Engineering Jobs. Just set your
preferences and Job Copilot will do the rest โ finding, filtering, and applying while you focus on what matters.
Help us maintain the quality of jobs posted on Empllo!
Is this position not a remote job?
Let us know!