Related skills
grpc python kubernetes distributed systems goπ Description
- Architect distributed AI infrastructure control plane services for LLM inference.
- Design high-scale, multi-tenant inference cloud with availability and resiliency.
- Define SLOs and observability to keep the platform healthy at scale.
- Partner with Product, TPMs, and Eng Mgmt to align roadmaps with priorities.
- Work on new architecture initiatives for fleet optimization and AI-native networking.
π― Requirements
- Distributed systems mastery: cloud services, messaging, databases, IaC, observability, security.
- Cloud infra: VPCs, load balancers, Kubernetes; storage; VLAN/RDMA/Ethernet networking.
- AI/ML domain: Gen AI platforms, LLM hosting, inference workflows.
- Operational track record: running high-availability services across multiple regions.
- Open source engagement: experience with OSS and technical ownership.
- Go or Python expert; familiar with gRPC.
π Benefits
- We innovate with purpose.
- Career development resources and LinkedIn Learning access.
- Well-being benefits: EAP, local meetups, flexible time off.
- Compensation and equity: market-based pay, bonus, equity, ESPP.
Meet JobCopilot: Your Personal AI Job Hunter
Automatically Apply to Engineering Jobs. Just set your
preferences and Job Copilot will do the rest β finding, filtering, and applying while you focus on what matters.
Help us maintain the quality of jobs posted on Empllo!
Is this position not a remote job?
Let us know!