Related skills
golang linux kubernetes multi-cloud gpus📋 Description
- Build and operate high-performance ML inference infrastructure at scale
- Deploy optimized NLP models to production with low latency and high throughput
- Ensure reliability, availability, and cost-effective operation of the AI platform
- Collaborate with cross-functional teams to deploy models to production
- Interface with customers for customized deployments meeting their needs
- Contribute to improving reliability and efficiency of the platform
🎯 Requirements
- 5+ years of engineering infra experience at scale
- Designing large, highly available distributed systems with Kubernetes and GPU workloads
- Kubernetes development, production coding and support
- Experience with GCP, Azure, AWS, OCI; multi-cloud on-prem/hybrid serving
- Experience in Linux-based computing environments
- Compute/storage/network resource and cost management
- Strong collaboration and troubleshooting skills
- Proficiency in Golang or C++ for high-performance servers
🎁 Benefits
- An open and inclusive culture and work environment
- Work closely with a team on the cutting edge of AI research
- Weekly lunch stipend, in-office lunches & snacks
- Full health and dental benefits, including mental health budget
- 100% Parental Leave top-up for up to 6 months
- Remote-flexible, offices in Toronto, New York, San Francisco, London and Paris, plus a co-working stipend
Meet JobCopilot: Your Personal AI Job Hunter
Automatically Apply to Engineering Jobs. Just set your
preferences and Job Copilot will do the rest — finding, filtering, and applying while you focus on what matters.
Help us maintain the quality of jobs posted on Empllo!
Is this position not a remote job?
Let us know!