Staff Software Engineer, Inference Infrastructure

Added
10 days ago
Type
Full time
Salary
Salary not provided

Related skills

golang linux kubernetes multi-cloud gpus

📋 Description

  • Build and operate high-performance ML inference infrastructure at scale
  • Deploy optimized NLP models to production with low latency and high throughput
  • Ensure reliability, availability, and cost-effective operation of the AI platform
  • Collaborate with cross-functional teams to deploy models to production
  • Interface with customers for customized deployments meeting their needs
  • Contribute to improving reliability and efficiency of the platform

🎯 Requirements

  • 5+ years of engineering infra experience at scale
  • Designing large, highly available distributed systems with Kubernetes and GPU workloads
  • Kubernetes development, production coding and support
  • Experience with GCP, Azure, AWS, OCI; multi-cloud on-prem/hybrid serving
  • Experience in Linux-based computing environments
  • Compute/storage/network resource and cost management
  • Strong collaboration and troubleshooting skills
  • Proficiency in Golang or C++ for high-performance servers

🎁 Benefits

  • An open and inclusive culture and work environment
  • Work closely with a team on the cutting edge of AI research
  • Weekly lunch stipend, in-office lunches & snacks
  • Full health and dental benefits, including mental health budget
  • 100% Parental Leave top-up for up to 6 months
  • Remote-flexible, offices in Toronto, New York, San Francisco, London and Paris, plus a co-working stipend
Share job

Meet JobCopilot: Your Personal AI Job Hunter

Automatically Apply to Engineering Jobs. Just set your preferences and Job Copilot will do the rest — finding, filtering, and applying while you focus on what matters.

Related Engineering Jobs

See more Engineering jobs →