Lead Member of Technical Staff, Inference Infrastructure

Added
42 minutes ago
Type
Full time
Salary
Salary not provided

Related skills

golang linux kubernetes distributed systems multi-cloud

📋 Description

  • Lead design of high-performance, scalable ML systems.
  • Set technical direction for AI platforms powering NLP apps.
  • Drive architecture for deploying optimized NLP models to production.
  • Ensure low latency, high throughput, high availability.
  • Serve as primary customer contact for customized deployments.
  • Mentor engineers to raise the technical bar across the team.

🎯 Requirements

  • 8+ years building production infra at scale with leadership.
  • Led architecture of large distributed systems with Kubernetes and GPU workloads.
  • Deep Kubernetes dev/prod coding expertise and standards.
  • Experience across GCP, Azure, AWS, OCI, and multi-cloud on-prem/hybrid.
  • Lead Linux-based compute environments at scale (design/deploy/support/troubleshoot).
  • Proficiency in Golang, C++, or similar; set standards and reviews.

🎁 Benefits

  • Open and inclusive culture and work environment.
  • Work with a team on the cutting edge of AI research.
  • Weekly lunch stipend, in-office lunches & snacks.
  • Full health and dental benefits, including mental health budget.
  • 100% Parental Leave top-up for up to 6 months.
  • Remote-flexible with offices in NYC, SF, Toronto, London and Paris.
Share job

Meet JobCopilot: Your Personal AI Job Hunter

Automatically Apply to Engineering Jobs. Just set your preferences and Job Copilot will do the rest — finding, filtering, and applying while you focus on what matters.

Related Engineering Jobs

See more Engineering jobs →