Related skills
apis observability sdks llm inference๐ Description
- Designing and operating scalable AI infra for LLM inference, prompts, and evaluation pipelines.
- Building self-service tools, SDKs, and APIs to move prototypes to production faster.
- Instrumenting production AI/ML workloads with standardized logging, tracing, and metrics.
- Implementing intelligent routing, caching, and provider optimisation via the LLM gateway.
- Driving adoption of shared platform services to replace bespoke solutions.
- Championing developer experience with documentation and responsive support.
๐ฏ Requirements
- Built and deployed production AI infra with enterprise-grade reliability and observability.
- Delivered self-service tools or APIs enabling multiple teams to accelerate AI/ML cycles.
- Implemented evaluation frameworks and monitoring to improve model performance, latency, cost.
- Led cost-reduction via intelligent routing or caching for AI compute.
- Migrated teams from bespoke to shared platform services, driving adoption.
- Prioritised developer experience through documentation and workflow enhancements.
๐ Benefits
- Share options at a Series B company
- 25 days of holiday + Polish bank holidays
- ยฃ5,000 GBP budget for Learning & Development
- Mental Health Support and Therapy via Spectrum Life
- Private Healthcare via Luxmed + Multisport (optional for B2B contractor)
- Top-spec laptop
Meet JobCopilot: Your Personal AI Job Hunter
Automatically Apply to Engineering Jobs. Just set your
preferences and Job Copilot will do the rest โ finding, filtering, and applying while you focus on what matters.
Help us maintain the quality of jobs posted on Empllo!
Is this position not a remote job?
Let us know!