Do you thrive on building the future of AI infrastructure?
Are you ready to lead a world-class team at the intersection of AI and edge computing?
Join the Akamai Inference Cloud Team!
The Akamai Inference Cloud team is part of Akamai's Cloud Technology Group. We design, implement, deploy and operate AI platforms that enable customers to run inference models and developers to create AI applications with unmatched performance, compliance, and economics.
Partner with the best
As a Senior Engineering Manager, you will build and lead a high-performing team of platform and ML engineers. Your team will design and develop our globally distributed AI inference platform, delivering OpenAI-compatible endpoints and orchestrating inference workloads across regions.
The ability to recruit, hire, and build a team of exceptional engineers while establishing technical vision is paramount for this role
As a Senior Engineering Manager, you will be responsible for:
- Building and scaling a world-class engineering team from the ground up, recruiting top talent in AI infrastructure and ML operations
- Leading the technical strategy for a global AI inference platform that is performant, compliant, economical, and explainable
- Providing the availability, performance, scalability, and security of Akamai Inference Cloud
- Designing global traffic orchestration for AI workloads and establishing platform standards and blueprints for production-grade AI applications
- Making critical decisions on AI tooling based on technical evaluation while ensuring compliance with regulatory requirements (i.e. FedRAMP, GDPR, SOX )
Do what you love
To be successful in this role you will:
- 10 years of relevant experience and a Bachelor's degree or its equivalent experience building and scaling high-performing teams that shipped successful AI/ML products
- Possess hands-on experience with AI inference optimization, model serving, and LLM deployment at scale with deep knowledge of inference frameworks (TensorRT, vLLM, TorchServe, Triton)
- Have an understanding of containerization strategies for AI workloads with hardware-specific optimizations, and possess opinions on what makes an AI platform successful
- Show proficiency with cloud-native technologies including Kubernetes and distributed systems with proven experience operating services at global scale
- Demonstrate expertise in building highly available, low-latency platforms with strict SLOs and cost optimization strategies for compute-intensive AI workloads
- Possess knowledge of AI application platforms, AI safety and GPU infrastructure and hardware acceleration is ideal
Work in a way that works for you
FlexBase, Akamai's Global Flexible Working Program, is based on the principles that are helping us create the best workplace in the world. When our colleagues said that flexible working was important to them, we listened. We also know flexible working is important to many of the incredible people considering joining Akamai. FlexBase, gives 95% of employees the choice to work from their home, their office, or both (in the country advertised). This permanent workplace flexibility program is consistent and fair globally, to help us find incredible talent, virtually anywhere. We are happy to discuss working options for this role and encourage you to speak with your recruiter in more detail when you apply.