About the Role
Our Inference team is responsible for building and maintaining critical systems that serve AI models to millions of users worldwide.
We bring these models to life by serving them via large-scale compute-agnostic inference deployments. Our goal is to maximize efficiency while enabling breakthrough research by giving scientists high-performance infrastructure to develop next-generation models.
Our Challenges:
* Tackle complex, distributed systems challenges across multiple accelerator families and emerging AI hardware running in multiple cloud platforms.
Your Key Skills:
* Significant software engineering experience with a focus on distributed systems.