About the job
Cerebras Systems is at the forefront of AI technology, recognized for developing the world's largest AI chip, which is 56 times larger than traditional GPUs. Our innovative wafer-scale architecture delivers AI compute power equivalent to dozens of GPUs on a single chip, simplifying programming and enhancing performance.
Our clients include leading model laboratories, global corporations, and pioneering AI-native startups. Notably, OpenAI has recently partnered with Cerebras to deploy 750 megawatts of scale, revolutionizing critical workloads through ultra-high-speed inference.
With our groundbreaking wafer-scale architecture, Cerebras Inference provides the fastest Generative AI inference solution globally, outperforming GPU-based hyperscale cloud inference services by over 10 times. This significant boost in speed is reshaping the user experience of AI applications, enabling real-time iterations and enhanced intelligence via advanced computation capabilities.
About The Role
We are looking for a skilled Platform Software Engineer to join our team dedicated to building the Cerebras Inference Platform. In this role, you will play a crucial part in designing, developing, and operating the core backend services and APIs that drive our Inference platform. Your contributions will empower our customers to effortlessly deploy, manage, and serve inference workloads on dedicated Cerebras systems.

