About the job
Cerebras Systems is pioneering the future of artificial intelligence with the world's largest AI chip, which is 56 times larger than traditional GPUs. Our innovative wafer-scale architecture allows us to deliver the AI compute power equivalent to dozens of GPUs on a single chip, simplifying programming and deployment. This unique approach enables Cerebras to achieve unparalleled training and inference speeds, allowing machine learning practitioners to run extensive ML applications without the complexity of managing multiple GPUs or TPUs.
Our clientele includes prestigious model labs, major global corporations, and groundbreaking AI-focused startups. Recently, OpenAI announced a multi-year partnership with Cerebras, aimed at leveraging 750 megawatts of scale to revolutionize critical workloads with ultra-high-speed inference capabilities.
Thanks to our cutting-edge wafer-scale architecture, Cerebras Inference delivers the fastest Generative AI inference solution globally, exceeding the speed of GPU-based hyperscale cloud inference services by more than 10 times. This significant enhancement in speed is redefining the user experience of AI applications, enabling real-time iteration and augmenting intelligence through enhanced computational capabilities.
The Role:
As a Senior/Staff Engineer, you will play a key role in the bring-up and optimization of Cerebras’s Wafer Scale Engine (WSE). The ideal candidate will possess substantial experience in delivering end-to-end solutions while collaborating closely with teams across chip design, system performance, software development, and productization.

