About the job
Cerebras Systems is at the forefront of AI technology, developing the largest AI chip in the world, which is 56 times the size of traditional GPUs. Our innovative wafer-scale architecture delivers the AI compute power equivalent to dozens of GPUs on a single chip, while maintaining the simplicity of programming a single device. This unique technology enables Cerebras to provide unparalleled training and inference speeds, allowing machine learning professionals to seamlessly execute large-scale ML applications without the complexity of managing numerous GPUs or TPUs.
Our clientele includes leading model laboratories, multinational corporations, and pioneering AI-driven startups. Notably, OpenAI has recently partnered with Cerebras to leverage 750 megawatts of scale, revolutionizing critical workloads with ultra-fast inference capabilities.
Thanks to our groundbreaking wafer-scale technology, Cerebras Inference offers the fastest Generative AI inference solution globally, outperforming GPU-based hyperscale cloud inference services by more than tenfold. This significant increase in speed enhances the user experience for AI applications, enabling real-time iterations and amplifying intelligence through enhanced computational capabilities.
About The Role
In this role as a Frontend Engineer on our AI cloud platform, you will be instrumental in developing our customer-facing inference, training, and administrative consoles as well as API experiences. You will design and implement responsive, user-friendly frontend interfaces that ensure an optimal experience for developers, efficiently managing high traffic and throughput.
Your expertise in the latest web development frameworks and best practices, along with a strong focus on design and user experience, will be key to our team's success.

