About the job
Cerebras Systems is revolutionizing the AI landscape with the world’s largest AI chip, which is 56 times larger than traditional GPUs. Our innovative wafer-scale architecture empowers AI compute capabilities equivalent to dozens of GPUs on a single chip while maintaining the programming ease of a singular device. This groundbreaking approach enables Cerebras to provide unmatched training and inference speeds, allowing machine learning practitioners to seamlessly execute large-scale ML applications without the complexity of managing multiple GPUs or TPUs.
Cerebras proudly supports leading model labs, global enterprises, and pioneering AI-native startups. Notably, OpenAI has recently forged a multi-year partnership with Cerebras, committing to leverage 750 megawatts of power to enhance essential workloads with ultra-fast inference capabilities.
Thanks to our state-of-the-art wafer-scale architecture, Cerebras Inference delivers the fastest Generative AI inference solution globally, exceeding the speed of GPU-based hyperscale cloud inference services by over tenfold. This substantial speed enhancement is transforming user experiences with AI applications, facilitating real-time iterations and augmenting intelligence through advanced agentic computation.
Location Options: Sunnyvale, Toronto, or Vancouver
About the Role
We are on the lookout for a Compiler Engineer to innovate and implement new functionalities in our low-level compiler toolchain, which encompasses the compiler mid-end, backend, assembler, and linker, specifically targeting the individual cores within this unique architecture. Your primary responsibilities will include working within the compiler's infrastructure to enhance performance and efficiency across various applications.
