Cerebras Systems, a pioneering company in AI hardware, is seeking an AI Infrastructure Operations Engineer to manage their cutting-edge machine learning compute clusters. The role involves working with the world's largest computer chip, the Wafer-Scale Engine (WSE), and the systems that harness its power. The position requires deep expertise in Linux systems, containerization, and distributed systems management.
The successful candidate will be responsible for ensuring the health, performance, and availability of Cerebras' infrastructure while maximizing compute capacity for AI initiatives. This is a critical role that combines hands-on technical work with strategic infrastructure management, requiring both deep technical knowledge and strong operational skills.
Cerebras Systems has established itself as a leader in AI computing, with partnerships across multiple industries including healthcare, where they recently announced a multi-year partnership with Mayo Clinic. Their technology delivers unprecedented AI computing power, with their chip being 56 times larger than traditional GPUs and their inference solution being 10 times faster than GPU-based cloud services.
The role offers an opportunity to work at the forefront of AI infrastructure, managing some of the most advanced computing systems in the world. The position demands a combination of technical expertise, operational excellence, and the ability to work in a fast-paced environment. The ideal candidate will have extensive experience with Linux systems, containerization, and large-scale infrastructure management, along with strong problem-solving and communication skills.
This is an excellent opportunity for someone passionate about AI infrastructure who wants to work with cutting-edge technology while making a significant impact on the future of AI computing. The role offers the chance to work with a team that's pushing the boundaries of what's possible in AI hardware and infrastructure.