OpenAI is seeking a Technical Lead Software Engineer for their Inference team, focusing on high-performance model inference and infrastructure scaling. This role combines technical leadership with hands-on engineering, requiring expertise in CUDA optimization and distributed systems. The position involves leading the design and implementation of core inference infrastructure for frontier AI models, optimizing CUDA-based systems, and collaborating with researchers to build scalable solutions. The ideal candidate will have deep expertise in GPU computing, PyTorch, and NVIDIA's software stack, along with proven experience leading complex engineering initiatives. The role offers competitive compensation ($460K-$685K plus equity) and comprehensive benefits, including healthcare, 401(k) matching, and generous parental leave. This is an opportunity to shape the future of AI technology at one of the industry's leading companies, working on cutting-edge infrastructure that powers products used by consumers and enterprises worldwide. The position requires both technical depth in performance optimization and the ability to mentor and guide other engineers while working cross-functionally with research and product teams.