Software Engineer, Model Inference

AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity.
$200,000 - $370,000
Backend
Senior Software Engineer
In-Person
501 - 1,000 Employees
3+ years of experience
AI
This job posting may no longer be active. You may be interested in these related jobs instead:
Software Engineer, Backend

Senior Backend Software Engineer role at OpenAI working on ChatGPT for Work team, building enterprise solutions with competitive compensation between $245K-$385K

Software Engineer, Financial Engineering

Senior Software Engineer role at OpenAI focusing on building and architecting next-generation billing and monetization systems.

Software Engineer, Internal Applications - Enterprise

Senior Software Engineer role at OpenAI focusing on internal applications and enterprise infrastructure automation

Software Development Engineer, FSx for OpenZFS

Senior Software Development Engineer position at Amazon FSx for OpenZFS team, building cloud storage solutions and managing high-performance file systems.

Senior Platform Software Engineer, PCIe

Senior Platform Software Engineer role at NVIDIA focusing on PCIe architecture and GPU-based AI server development, offering competitive compensation and the opportunity to work on cutting-edge AI computing infrastructure.

Description For Software Engineer, Model Inference

OpenAI is seeking a Software Engineer for Model Inference to join their Applied AI Engineering team in San Francisco. This role is crucial for scaling up critical inference infrastructure that efficiently services customer requests for state-of-the-art AI models like GPT-4 and DALL-E.

Key responsibilities include:

  • Collaborating with ML researchers, engineers, and product managers to productionize latest technologies
  • Implementing new techniques, tools, and architecture to enhance model performance, latency, throughput, and efficiency
  • Developing tools for identifying bottlenecks and instability sources, then designing and implementing solutions
  • Optimizing code and Azure VM fleet to maximize hardware utilization

Ideal candidates should have:

  • Understanding of modern ML architectures and optimization for inference
  • End-to-end problem-solving skills
  • At least 3 years of professional software engineering experience
  • Expertise in HPC technologies (InfiniBand, MPI, CUDA)
  • Experience with production distributed systems
  • Self-direction and ability to identify important problems
  • Humble attitude and eagerness to help colleagues

OpenAI offers a competitive salary range of $200K – $370K and is committed to diversity, equal opportunity, and providing reasonable accommodations to applicants with disabilities.

Join OpenAI in shaping the future of AI technology and ensuring its benefits are widely shared.

Last updated 7 months ago

Responsibilities For Software Engineer, Model Inference

  • Work with ML researchers, engineers, and product managers to bring latest technologies into production
  • Introduce new techniques, tools, and architecture to improve model performance, latency, throughput, and efficiency
  • Build tools to identify bottlenecks and sources of instability, then design and implement solutions
  • Optimize code and Azure VM fleet to maximize hardware utilization

Requirements For Software Engineer, Model Inference

Python
  • Understanding of modern ML architectures and optimization for inference
  • At least 3 years of professional software engineering experience
  • Expertise in core HPC technologies: InfiniBand, MPI, CUDA
  • Experience architecting, observing, and debugging production distributed systems
  • Self-direction and ability to identify important problems to work on

Interested in this job?